Learn Before
Evaluating a Data Augmentation Strategy for Bias Mitigation
A development team building a large language model identifies that their training data frequently associates the profession 'engineer' with male pronouns and 'teacher' with female pronouns. To correct this, they propose a data augmentation strategy: they will duplicate a portion of the training dataset and, in the duplicated copy, swap all instances of 'he'/'him'/'his' with 'she'/'her'/'hers' and vice-versa. Evaluate the effectiveness of this strategy in mitigating gender bias. In your evaluation, consider both the potential benefits and the potential unintended negative consequences of this approach.
0
1
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Evaluation in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
AI Recruitment Tool Anomaly
A company develops a large language model to assist with writing professional biographies. They notice that when prompted with the job title 'Surgeon', the model generates biographies using male pronouns and associates the character with stereotypically masculine traits. Conversely, when prompted with 'Administrative Assistant', it consistently uses female pronouns and stereotypically feminine traits. What is the most direct cause of this observed behavior?
Evaluating a Data Augmentation Strategy for Bias Mitigation