Learn Before
Challenge of Articulating Human Preferences for Data Annotation
A primary difficulty in LLM alignment is that human values and expectations are complex and hard to describe. It can be challenging for people to articulate what is ethically correct or culturally appropriate, which makes the task of collecting and annotating fine-tuning data far less straightforward compared to tasks with objective outputs.
0
1
References
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Ch.2 Generative Models - Foundations of Large Language Models
Related
Desired Qualities of Value-Aligned LLMs
Example of Value Alignment: Refusing Harmful Requests
Difficulty of Encoding Human Values in Datasets
Reinforcement Learning from Human Feedback (RLHF)
A user asks a large language model: "Summarize the arguments for and against using genetically modified organisms (GMOs) in agriculture." Consider two possible responses:
Model A's Response: "Genetically modified organisms are a triumph of modern science, allowing for higher crop yields and resistance to pests. They are essential for feeding the world's growing population and concerns about them are largely unscientific and based on fear."
Model B's Response: "Arguments for GMOs often highlight benefits such as increased crop yields, enhanced nutritional content, and resistance to pests and diseases, which can contribute to food security. Arguments against them frequently raise concerns about potential long-term environmental impacts, the risk of cross-pollination with non-GMO crops, and the socio-economic effects on small-scale farmers."
Which model's response better demonstrates successful alignment with human values, and why?
Evaluating an LLM's Response to a Sensitive Request
Challenge of Articulating Human Preferences for Data Annotation
A large language model that accurately and efficiently follows every user instruction without deviation is considered perfectly aligned with human values.
Role of Fine-Tuning in Value Alignment