Evaluating Fine-Tuning Strategies for a General-Purpose LLM
Based on the scenario below, which dataset should the team choose to best achieve their goal? Justify your decision by explaining the relationship between the diversity of adaptation tasks and the resulting model's capabilities.
0
1
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Evaluation in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Evaluating Fine-Tuning Strategies for a General-Purpose LLM
A development team is fine-tuning a large language model to serve as a general-purpose assistant capable of handling a wide variety of user queries. They have two potential datasets for this process:
- Dataset A: A large dataset with 2 million examples, all focused on a single, complex task: summarizing scientific research papers.
- Dataset B: A smaller dataset with 200,000 examples, but spread across 150 different tasks, such as question-answering, creative writing, translation, and code generation.
Based on principles of effective model fine-tuning, which dataset is more likely to produce a better general-purpose assistant, and why?
Evaluating a Fine-Tuning Strategy for a Specialized LLM