Analyzing Fine-Tuning Strategies for General-Purpose LLMs
A research team is developing a general-purpose chatbot. They have two options for fine-tuning their pre-trained language model:
- Use a small, highly-curated dataset of 1,000 high-quality examples focused exclusively on customer service conversations.
- Use a large, diverse dataset of 500,000 examples covering a wide range of topics including creative writing, coding, summarization, and general Q&A.
Analyze the likely outcome for the model's general capabilities with each approach. Explain the underlying principle that makes one strategy more suitable than the other for creating a model that can follow a wide range of instructions.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
An AI development team has fine-tuned a large language model that excels at summarizing legal documents. However, the model performs poorly on tasks outside of the legal domain, such as creative writing and answering general science questions. Which of the following strategies is most likely to broaden the model's capabilities to handle this wider range of instructions effectively?
Evaluating a Fine-Tuning Strategy for a General-Purpose AI
Analyzing Fine-Tuning Strategies for General-Purpose LLMs