Adapting a Language Model for Different Conversational Tasks
A machine learning team is using a pre-trained language model for two different projects. Project A involves translating individual, unrelated sentences from English to French. Project B involves building a customer service chatbot that engages in extended, back-and-forth conversations. Explain why the structure of the input data used to fine-tune the model for Project B must be fundamentally different from the data used for Project A.
0
1
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Application in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Example of a Multi-Turn Conversation for LLM Fine-Tuning
A development team is preparing two separate datasets to fine-tune a language model. The first dataset is for a tool that summarizes individual, self-contained documents. The second is for a conversational assistant designed to help users troubleshoot problems over several back-and-forth exchanges. Which statement best analyzes the fundamental difference in how the input data must be structured for these two tasks?
Diagnosing a Conversational AI Fine-Tuning Issue
Adapting a Language Model for Different Conversational Tasks