Rationale for Modifying a Pre-trained Model
A language model has been pre-trained on a large dataset to predict the next word in a sentence. You now want to adapt this model for a new task: classifying news articles into categories like 'Sports', 'Technology', and 'Politics'. Explain why the final layer of the pre-trained model must be replaced before you can begin training on the new classification task.
0
1
Tags
Ch.1 Pre-training - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Comprehension in Revised Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Troubleshooting a Model Adaptation Pipeline
A machine learning engineer has successfully pre-trained a large language model on a massive text corpus with the objective of predicting the next word in a sequence. To adapt this model for a new task of classifying customer reviews as 'positive', 'negative', or 'neutral', the engineer's first step is to remove the model's final output layer. What is the most accurate justification for this action?
Rationale for Modifying a Pre-trained Model