Evaluating a Model's Training Task
A language model is trained to determine if Sentence B is the direct successor to Sentence A. The training data consists of 'positive' pairs (consecutive sentences from the same document) and 'negative' pairs (two sentences randomly selected from different documents). The model achieves 99% accuracy on this task but fails to perform well on tasks requiring nuanced language understanding. Based on the training setup, what is the most probable reason for this discrepancy? Explain how the model likely achieved its high accuracy without developing a deep understanding of language.
0
1
Tags
Ch.1 Pre-training - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Analysis of a Language Model's Training Objective
A language model is being trained to determine if two sentences are consecutive. For 'positive' examples, it is given two sentences that appear one after the other in a book. For 'negative' examples, the first sentence is from a book about astrophysics, and the second is always from a children's fairy tale. What is the most significant risk associated with this training design?
Evaluating a Model's Training Task