Learn Before
Contrasting Learning Target Methodologies
Consider two distinct model training scenarios:
Scenario A: A model is trained to identify objects in images. For each image, the training process uses a pre-defined, human-verified label (e.g., 'cat', 'dog') as the correct answer to guide learning.
Scenario B: A model is trained to generate plausible text. For a given input, the model first calculates which potential output sequence it currently considers to be the most likely. This most likely sequence is then treated as the 'correct' answer for the purpose of updating the model's parameters.
Based on these descriptions, contrast the fundamental difference in how the learning target (the 'correct' answer) is determined in Scenario A versus Scenario B.
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Log-Probability Loss with Model-Generated Target
A research team is training a generative model using a method where the learning target for any given input is the output that the model itself currently calculates as having the highest probability. This self-generated target is then used to update the model's parameters. Which statement best analyzes a key implication of this training approach?
Self-Reinforcing Training Strategy for a Chatbot
Contrasting Learning Target Methodologies