Learn Before
An engineer is training a language model where the training objective is to adjust the model's parameters to maximize a utility score for its generated outputs. The loss function is defined as the negative of the expected utility score. During a training run, the engineer observes that the calculated loss value is consistently increasing over several iterations (e.g., moving from -15.0 to -12.5 to -10.0). What is the most direct interpretation of this observation?
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
An engineer is training a language model where the training objective is to adjust the model's parameters to maximize a utility score for its generated outputs. The loss function is defined as the negative of the expected utility score. During a training run, the engineer observes that the calculated loss value is consistently increasing over several iterations (e.g., moving from -15.0 to -12.5 to -10.0). What is the most direct interpretation of this observation?
Rationale for the Negative Expected Utility Loss Function
Consequences of an Incorrect Loss Function Implementation