Learn Before
Theoretical Loss Limit with Infinite Data
Consider the empirical formula for the Chinchilla scaling law: If a research team had access to a dataset of infinite size (i.e., as D approaches infinity), what would be the theoretical minimum loss for a model of a fixed size N? Explain how you arrived at your answer by analyzing the components of the formula.
0
1
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Optimizing Training Resources
Theoretical Loss Limit with Infinite Data
A research team is using the following empirical formula to guide their training strategy for a large language model, where
Lis the test loss,Nis the model size, andDis the dataset size: To achieve the most substantial reduction in test loss, which of the following strategies is predicted by this formula to be more effective?