When the supervised fine-tuning objective is written as , the parameters denoted by are typically initialized from a random distribution before the optimization process begins.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Computing Sciences
Foundations of Large Language Models Course
Comprehension in Revised Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
A researcher is fine-tuning a pre-trained language model on a new dataset. They represent the optimization objective using the following simplified notation:
Based on standard conventions in this field, what is the most accurate interpretation of the parameters
θbeing optimized in this formula?When the supervised fine-tuning objective is written as , the parameters denoted by are typically initialized from a random distribution before the optimization process begins.
Potential Misinterpretation of Fine-Tuning Notation