Evaluating an Alternative Fine-Tuning Objective
Based on the provided scenario, critique Objective B and explain the mathematical and practical reasons why Objective A is the preferred standard for supervised fine-tuning.
0
1
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Ch.4 Alignment - Foundations of Large Language Models
Evaluation in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Notational Simplification in Fine-Tuning Formulas
A language model is being fine-tuned on a dataset of customer support chat logs to improve its ability to generate helpful responses. The training process is guided by the objective function: During one step of this process, the model processes a single
(query, response)pair from the dataset. What is the role of the specific componentlog Pr(response|query)for this single pair?The following equation represents the primary goal of a common model training process. Match each mathematical symbol from the equation to its correct description.
Evaluating an Alternative Fine-Tuning Objective
Token-Level Conditional Log-Probability in Supervised Fine-Tuning