Case Study

Analyzing Model Behavior Under Maximum Likelihood Estimation

A development team is fine-tuning a language model on a dataset of instructions and ideal responses. The goal is for the model to be a direct and concise assistant. For an input like List the primary colors, the ideal response in the training data is Red, yellow, and blue. However, after training, the model consistently generates Of course! The primary colors are red, yellow, and blue.

Assuming the training process is correctly configured to maximize the probability of generating the ideal responses, analyze this situation. Explain why the model might learn to generate the conversational preamble (Of course! The...) despite it not being present in the ideal response for this specific example. Your explanation should focus on how the training objective functions on a token-by-token basis across the entire dataset.

0

1

Updated 2025-10-08

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science