Objective Function for Fine-Tuning a Strong LLM with Weak Supervision
The process of fine-tuning a strong Large Language Model using synthetic data generated by a weak model can be mathematically formalized. Given a collection of inputs , where each input includes an instruction and any necessary user input, a weak LLM denoted by generates a prediction . The strong LLM, denoted by , is then trained on these predictions. The objective is to find the optimal model parameters that maximize the log-probability of the weak model's generated predictions: .

0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Objective Function for Fine-Tuning a Strong LLM with Weak Supervision
A research team is developing a powerful new language model for summarizing scientific papers. Lacking a large, human-curated dataset of summaries, they use an older, less accurate model to generate summaries for 100,000 papers. They then fine-tune their powerful new model on this machine-generated dataset, with the goal of teaching it to produce summaries that match the ones from the older model. What is the most significant inherent risk in this training strategy?
Training Strategy for a Legal AI
Visual Diagram of Weak-to-Strong Generalization via Data Selection
A team is implementing a strategy where a powerful language model learns from a less capable one. Arrange the following steps into the correct chronological order to describe this process.
Your company is rolling out an instruction-tuned L...
You lead an LLM enablement team building an instru...
You’re leading an LLM platform team building an in...
Your company is building an internal IT helpdesk a...
Deciding Whether (and How) to Use Weak-Model Synthetic Data for Instruction Fine-Tuning
Diagnosing and Fixing a Synthetic Instruction-Tuning Data Flywheel That Degrades Model Behavior
Designing a Synthetic Instruction Fine-Tuning Pipeline Under Budget and Quality Constraints
Stabilizing an Instruction-Tuned Support Assistant When Synthetic Data Conflicts with Human Policy
Selecting and Filtering Self-Generated Instruction Data When Bootstrapping a Strong Model from a Weak Supervisor
Choosing a Weak-Model + Self-Instruct Data Strategy for Instruction Fine-Tuning Without Regressions
Learn After
Weak-to-Strong Fine-Tuning as a Knowledge Distillation Problem
A research team is adapting a large, powerful language model (the 'strong model') for a specialized task. They lack a large set of human-verified labels, but they have a smaller, less accurate model (the 'weak model') that can generate plausible, albeit imperfect, labels. The team's strategy is to use the weak model to label a large unlabeled dataset and then fine-tune the strong model to mimic the weak model's labeling behavior on this dataset. Which of the following mathematical objectives best represents the goal of finding the optimal strong model parameters, , that maximize the strong model's ability to predict the labels, , generated by the weak model for a given set of inputs, ?
Analyzing Overfitting in Weak-to-Strong Fine-Tuning
Deconstructing the Weak-to-Strong Fine-Tuning Objective