Learn Before
Analysis of Text Generation Combination Methods
A team is developing a system to generate a single, high-quality sentence by leveraging two different language models. They are considering two approaches:
- Approach 1: Generate a complete sentence from Model A and a complete sentence from Model B. Then, analyze these two final sentences to create a single, blended output.
- Approach 2: To decide on the very first word, get the probability scores for all possible words from both Model A and Model B, average these scores, and pick the word with the highest average. Repeat this step-by-step process for each subsequent word until a full sentence is formed.
Based on these descriptions, what is the fundamental difference in what is being combined and when the combination occurs in Approach 2 compared to Approach 1?
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Formula for Token-Level Model Averaging in Prompt Ensembling
Imagine two language models are tasked with completing the sentence: 'The weather today is exceptionally...'. At this specific step, they must choose the very next word. Their internal calculations produce the following probability scores for the top three candidate words:
- Model 1:
warm(0.6),sunny(0.3),bright(0.1) - Model 2:
warm(0.2),sunny(0.7),bright(0.1)
If a system combines these models by averaging their token-level probability distributions to make a decision, which word will it select as the next word in the sequence, and why?
- Model 1:
Analysis of Text Generation Combination Methods
Choosing a Generation Combination Strategy