Learn Before
A language model is configured to generate text by sampling from the smallest set of tokens whose cumulative probability exceeds a predefined threshold 'p'. Arrange the following steps of this process in the correct chronological order.
0
1
Tags
Ch.5 Inference - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Comprehension in Revised Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Candidate Pool Size in Top-p Sampling (kp)
Forming the Candidate Pool in Top-p Sampling
A language model is generating text and has calculated the following probabilities for the next possible token: 'the' (0.45), 'a' (0.25), 'one' (0.15), 'it' (0.10), 'she' (0.05). If the model uses a sampling strategy with a probability threshold of
p = 0.8, which set of tokens will form the final candidate pool (the 'nucleus') from which the next token is actually sampled?A language model is configured to generate text by sampling from the smallest set of tokens whose cumulative probability exceeds a predefined threshold 'p'. Arrange the following steps of this process in the correct chronological order.
Applying the Top-p Sampling Process