Search Space Growth Analysis
A language model with a vocabulary of 50,000 tokens is generating a sequence. Explain why generating the 10th token in a sequence is computationally a much larger problem than generating the 2nd token, even though both involve selecting just one token. Base your explanation on the growth of the total number of possible sequences.
0
1
Tags
Ch.5 Inference - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
A language model with a vocabulary of 5,000 unique tokens is generating text. If, at a certain stage, the model is considering 20 different unique sequences of tokens, each 4 tokens long, how many total unique sequences of 5 tokens can possibly be generated in the very next step?
Search Space Growth Analysis
Comparing Computational Load in Language Models