Debugging a Text Generation Process
A developer is building a text-generation model. They provide the input sequence of tokens ['The', 'sky', 'is']. The model's first generated token is ['blue']. For the next step, to generate the token that follows 'blue', the developer's code incorrectly uses only the token ['blue'] as the new input for the model. Explain why this approach is flawed and describe the correct input sequence the model should process.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
A language model is given the input 'The cat sat on' and begins to generate the next part of the sentence. The model's first generated word is 'the'. To predict the second word of its output, what sequence of words does the model use as its new input for the next processing step?
A text-generating model is given the input sequence of tokens:
['The', 'quick', 'brown', 'fox']. It then generates the output sequence:['jumps', 'over', 'the']. Arrange the following processing steps in the correct chronological order that the model would take to produce this output.Debugging a Text Generation Process