Learn Before
Separating Input and Output Variables in LLM Formulation
Although the input and output tokens of a Large Language Model can technically be viewed as sub-sequences of a single, continuous sequence, it is common practice to employ separate variables—typically for the input and for the output. Adopting this distinct notation helps to clearly separate the given context from the generated text, resulting in mathematical formulations that closely resemble those used in other natural language processing text generation models, such as neural machine translation.
0
1
Tags
Foundations of Large Language Models
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Formal Definition of LLM Inference
Notation for Preceding Output Subsequence
Deconstructing a Model's Generated Text
Representing Model Output as a Token Sequence
A Large Language Model generates the sentence: 'AI is transforming our world.' How is this output fundamentally structured by the model before being presented to the user?
Separating Input and Output Variables in LLM Formulation