Short Answer

Positional Context in Concatenated Sequences

Consider two token sequences, x and y. A language model calculates two separate quantities: one by processing the sequence x followed by y (concatenated as [x, y]), and another by processing the sequence y by itself. Explain why the initial input vector for the first token of y will be different in these two scenarios. Your explanation should identify the specific component of the input vector that changes and the reason for this change.

0

1

Updated 2025-10-05

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science