Learn Before
Multiple Choice

A language model needs to process a group of sentences simultaneously. For computational efficiency, all input sequences in the group must be the same length. This is achieved by adding a special, non-word token to the end of any shorter sequences. Given the two tokenized sentences below, which option correctly demonstrates this preparation process?

Sentence A: ['The', 'quick', 'fox'] (length 3) Sentence B: ['A', 'lazy', 'dog', 'sleeps'] (length 4)

0

1

Updated 2025-09-28

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Application in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science