Multiple Choice

A batch of four text sequences is being prepared for processing by a language model. The lengths of the sequences are 25, 28, 30, and 60 tokens. To process them together, all sequences must be extended to the length of the longest one by adding non-informative 'padding' tokens. What percentage of the total tokens in the final prepared batch consists of these non-informative padding tokens?

0

1

Updated 2025-10-07

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Computing Sciences

Foundations of Large Language Models Course

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science