Multiple Choice

A large language model was originally trained with a maximum context window of 2048 tokens. You are now tasked with enabling it to process a sequence of 4096 tokens using a technique that scales the position indices of the longer sequence to fit within the model's original learned range. How should the position index for the token at position 3072 in the 4096-token sequence be handled before being passed to the embedding layer?

0

1

Updated 2025-10-03

Contributors are:

Who are from:

Tags

Ch.3 Prompting - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Application in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science