Learn Before
Multiple Choice

In a sequence-processing model using an attention mechanism, the model needs to determine which words in an input sentence are most relevant to the current word it is processing. If the 'Key' vectors associated with every word in the input sentence were made identical to each other, what would be the most direct consequence for the attention calculation?

0

1

Updated 2025-09-26

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science