Multiple Choice

An autoregressive model calculates a square attention weight matrix using the formula: Softmax((QK^T / sqrt(d)) + Mask). The purpose of the Mask component is to prevent any token from attending to subsequent tokens in the sequence. Which statement best describes the resulting attention weight matrix?

0

1

Updated 2025-09-29

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science