Learn Before
Consider a large, pre-trained language model being adapted for a specific task. During this adaptation process, a small sequence of new, trainable vectors is prepended to the input of each transformer layer. The original weights of the pre-trained model are not modified. The training objective is to minimize a task-specific loss by only updating the parameters of these newly introduced vector sequences. Which statement best analyzes how this adaptation method functions?
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Consider a large, pre-trained language model being adapted for a specific task. During this adaptation process, a small sequence of new, trainable vectors is prepended to the input of each transformer layer. The original weights of the pre-trained model are not modified. The training objective is to minimize a task-specific loss by only updating the parameters of these newly introduced vector sequences. Which statement best analyzes how this adaptation method functions?
Information Flow in a Multi-Layer Tuning Process