Multiple Choice

A research lab has developed a very large language model that was pre-trained on a vast and diverse dataset from the internet. The lab now wants to adapt this model to be a helpful assistant that follows specific user commands, but they have a very limited budget for creating new training data. Based on the relationship between extensive pre-training and model adaptation, which of the following approaches is the most logical and resource-efficient for the lab to pursue?

0

1

Updated 2025-10-02

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science