Concept

Limitation of Pre-trained LLMs: Next-Token Prediction vs. Instruction Following

Pre-trained Large Language Models are fundamentally optimized for next-token prediction, meaning their primary training objective is to predict the next word in a sequence. This training method does not inherently equip them to follow explicit instructions. Consequently, when presented with a task-oriented prompt, such as a request for summarization, a pre-trained LLM is likely to simply continue the input text rather than executing the given command.

0

1

Updated 2025-10-06

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences