Short Answer

Mismatch Between Pre-training and Downstream Objectives

A research lab develops a large-scale language model by training it on a massive corpus of historical literature to become an expert at completing sentences in the style of 18th-century authors. A separate team wants to use this model, without any modifications, to power a modern-day customer service chatbot. Explain the fundamental reason why this approach is likely to fail, focusing on the model's internal configuration.

0

1

Updated 2025-10-06

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science