Multiple Choice

A machine learning team is training a new 10-billion-parameter language model on a novel, specialized dataset. They meticulously copy the exact training configuration (optimizer, learning rate schedule, parallelism strategy) from a famous research paper that successfully trained a model of a similar size. After several days, their training run becomes unstable and the model's performance collapses. What is the most probable explanation for this failure?

0

1

Updated 2025-09-26

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science