Multiple Choice

A research team is developing a powerful new language model for summarizing scientific papers. Lacking a large, human-curated dataset of summaries, they use an older, less accurate model to generate summaries for 100,000 papers. They then fine-tune their powerful new model on this machine-generated dataset, with the goal of teaching it to produce summaries that match the ones from the older model. What is the most significant inherent risk in this training strategy?

0

1

Updated 2025-09-26

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science

Related