Multiple Choice

A data scientist trains two models, Model X and Model Y, on the same dataset for the same task. The training objective for each is to find the set of parameters, θ, that minimizes a loss function, L(θ), according to the principle: θ~=argminθL(θ)\tilde{\theta} = \arg \min_{\theta} \mathcal{L}(\theta) After training, the results are as follows:

  • For Model X, the lowest achieved loss is 50, using parameters θ_X.
  • For Model Y, the lowest achieved loss is 100, using parameters θ_Y.

Based only on this information and the definition of the training objective, what is the most valid conclusion?

0

1

Updated 2025-10-04

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science