Learn Before
A machine learning engineer needs to deploy a large Transformer-based language model on a device with very limited memory. The primary objective is to significantly reduce the model's file size on disk. Which of the following strategies directly achieves this by changing the numerical precision of the model's parameters?
0
1
Tags
Ch.1 Pre-training - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Application in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Evaluating a Model Compression Strategy
A machine learning engineer needs to deploy a large Transformer-based language model on a device with very limited memory. The primary objective is to significantly reduce the model's file size on disk. Which of the following strategies directly achieves this by changing the numerical precision of the model's parameters?
A development team successfully reduces the size of a large Transformer-based language model by converting its 32-bit floating-point parameters into 8-bit integers. What is the primary trade-off they must evaluate to ensure the compressed model is still effective for its intended task?