Learn Before
Multiple Choice

A research team is tasked with deploying a large language model on edge devices with limited memory and processing power. Their primary goal is to reduce the model's memory footprint. They achieve this by converting the model's 32-bit floating-point weights and activations into 8-bit integers. While this significantly reduces the model's size, they observe a minor drop in performance. Which model compression technique does this scenario describe?

0

1

Updated 2025-10-01

Contributors are:

Who are from:

Tags

Data Science

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Application in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science