Learn Before
Relation
Memory Efficient Models
Most modern neural language models require a significant amount of memory for training and inference. These models have to be compressed in order to meet the computation and storage constraints of edge applications. This can be done either by building student models using knowledge distillation or by using model compression techniques. Developing a task-agnostic model compression method is an active research topic.
0
1
Updated 2025-09-16
Tags
Data Science