Learn Before
Concept
Model Compression
Its crucial that the time and memory cost of running inference in a ML model is lower than the time and memory cost of training. Model compression is a strategy for reducing the cost of inference. Its main idea is to replace the original, expensive model with a smaller model that requires less memory and runtime to store and evaluate. This strategy can be used when size of the original model needs to prevent overfitting.
0
0
Updated 2021-07-01
Tags
Data Science