Learn Before
Concept
Model Compression and Acceleration Method Categories
- Parameter pruning and sharing: Focus on removing insignificant parameters from deep neural networks. Sub-categories are model quantization, model binarization, structural matrices, and parameter sharing.
- Low-rank factorization: Use matrix and tensor decomposition to identify redundant parameters
- Transferred compact convolutional filters: Transfer or compress convolutional filters to remove unnecessary parameters
- Knowledge Distillation (KD): Filter knowledge from a deep neural network into a smaller network
0
1
Updated 2022-10-21
Tags
Deep Learning (in Machine learning)
Data Science