Learn Before
Concept

Model Compression and Acceleration Method Categories

  • Parameter pruning and sharing: Focus on removing insignificant parameters from deep neural networks. Sub-categories are model quantization, model binarization, structural matrices, and parameter sharing.
  • Low-rank factorization: Use matrix and tensor decomposition to identify redundant parameters
  • Transferred compact convolutional filters: Transfer or compress convolutional filters to remove unnecessary parameters
  • Knowledge Distillation (KD): Filter knowledge from a deep neural network into a smaller network

0

1

Updated 2022-10-21

Tags

Deep Learning (in Machine learning)

Data Science