Learn Before
Concept
Distillation Loss
Distillation loss for feature-based learning transfer is
- and are the feature maps of the intermediate layers of the teacher and student models - and are applied when feature maps of the two models have different shape - is the similarity function for matching the models’ feature maps
0
1
Updated 2022-10-22
Tags
Deep Learning (in Machine learning)
Data Science