Learn Before
Concept
Offline Distillation
Offline methods assume the teacher knowledge/model is already defined, and focus on improving knowledge transfer, such as design of knowledge and loss functions for feature or distribution matching. They usually have one-way knowledge transfer and two-phase training procedure.
0
1
Updated 2022-10-29
Tags
Deep Learning (in Machine learning)
Data Science