Learn Before
Concept

Offline Distillation

Offline methods assume the teacher knowledge/model is already defined, and focus on improving knowledge transfer, such as design of knowledge and loss functions for feature or distribution matching. They usually have one-way knowledge transfer and two-phase training procedure.

0

1

Updated 2022-10-29

Tags

Deep Learning (in Machine learning)

Data Science

Related
Learn After