Restructuring the Teacher and Student in Self-Distillation
Knowledge distillation aims to achieve model compression by transferring knowledge from complex teacher models to lightweight student models. To reduce reliance on pre-trained teacher models, self-distillation methods utilize knowledge from the model itself as additional supervision. However, their...
Ausführliche Beschreibung
Bibliographische Detailangaben
| Veröffentlicht in: | IEEE transactions on image processing : a publication of the IEEE Signal Processing Society. - 1992. - 33(2024) vom: 24., Seite 5551-5563
|
| 1. Verfasser: |
Zheng, Yujie
(VerfasserIn) |
| Weitere Verfasser: |
Wang, Chong,
Tao, Chenchen,
Lin, Sunqi,
Qian, Jiangbo,
Wu, Jiafei |
| Format: | Online-Aufsatz
|
| Sprache: | English |
| Veröffentlicht: |
2024
|
| Zugriff auf das übergeordnete Werk: | IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
|
| Schlagworte: | Journal Article |