1. G. Hinton, O. Vinyals, J. Dean, Distilling the knowledge in a neural network, in: NeurIPS Workshop on Deep Learning and Representation Learning, 2015.
2. Learning what and where to transfer;Jang,2019
3. S.I. Mirzadeh, M. Farajtabar, A. Li, N. Levine, A. Matsukawa, H. Ghasemzadeh, Improved knowledge distillation via teacher assistant, in: Proc. AAAI, Vol. 34, 2020, pp. 5191–5198.
4. L. Yuan, F.E. Tay, G. Li, T. Wang, J. Feng, Revisiting Knowledge Distillation via Label Smoothing Regularization, in: Proc. CVPR, 2020, pp. 3903–3911.
5. K. Kim, B. Ji, D. Yoon, S. Hwang, Self-Knowledge Distillation With Progressive Refinement of Targets, in: Proc. ICCV, 2021, pp. 6567–6576.