1. Hinton, G.E., Vinyals, O., and Dean, J. (2015). Distilling the Knowledge in a Neural Network. arXiv.
2. Tian, Y., Krishnan, D., and Isola, P. (2022). Contrastive Representation Distillation. arXiv.
3. Tung, F., and Mori, G. (November, January 27). Similarity-Preserving Knowledge Distillation. Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea.
4. LeCun, Y., Denker, J., and Solla, S. (1989). Proceedings of the Advances in Neural Information Processing Systems, Morgan-Kaufmann.
5. Zagoruyko, S., and Komodakis, N. (2017). Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer. arXiv.