Knowledge Distillation in Image Classification: The Impact of Datasets
-
Published:2024-07-24
Issue:8
Volume:13
Page:184
-
ISSN:2073-431X
-
Container-title:Computers
-
language:en
-
Short-container-title:Computers
Author:
Belinga Ange Gabriel1ORCID, Tekouabou Koumetio Cédric Stéphane23ORCID, El Haziti Mohamed14, El Hassouni Mohammed5
Affiliation:
1. Laboratory of Research in Computer Science and Telecommunications (LRIT), Faculty of Sciences in Rabat, Mohammed V University in Rabat, Rabat 10000, Morocco 2. Laboratory in Computer Science and Educational Technologies (LITE), Higher Teacher Training College (HTTC), University of Yaoundé 1, Yaounde P.O. Box 47, Cameroon 3. Department of Computer Science and Educational Technologies (DITE), Higher Teacher Training College (HTTC), University of Yaoundé 1, Yaounde P.O. Box 47, Cameroon 4. High School of Technology, Mohammed V University in Rabat, Sale 11000, Morocco 5. FLSH, Mohammed V University in Rabat, 3 Av. Ibn Batouta, Rabat 10090, Morocco
Abstract
As the demand for efficient and lightweight models in image classification grows, knowledge distillation has emerged as a promising technique to transfer expertise from complex teacher models to simpler student models. However, the efficacy of knowledge distillation is intricately linked to the choice of datasets used during training. Datasets are pivotal in shaping a model’s learning process, influencing its ability to generalize and discriminate between diverse patterns. While considerable research has independently explored knowledge distillation and image classification, a comprehensive understanding of how different datasets impact knowledge distillation remains a critical gap. This study systematically investigates the impact of diverse datasets on knowledge distillation in image classification. By varying dataset characteristics such as size, domain specificity, and inherent biases, we aim to unravel the nuanced relationship between datasets and the efficacy of knowledge transfer. Our experiments employ a range of datasets to comprehensively explore their impact on the performance gains achieved through knowledge distillation. This study contributes valuable guidance for researchers and practitioners seeking to optimize image classification models through kno-featured applications. By elucidating the intricate interplay between dataset characteristics and knowledge distillation outcomes, our findings empower the community to make informed decisions when selecting datasets, ultimately advancing the field toward more robust and efficient model development.
Reference49 articles.
1. Chen, G., Choi, W., Yu, X., Han, T., and Chandraker, M. (2017, January 4–9). Learning efficient object detection models with knowledge distillation. Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA. 2. Feng, Y., Wang, H., Hu, H.R., Yu, L., Wang, W., and Wang, S. (2020, January 25–28). Triplet distillation for deep face recognition. Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Virtual. 3. Jiao, X., Yin, Y., Shang, L., Jiang, X., Chen, X., Li, L., Wang, F., and Liu, Q. (2019). Tinybert: Distilling bert for natural language understanding. arXiv. 4. Wang, H., Li, Y., Wang, Y., Hu, H., and Yang, M.H. (2020, January 13–19). Collaborative distillation for ultra-resolution universal style transfer. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA. 5. Krizhevsky, A., and Hinton, G. (2009). Learning Multiple Layers of Features from Tiny Images. [Master’s Thesis, Department of Computer Science, University of Toronto]. Available online: https://www.cs.toronto.edu/~kriz/learningfeatures-2009-TR.pdf.
|
|