MKTN: Adversarial-Based Multifarious Knowledge Transfer Network from Complementary Teachers
-
Published:2024-03-28
Issue:1
Volume:17
Page:
-
ISSN:1875-6883
-
Container-title:International Journal of Computational Intelligence Systems
-
language:en
-
Short-container-title:Int J Comput Intell Syst
Author:
Zhang XiaobingORCID, Chang Heyu, Hao Yaohui, Chang Dexian
Abstract
AbstractWith the demands for light deep networks models in various front-end devices, network compression has attracted increasing interest for reducing model sizes yet without sacrificing much model accuracy. This paper presents a multifarious knowledge transfer network (MKTN) that aims to produce a compact yet powerful student network from two complementary teacher networks. Instead of learning homogeneous features, the idea is to pre-train one teacher to capture generative and low-level image features under a reconstruction objective, and another teacher to capture discriminative and task-specific features under the same objective as the student network. During knowledge transfer, the student learns multifarious and complementary knowledge from the two teacher networks under the guidance of the proposed adversarial loss and feature loss respectively. Experimental results indicate that the proposed training losses can effectively guide the student to learn spatial-level and pixel-level information as distilled from teacher networks. On the other hand, our study over a number of widely used datasets shows that transferring multifarious features from complementary teachers equipped with different types of knowledge helps to teach a compact yet powerful student effectively.
Funder
National Social Science Fund Project
Publisher
Springer Science and Business Media LLC
Reference59 articles.
1. Tsiakmaki, M., Kostopoulos, G., Kotsiantis, S., Ragos, O.: Transfer learning from deep neural networks for predicting student performance. Appl. Sci. 10(6), 2145 (2020) 2. Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., Rabinovich, A.: Going Deeper with Convolutions. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, USA, Jun 7–12, pp. 1–9 (2015) 3. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, USA, Jun 26–Jul 1, pp. 770–778 (2016) 4. Zhang, X., Gong, H., Dai, X., Yang, F., Liu, N., Liu, M.: Understanding pictograph with facial features: end-to-end sentence-level lip reading of Chinese. In: Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, USA, Jan 2–Feb 1, pp. 9211–9218 (2019) 5. Madan, P., Singh, V., Chaudhari, V., Albagory, Y., Dumka, A., Singh, R., Gehlot, A., Rashid, M., Alshamrani, S.S., AlGhamdi, A.S.: An optimization-based diabetes prediction model using CNN and bi-directional LSTM in real-time environment. Appl. Sci. 12(8), 3989 (2022)
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|