TransConv: Transformer Meets Contextual Convolution for Unsupervised Domain Adaptation
Author:
Liu Junchi1ORCID, Zhang Xiang1, Luo Zhigang1
Affiliation:
1. School of Computer Science, National University of Defense Technology, Changsha 410073, China
Abstract
Unsupervised domain adaptation (UDA) aims to reapply the classifier to be ever-trained on a labeled source domain to a related unlabeled target domain. Recent progress in this line has evolved with the advance of network architectures from convolutional neural networks (CNNs) to transformers or both hybrids. However, this advance has to pay the cost of high computational overheads or complex training processes. In this paper, we propose an efficient alternative hybrid architecture by marrying transformer to contextual convolution (TransConv) to solve UDA tasks. Different from previous transformer based UDA architectures, TransConv has two special aspects: (1) reviving the multilayer perception (MLP) of transformer encoders with Gaussian channel attention fusion for robustness, and (2) mixing contextual features to highly efficient dynamic convolutions for cross-domain interaction. As a result, TransConv enables to calibrate interdomain feature semantics from the global features and the local ones. Experimental results on five benchmarks show that TransConv attains remarkable results with high efficiency as compared to the existing UDA methods.
Reference55 articles.
1. He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27–30). Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA. 2. Deep visual domain adaptation: A survey;Wang;Neurocomputing,2018 3. Kuroki, S., Charoenphakdee, N., Bao, H., Honda, J., Sato, I., and Sugiyama, M. (February, January 27). Unsupervised domain adaptation based on source-guided discrepancy. Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA. 4. Xu, T., Chen, W., Wang, P., Wang, F., Li, H., and Jin, R. (2021). Cdtrans: Cross-domain transformer for unsupervised domain adaptation. arXiv. 5. Yang, J., Liu, J., Xu, N., and Huang, J. (2023, January 2–7). Tvt: Transferable vision transformer for unsupervised domain adaptation. Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA.
|
|