Author:
Wang Yu,Liu Bo,Zhou Fugen
Publisher
Springer Nature Switzerland
Reference31 articles.
1. Dosovitskiy, A., Beyer, L., Kolesnikov, A., et al.: An image is worth 16 x 16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, (2020)
2. Hatamizadeh, A, Tang, Y., Nath, V., et al.: Unetr: transformers for 3d medical image segmentation. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 574–584 (2022)
3. Tang, Y, Yang, D., Li, W., et al.: Self-supervised pre-training of swin transformers for 3d medical image analysis. arXiv preprint arXiv:2111.14791 (2021)
4. Bao, H., Dong, L., Wei, F.: Beit: BERT pre-training of image transformers. arXiv preprint arXiv:2106.08254 (2021)
5. He, K., Chen, X., Xie, S., et al.: Masked autoencoders are scalable vision learners. arXiv preprint arXiv:2111.06377 (2021)