1. Radford, A., Narasimhan, et al.: Improving language understanding by generative pre-training. OpenAI (2018)
2. Devlin, J., Chang, M-W., et al.: Bert: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of NAACL-HLT, pp. 4171–4186 (2019)
3. Alsentzer, E., Murphy, J., et al.: Publicly available clinical BERT embeddings. In: Proceedings of the 2nd Clinical Natural Language Processing Workshop, pp. 72–78. Association for Computational Linguistics (2019)
4. Yang, F.: Research on BERT Model for Chinese Clinical Natural Language Processing. Peking Union Medical College (2021)
5. Brown, T.B., et al: Language models are few-shot learners. In: Proceedings of the Annual Conference on Neural Information Processing Systems, pp. 18741–18755 (2020)