1. J. Devlin , M.-W. Chang , K. Lee , and K. Toutanova , ‘BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding’, ArXiv181004805 Cs, May 2019, Accessed: Jan. 26, 2021. [Online]. Available: http://arxiv.org/abs/1810.04805
2. R. Rao et al., ‘Evaluating Protein Transfer Learning with TAPE’, in Advances in Neural Information Processing Systems, Curran Associates, Inc., 2019. Accessed: Apr. 05, 2023. [Online]. Available: https://proceedings.neurips.cc/paper/2019/hash/37f65c068b7723cd7809ee2d31d7861c-Absract.htmlt
3. ProtTrans: Towards Cracking the Language of Lifes Code Through Self-Supervised Deep Learning and High Performance Computing
4. ProteinBERT: a universal deep-learning model of protein sequence and function
5. BERTMHC: improved MHC–peptide class II interaction prediction with transformer and multiple instance learning