1. Ofer D , Brandes N , Linial M (2021) The language of proteins: NLP, machine learning \& protein sequences. Comput Struct Biotechnol J
2. Devlin J , Chang M-W , Lee K , Toutanova K (2018) Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv Prepr arXiv 181004805
3. Vaswani A , Shazeer N , Parmar N , et al (2017) Attention is all you need. arXiv Prepr arXiv 170603762
4. Language models are unsupervised multitask learners;OpenAI blog,2019
5. Brown TB , Mann B , Ryder N , et al (2020) Language models are few-shot learners. arXiv Prepr arXiv 200514165