1. BERT: pre-training of deep bidirectional transformers for language understanding;Devlin,2019
2. Finetuned language models are zero-shot learners;Wei,2022
3. Parameter-efficient transfer learning for nlp;Houlsby,2019
4. LoRA: low-rank adaptation of large language models;Hu,2021
5. Differentially private fine-tuning of language models;Yu,2022