1. BERT: Pre-training of deep bidirectional transformers for language understanding;Devlin
2. Roberta: A robustly optimized bert pretraining approach;Liu,2019
3. LLaMA: Open and efficient foundation language models;Touvron,2023
4. How robust is gpt-3.5 to predecessors? a comprehensive study on language understanding tasks;Chen,2023
5. On the robustness of chatgpt: An adversarial and out-of-distribution perspective;Wang,2023