1. China Internet Network Information Center (CNNIC), The 50th Statistical Report on the Development of the Internet in China, 2022.
2. J. Li, T. Du, S. Ji, et al., TextShield: Robust Text Classification Based on Multimodal Embedding and Neural Machine Translation, in: 29th USENIX Security Symposium, USENIX Security 20, 2020, pp. 1381–1398.
3. Roberta: A robustly optimized BERT pretraining approach;Liu,2019
4. K. Lai, Y. Long, B. Wu, et al., Semorph: A Morphology Semantic Enhanced Pre-trained Model for Chinese Spam Text Detection, in: Proceedings of the 31st ACM International Conference on Information and Knowledge Management, 2022, pp. 1003–1013.
5. J. Devlin, M.W. Chang, K. Lee, et al., Bert: Pre-training of deep bidirectional transformers for language understanding, in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 2019, pp. 4171–4186.