1. Ao J Y,Wang R,Zhou L,Wang C Y,Ren S,Wu Y,Liu S J,Ko T,Li Q,Zhang Y,Wei Z H,Qian Y,Li J Y and Wei F R.2022.Speecht5:unified-modal encoder-decoder pre-training for spoken language processing[EB/OL].[2023-07-25].https://arxiv.org/pdf/2110.07205.pdf
2. Barradas D,Santos N,Rodrigues L and Nunes V.2020.Poking a hole in the wall:efficient censorship-resistant internet communications by parasitizing on WebRTC//Proceedings of 2020 ACM SIGSACConference on Computer and Communications Security.Virtual Event,USA:ACM:35-48[DOI:10.1145/3372297.3417874]
3. Brown T B,Mann B,Ryder N,Subbiah M,Kaplan J,Dhariwal P,Neelakantan A,Shyam P,Sastry G,Askell A,Agarwal S,Herbert-Voss A,Krueger G,Henighan T,Child R,Ramesh A,Ziegler D M,Wu J,Winter C,Hesse C,Chen M,Sigler E,Litwin M,Gray S,Chess B,Clark J,Berner C,McCandlish S,Radford A,Sutskever I and Amodei D.2020.Language models are fewshot learners//Proceedings of the 34th International Conference on Neural Information Processing Systems.Vancouver,Canada:Curran Associates Inc.:1877-1901
4. Chen K J,Zhou H,Zhao H Q,Chen D D,Zhang W M and Yu N H.2022.Distribution-preserving steganography based on text-tospeech generative models.IEEE Transactions on Dependable and Secure Computing,19(5):3343-3356[DOI:10.1109/TDSC.2021.3095072]
5. Devlin J,Chang M W,Lee K and Toutanova K.2019.Bert:pretraining of deep bidirectional transformers for language understanding[EB/OL].[2023-07-25].https://arxiv.org/pdf1810.04805.pdf