Publisher
Springer Nature Switzerland
Reference27 articles.
1. Adelani, D., et al.: MasakhaNER 2.0: Africa-centric transfer learning for named entity recognition. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 4488–4508. Association for Computational Linguistics, Abu Dhabi (2022). https://doi.org/10.18653/v1/2022.emnlp-main.298
2. Ba, J.L., Kiros, J.R., Hinton, G.E.: Layer normalization. arXiv preprint arXiv:1607.06450 (2016). https://doi.org/10.48550/arXiv.1607.06450
3. Chi, E.A., Hewitt, J., Manning, C.D.: Finding universal grammatical relations in multilingual BERT. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 5564–5577. Association for Computational Linguistics, Online (2020). https://doi.org/10.18653/v1/2020.acl-main.493
4. Clark, K., Khandelwal, U., Levy, O., Manning, C.D.: What does BERT look at? An analysis of BERT’s attention. In: Linzen, T., Chrupała, G., Belinkov, Y., Hupkes, D. (eds.) Proceedings of the 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, pp. 276–286. Association for Computational Linguistics, Florence (2019). https://doi.org/10.18653/v1/W19-4828
5. Clark, K., Luong, M.T., Le, Q.V., Manning, C.D.: ELECTRA: pre-training text encoders as discriminators rather than generators. In: International Conference on Learning Representations. Online (2020). https://openreview.net/forum?id=r1xMH1BtvB