Testing the Effectiveness of the Diagnostic Probing Paradigm on Italian Treebanks
-
Published:2023-02-22
Issue:3
Volume:14
Page:144
-
ISSN:2078-2489
-
Container-title:Information
-
language:en
-
Short-container-title:Information
Author:
Miaschi Alessio1ORCID, Alzetta Chiara1ORCID, Brunato Dominique1ORCID, Dell’Orletta Felice1ORCID, Venturi Giulia1ORCID
Affiliation:
1. CNR—Institute for Computational Linguistics “A. Zampolli”, ItaliaNLPLab, Via G. Moruzzi 1, 56124 Pisa, Italy
Abstract
The outstanding performance recently reached by neural language models (NLMs) across many natural language processing (NLP) tasks has steered the debate towards understanding whether NLMs implicitly learn linguistic competence. Probes, i.e., supervised models trained using NLM representations to predict linguistic properties, are frequently adopted to investigate this issue. However, it is still questioned if probing classification tasks really enable such investigation or if they simply hint at surface patterns in the data. This work contributes to this debate by presenting an approach to assessing the effectiveness of a suite of probing tasks aimed at testing the linguistic knowledge implicitly encoded by one of the most prominent NLMs, BERT. To this aim, we compared the performance of probes when predicting gold and automatically altered values of a set of linguistic features. Our experiments were performed on Italian and were evaluated across BERT’s layers and for sentences with different lengths. As a general result, we observed higher performance in the prediction of gold values, thus suggesting that the probing model is sensitive to the distortion of feature values. However, our experiments also showed that the length of a sentence is a highly influential factor that is able to confound the probing model’s predictions.
Subject
Information Systems
Reference51 articles.
1. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, Ł., and Polosukhin, I. (2017). Attention is all you need. Adv. Neural Inf. Process. Syst., 6000–6010. Available online: https://dl.acm.org/doi/abs/10.5555/3295222.3295349. 2. Superglue: A stickier benchmark for general-purpose language understanding systems;Wang;Adv. Neural Inf. Process. Syst.,2019 3. Yang, W., Xie, Y., Lin, A., Li, X., Tan, L., Xiong, K., Li, M., and Lin, J. (2019, January 2–7). End-to-End Open-Domain Question Answering with BERTserini. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics (Demonstrations), Minneapolis, MN, USA. 4. Transformer based deep intelligent contextual embedding for twitter sentiment analysis;Naseem;Future Gener. Comput. Syst.,2020 5. Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., and Stoyanov, V. (2020). RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv.
|
|