Affiliation:
1. Vietnam National University, Hanoi
2. University of Brussels (VUB)
Abstract
Abstract
Automated tools for syntactic complexity measurement are increasingly used for analyzing various kinds of second
language corpora, even though these tools were originally developed and tested for texts produced by advanced learners. This study
investigates the reliability of automated complexity measurement for beginner and lower-intermediate L2 English data by comparing
manual and automated analyses of a corpus of 80 texts written by Dutch-speaking learners. Our quantitative and qualitative
analyses reveal that the reliability of automated complexity measurement is substantially affected by learner errors, parser
errors, and Tregex pattern undergeneration. We also demonstrate the importance of aligning the definitions of
analytical units between the computational tool and human annotators. In order to enhance the reliability of automated analyses,
it is recommended that certain modifications are made to the system, and non-advanced L2 English data are preprocessed prior to
automated analyses.
Publisher
John Benjamins Publishing Company
Subject
Linguistics and Language,Language and Linguistics
Cited by
6 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献