Abstract
AbstractQuestioning plays a vital role in education, directing knowledge construction and assessing students’ understanding. However, creating high-level questions requires significant creativity and effort. Automatic question generation is expected to facilitate the generation of not only fluent and relevant but also educationally valuable questions. While rule-based methods are intuitive for short inputs, they struggle with longer and more complex inputs. Neural question generation (NQG) has shown better results in this regard. This review summarizes the advancements in NQG between 2016 and early 2022. The focus is on the development of NQG for educational purposes, including challenges and research opportunities. We found that although NQG can generate fluent and relevant factoid-type questions, few studies focus on education. Specifically, there is limited literature using context in the form of multi-paragraphs, which due to the input limitation of the current deep learning techniques, require key content identification. The desirable key content should be important to specific topics or learning objectives and be able to generate certain types of questions. A further research opportunity is controllable NQG systems, which can be customized by taking into account factors like difficulty level, desired answer type, and other individualized needs. Equally important, the results of our review also suggest that it is necessary to create datasets specific to the question generation tasks with annotations that support better learning for neural-based methods.
Funder
Kementerian Riset, Teknologi dan Pendidikan Tinggi
Publisher
Springer Science and Business Media LLC
Subject
Computational Theory and Mathematics,Education
Reference124 articles.
1. Abu-Salih, B. (2020). Domain-specific knowledge graphs: A survey. arXiv:2011.00235
2. Ali, H., Chali, Y., Hasan, S.A. (2010). Automatic question generation from sentences. In Actes de la 17e conférence sur le traitement automatique des langues naturelles. articles courts (pp. 213–218). ATALA
3. Amidei, J., Piwek, P., Willis, A. (2018). Evaluation methodologies in automatic question generation 2013-2018. In Proceedings of the 11th international natural language generation conference (pp. 307-317). Association for Computational Linguistics
4. Bai, Y., Wang, D.Z. (2021). More than reading comprehension: a survey on datasets and metrics of textual question answering. arXiv:2109.12264 [cs.CL]
5. Bajaj, P., Campos, D., Craswell, N., Deng, L., Gao, J., Liu, X., et al. (2016). MS MARCO: a human generated MAchine reading comprehension dataset. arXiv:1611.09268