Affiliation:
1. Yale University, USA. linyong.nan@yale.edu
2. Yale University, USA
3. Yale University, USA. ziming.mao@yale.edu
4. Salesforce Research, USA
5. Penn State University, USA
6. Archbishop Mitty High School, USA
Abstract
AbstractExisting table question answering datasets contain abundant factual questions that primarily evaluate a QA system’s comprehension of query and tabular data. However, restricted by their short-form answers, these datasets fail to include question–answer interactions that represent more advanced and naturally occurring information needs: questions that ask for reasoning and integration of information pieces retrieved from a structured knowledge source. To complement the existing datasets and to reveal the challenging nature of the table-based question answering task, we introduce FeTaQA, a new dataset with 10K Wikipedia-based {table, question, free-form answer, supporting table cells} pairs. FeTaQA is collected from noteworthy descriptions of Wikipedia tables that contain information people tend to seek; generation of these descriptions requires advanced processing that humans perform on a daily basis: Understand the question and table, retrieve, integrate, infer, and conduct text planning and surface realization to generate an answer. We provide two benchmark methods for the proposed task: a pipeline method based on semantic parsing-based QA systems and an end-to-end method based on large pretrained text generation models, and show that FeTaQA poses a challenge for both methods.
Subject
Artificial Intelligence,Computer Science Applications,Linguistics and Language,Human-Computer Interaction,Communication
Reference50 articles.
1. With few eyes, all hoaxes are deep;Asthana;Proceedings of the ACM on Human Computer Interaction,2018
2. METEOR: An automatic metric for MT evaluation with improved correlation with human judgments;Banerjee,2005
3. Question answering from frequently asked question files: Experiences with the faq finder system;Burke;AI Magazine,1997
4. Reading Wikipedia to answer open-domain questions;Chen,2017
5. Open question answering over tables and text;Chen,2021
Cited by
16 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献