FeTaQA: Free-form Table Question Answering

Author:

Nan Linyong1,Hsieh Chiachun2,Mao Ziming3,Lin Xi Victoria4,Verma Neha2,Zhang Rui5,Kryściński Wojciech4,Schoelkopf Hailey2,Kong Riley6,Tang Xiangru2,Mutuma Mutethia1,Rosand Ben2,Trindade Isabel2,Bandaru Renusree5,Cunningham Jacob5,Xiong Caiming4,Radev Dragomir2,Radev Dragomir4

Affiliation:

1. Yale University, USA. linyong.nan@yale.edu

2. Yale University, USA

3. Yale University, USA. ziming.mao@yale.edu

4. Salesforce Research, USA

5. Penn State University, USA

6. Archbishop Mitty High School, USA

Abstract

AbstractExisting table question answering datasets contain abundant factual questions that primarily evaluate a QA system’s comprehension of query and tabular data. However, restricted by their short-form answers, these datasets fail to include question–answer interactions that represent more advanced and naturally occurring information needs: questions that ask for reasoning and integration of information pieces retrieved from a structured knowledge source. To complement the existing datasets and to reveal the challenging nature of the table-based question answering task, we introduce FeTaQA, a new dataset with 10K Wikipedia-based {table, question, free-form answer, supporting table cells} pairs. FeTaQA is collected from noteworthy descriptions of Wikipedia tables that contain information people tend to seek; generation of these descriptions requires advanced processing that humans perform on a daily basis: Understand the question and table, retrieve, integrate, infer, and conduct text planning and surface realization to generate an answer. We provide two benchmark methods for the proposed task: a pipeline method based on semantic parsing-based QA systems and an end-to-end method based on large pretrained text generation models, and show that FeTaQA poses a challenge for both methods.

Publisher

MIT Press

Subject

Artificial Intelligence,Computer Science Applications,Linguistics and Language,Human-Computer Interaction,Communication

Reference50 articles.

1. With few eyes, all hoaxes are deep;Asthana;Proceedings of the ACM on Human Computer Interaction,2018

2. METEOR: An automatic metric for MT evaluation with improved correlation with human judgments;Banerjee,2005

3. Question answering from frequently asked question files: Experiences with the faq finder system;Burke;AI Magazine,1997

4. Reading Wikipedia to answer open-domain questions;Chen,2017

5. Open question answering over tables and text;Chen,2021

Cited by 16 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3