Affiliation:
1. Language Intelligence Research Group Electronics and Telecommunications Research Institute Daejeon Republic of Korea
Abstract
AbstractWe focus on open‐domain question‐answering tasks that involve a chain‐of‐reasoning, which are primarily implemented using large language models. With an emphasis on cost‐effectiveness, we designed EffiChainQA, an architecture centered on the use of small language models. We employed a retrieval‐based language model to address the limitations of large language models, such as the hallucination issue and the lack of updated knowledge. To enhance reasoning capabilities, we introduced a question decomposer that leverages a generative language model and serves as a key component in the chain‐of‐reasoning process. To generate training data for our question decomposer, we leveraged ChatGPT, which is known for its data augmentation ability. Comprehensive experiments were conducted using the HotpotQA dataset. Our method outperformed several established approaches, including the Chain‐of‐Thoughts approach, which is based on large language models. Moreover, our results are on par with those of state‐of‐the‐art Retrieve‐then‐Read methods that utilize large language models.
Funder
Institute for Information and Communications Technology Promotion
Reference35 articles.
1. J.Wei X.Wang D.Schuurmans M.Bosma B.Ichter F.Xia E. H.Chi Q. V.Le andD.Zhou Chain‐of‐Thought prompting elicits reasoning in large language models Vol. 35 2022 pp.24824–24837.
2. J.Maynez S.Narayan B.Bohnet andR.McDonald On faithfulness and factuality in abstractive summarization arXiv preprint 2020 DOI10.48550/arXiv.2005.00661
3. A.Lazaridou E.Gribovskaya W.Stokowiec andN.Grigorev Internet‐augmented language models through few‐shot prompting for open‐domain question answering arXiv preprint 2022 DOI10.48550/arXiv.2203.05115
4. H.He H.Zhang andD.Roth Rethinking with retrieval: faithful large language model inference arXiv preprint 2022 DOI10.48550/arXiv.2301.00303
5. X.Ma Y.Gong P.He H.Zhao andN.Duan Query rewriting for retrieval‐augmented large language models arXiv preprint 2023 DOI10.48550/arXiv.2305.14283
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献