Affiliation:
1. University of Southern California Los Angeles California USA
Abstract
AbstractThe recently released BARD and ChatGPT have generated substantial interest from a range of researchers and institutions concerned about the impact on education, medicine, law and more. This paper uses questions from the Watson Jeopardy! Challenge to compare BARD, ChatGPT, and Watson. Using those, Jeopardy! questions, we find that for high confidence Watson questions the three systems perform with similar accuracy as Watson. We also find that both BARD and ChatGPT perform with the accuracy of a human expert and that the sets of their correct answers are rated highly similar using a Tanimoto similarity score. However, in addition, we find that both systems can change their solutions to the same input information on subsequent uses. When given the same Jeopardy! category and question multiple times, both BARD and ChatGPT can generate different and conflicting answers. As a result, the paper examines the characteristics of some of those questions that generate different answers to the same inputs. Finally, the paper discusses some of the implications of finding the different answers and the impact of the lack of reproducibility on testing such systems.
Reference24 articles.
1. GPT-3: What’s it good for?
2. Dave V.2023.10 Insightful Google Bard Statistics & Facts to Know in 2023.https://www.enterpriseappstoday.com/stats/google‐bard‐statistics‐unfolding‐google‐bard.html
3. Duarte F.2023.Number of Chat GPT Users.https://explodingtopics.com/blog/chatgpt‐users
4. Elad B.2023.Google Bard Statistics – Unfolding Google Bard.https://www.enterpriseappstoday.com/stats/google‐bard‐statistics‐unfolding‐google‐bard.html
Cited by
9 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献