Author:
Ali Hussein Al Naffakh Hussein,Dheyaa Radhi Ahmed,A Hakim Baqer,Fuqdan AL-Ibraheemi,Al-Attar Bourair
Abstract
Using complex language models has been a common strategy as personalized news recommendation systems are adopted by online news sites more and more. Before the epoch of GPT-3, news recommendation systems underwent a progression from rule-based and collaborative filtering approaches in the pre-2010s, through the integration of neural networks in the 2010s, to the emergence of earlier iterations of large language models like GPT-2 in 2019. Pre-trained language models have ushered in a new era of recommendation paradigms, thanks to the emergence of huge language models like GPT-3 and T-5. With its easy-to-use interface, ChatGPT is becoming more and more popular for text-based jobs. Focusing on news provider fairness, individualized news recommendations, and fake news identification, this study starts an inquiry into ChatGPT's efficacy in news recommendations. We acknowledge that ChatGPT's output sensitivity to input phrasing is a limitation, and our goal is to investigate these limitations from each angle. We also investigate whether certain prompt formats can help to alleviate these constraints or if more research is needed. To go beyond set assessments, we create a webpage where ChatGPT's performance on the examined activities and prompts is tracked once a week. Through the use of big language models, this work seeks to improve news recommendation performance and stimulate more research in this area.
Reference34 articles.
1. Mingxiao An, Fangzhao Wu, Chuhan Wu, Kun Zhang, Zheng Liu, and Xing Xie. 2019. Neural news recommendation with long-and short-term user representations. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 336–345 (2019).
2. Bang Yejin, Cahyawijaya Samuel, Lee Nayeon, Dai Wenliang, Su Dan, Wilie Bryan, Lovenia Holy, Ji Ziwei, Yu Tiezheng, Chung Willy, et al. 2023. A multitask, multilingual, multimodal evaluation of chatgpt on reasoning, hallucination, and interactivity. arXiv preprint 2302.04023 (2023).
3. Chen Yahui. 2015. Convolutional neural network for sentence classification. Master’s thesis. University of Waterloo (2015).
4. Cho Kyunghyun, Van Merriënboer Bart, Gulcehre Caglar, Bahdanau Dzmitry, Bougares Fethi, Schwenk Holger, and Bengio Yoshua. 2014. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint 1406.1078 (2014).
5. Cui Zeyu, Ma Jianxin, Zhou Chang, Zhou Jingren, and Yang Hongxia. 2022. M6-Rec: Generative Pretrained Language Models are Open-Ended Recommender Systems. arXiv preprint 2205.08084 (2022).