Experimental Evaluation: Can Humans Recognise Social Media Bots?

Author:

Kolomeets Maxim12ORCID,Tushkanova Olga2ORCID,Desnitsky Vasily2ORCID,Vitkova Lidia2ORCID,Chechulin Andrey2ORCID

Affiliation:

1. School of Computing, Newcastle University, Newcastle upon Tyne NE4 5TG, UK

2. St. Petersburg Federal Research Center of the Russian Academy of Sciences, 14th Line of V.O. 39, St. Petersburg 199178, Russia

Abstract

This paper aims to test the hypothesis that the quality of social media bot detection systems based on supervised machine learning may not be as accurate as researchers claim, given that bots have become increasingly sophisticated, making it difficult for human annotators to detect them better than random selection. As a result, obtaining a ground-truth dataset with human annotation is not possible, which leads to supervised machine-learning models inheriting annotation errors. To test this hypothesis, we conducted an experiment where humans were tasked with recognizing malicious bots on the VKontakte social network. We then compared the “human” answers with the “ground-truth” bot labels (‘a bot’/‘not a bot’). Based on the experiment, we evaluated the bot detection efficiency of annotators in three scenarios typical for cybersecurity but differing in their detection difficulty as follows: (1) detection among random accounts, (2) detection among accounts of a social network ‘community’, and (3) detection among verified accounts. The study showed that humans could only detect simple bots in all three scenarios but could not detect more sophisticated ones (p-value = 0.05). The study also evaluates the limits of hypothetical and existing bot detection systems that leverage non-expert-labelled datasets as follows: the balanced accuracy of such systems can drop to 0.5 and lower, depending on bot complexity and detection scenario. The paper also describes the experiment design, collected datasets, statistical evaluation, and machine learning accuracy measures applied to support the results. In the discussion, we raise the question of using human labelling in bot detection systems and its potential cybersecurity issues. We also provide open access to the datasets used, experiment results, and software code for evaluating statistical and machine learning accuracy metrics used in this paper on GitHub.

Publisher

MDPI AG

Reference31 articles.

1. Dialektakis, G., Dimitriadis, I., and Vakali, A. (2022). CALEB: A Conditional Adversarial Learning Framework to Enhance Bot Detection. arXiv.

2. Cresci, S., Petrocchi, M., Spognardi, A., and Tognazzi, S. (July, January 30). Better Safe Than Sorry: An Adversarial Approach to Improve Social Bot Detection. Proceedings of the 10th ACM Conference on Web Science, Boston, MA, USA.

3. Detecting Malicious Social Bots Based on Clickstream Sequences;Shi;IEEE Access,2019

4. A decade of social bot detection;Cresci;Commun. ACM,2020

5. Kolomeets, M., and Chechulin, A. (2021, January 12–14). Analysis of the malicious bots market. Proceedings of the 2021 29th Conference of Open Innovations Association (FRUCT), Tampere, Finland.

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3