Author:
Abbasi Ahmed,Javed Abdul Rehman,Iqbal Farkhund,Kryvinska Natalia,Jalil Zunera
Abstract
AbstractWith time, numerous online communication platforms have emerged that allow people to express themselves, increasing the dissemination of toxic languages, such as racism, sexual harassment, and other negative behaviors that are not accepted in polite society. As a result, toxic language identification in online communication has emerged as a critical application of natural language processing. Numerous academic and industrial researchers have recently researched toxic language identification using machine learning algorithms. However, Nontoxic comments, including particular identification descriptors, such as Muslim, Jewish, White, and Black, were assigned unrealistically high toxicity ratings in several machine learning models. This research analyzes and compares modern deep learning algorithms for multilabel toxic comments classification. We explore two scenarios: the first is a multilabel classification of Religious toxic comments, and the second is a multilabel classification of race or toxic ethnicity comments with various word embeddings (GloVe, Word2vec, and FastText) without word embeddings using an ordinary embedding layer. Experiments show that the CNN model produced the best results for classifying multilabel toxic comments in both scenarios. We compared the outcomes of these modern deep learning model performances in terms of multilabel evaluation metrics.
Publisher
Springer Science and Business Media LLC
Reference60 articles.
1. Abbasi, A. et al. Elstream: An ensemble learning approach for concept drift detection in dynamic social big data stream learning. IEEE Access 9, 66408–66419 (2021).
2. Khan, M. U., Javed, A. R., Ihsan, M. & Tariq, U. A novel category detection of social media reviews in the restaurant industry. Multimed. Syst. 1–14 (2020).
3. Burnap, P. & Williams, M. L. Cyber hate speech on twitter: An application of machine classification and statistical modeling for policy and decision making. Policy Internet 7, 223–242 (2015).
4. Davidson, T., Bhattacharya, D. & Weber, I. Racial bias in hate speech and abusive language detection datasets. arXiv preprint arXiv:1905.12516 (2019).
5. Kumar, R., Ojha, A. K., Malmasi, S. & Zampieri, M. Benchmarking aggression identification in social media. In Proceedings of the first workshop on trolling, aggression and cyberbullying (TRAC-2018), 1–11 (2018).
Cited by
20 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献