Journal Article |
Anatomy of Online Hate: Developing a Taxonomy and Machine Learning Models for Identifying and Classifying Hate in Online News Media
View Abstract
Online social media platforms generally attempt to mitigate hateful expressions, as these comments can be detrimental to the health of the community. However, automatically identifying hateful comments can be challenging. We manually label 5,143 hateful expressions posted to YouTube and Facebook videos among a dataset of 137,098 comments from an online news media. We then create a granular taxonomy of different types and targets of online hate and train machine learning models to automatically detect and classify the hateful comments in the full dataset. Our contribution is twofold: 1) creating a granular taxonomy for hateful online comments that includes both types and targets of hateful comments, and 2) experimenting with machine learning, including Logistic Regression, Decision Tree, Random Forest, Adaboost, and Linear SVM, to generate a multiclass, multilabel classification model that automatically detects and categorizes hateful comments in the context of online news media. We find that the best performing model is Linear SVM, with an average F1 score of 0.79 using TF-IDF features. We validate the model by testing its predictive ability, and, relatedly, provide insights on distinct types of hate speech taking place on social media.
|
2018 |
Salminen, J., Almerekhi, H., Milenković, M., Jung, S.G., An, J., Kwak, H. and Jansen, B.J. |
View
Publisher
|
Journal Article |
Hate Speech Detection on Twitter: Feature Engineering v.s. Feature Selection
View Abstract
The increasing presence of hate speech on social media has drawn significant investment from governments, companies, and empirical research. Existing methods typically use a supervised text classification approach that depends on carefully engineered features. However, it is unclear if these features contribute equally to the performance of such methods. We conduct a feature selection analysis in such a task using Twitter as a case study, and show findings that challenge conventional perception of the importance of manual feature engineering: automatic feature selection can drastically reduce the carefully engineered features by over 90% and selects predominantly generic features often used by many other language related tasks; nevertheless, the resulting models perform better using automatically selected features than carefully crafted task-specific features.
|
2018 |
Robinson, D., Zhang, Z. and Tepper, J. |
View
Publisher
|
Journal Article |
Hierarchical CVAE for Fine-Grained Hate Speech Classification
View Abstract
Existing work on automated hate speech detection typically focuses on binary classification or on differentiating among a small set of categories. In this paper, we propose a novel method on a fine-grained hate speech classification task, which focuses on differentiating among 40 hate groups of 13 different hate group categories. We first explore the Conditional Variational Autoencoder (CVAE) as a discriminative model and then extend it to a hierarchical architecture to utilize the additional hate category information for more accurate prediction. Experimentally, we show that incorporating the hate category information for training can significantly improve the classification performance and our proposed model outperforms commonly-used discriminative models.
|
2018 |
Qian, J., ElSherief, M., Belding, E. and Wang, W.Y. |
View
Publisher
|
Journal Article |
A Survey on Automatic Detection of Hate Speech in Text
View Abstract
The scientific study of hate speech, from a computer science point of view, is recent. This survey organizes and describes the current state of the field, providing a structured overview of previous approaches, including core algorithms, methods, and main features used. This work also discusses the complexity of the concept of hate speech, defined in many platforms and contexts, and provides a unifying definition. This area has an unquestionable potential for societal impact, particularly in online communities and digital media platforms. The development and systematization of shared resources, such as guidelines, annotated datasets in multiple languages, and algorithms, is a crucial step in advancing the automatic detection of hate speech.
|
2018 |
Fortuna, P. and Nunes, S. |
View
Publisher
|
Journal Article |
Automatic Identification and Classification of Misogynistic Language on Twitter
View Abstract
Hate speech may take different forms in online social media. Most of the investigations in the literature are focused on detecting abusive language in discussions about ethnicity, religion, gender identity and sexual orientation. In this paper, we address the problem of automatic detection and categorization of misogynous language in online social media. The main contribution of this paper is two-fold: (1) a corpus of misogynous tweets, labelled from different perspective and (2) an exploratory investigations on NLP features and ML models for detecting and classifying misogynistic language.
|
2018 |
Anzovino, M., Fersini, E. and Rosso, P. |
View
Publisher
|
Journal Article |
Natural Language Understanding and Multimodal Discourse Analysis for Interpreting Extremist Communications and the Re-Use of These Materials Online
View Abstract
This paper reports on a study that is part of a project which aims to develop a multimodal analytical approach for big data analytics, initially in the context of violent extremism. The findings reported here tested the application of natural language processing models to the text of a sample of articles from the online magazines Dabiq and Rumiyah, produced by the Islamic extremist organisation ISIS. For comparison, text of articles found by reverse image search software which re-used the lead images from the original articles in text which either reported on or opposed extremist activities was also analysed. The aim was to explore what insights the natural language processing models could provide to distinguish between texts produced as propaganda to incite violent extremism and texts which either reported on or opposed violent extremism. The results showed that some valuable insights can be gained from such an approach and that these results could be improved through integrating automated analyses with a theoretical approach with analysed language and images in their immediate and social contexts. Such an approach will inform the interpretation of results and will be used in training software so that stronger results can be achieved in the future.
|
2018 |
Wignell, P., Chai, K., Tan, S., O’Halloran, K. and Lange, R. |
View
Publisher
|