Journal Article |
Messaging and mobilization: Rebel groups, social media communication, and audience engagement
View Abstract
Mobilization is central to the emergence, survival and success of armed groups challenging the state, and has lately expanded to new arenas with the rise of social media. Using a new dataset of rebel group Twitter use, we examined the topics contained in rebel group social media communications to understand how different messaging strategies impact civilian engagement with rebel messages. Rather than benefiting solely from direct calls to action, we found that rebel groups also increased civilian engagement through indirect messages of self-promotion. While direct appeals received more engagement than indirect appeals, their effects were tempered by audience fatigue when relied on too heavily. We additionally found that including images further enhanced the impact of a mobilizing message. These findings expanded our understanding of rebel communications and mobilization, with important implications for combating the use of social media as a recruitment tool for violent extremism.
|
2023 |
Bestvater, S.E. and Loyle, C.E. |
View
Publisher
|
Journal Article |
Sex Sells Terrorism: How Sexual Appeals in Fringe Online Communities Contribute to Self-Radicalization
View Abstract
The past several years have seen rising hate crimes, terrorist attacks, and broader extremist movements, with news reports often noting that these movements can be traced back to fringe online communities. Yet the question remains why such online groups appear more likely to foster radicalization than those in other contexts. This netnographic case study demonstrates how sexual appeals in fringe online communities facilitate the development of extremist ideologies. Specifically, the cognitive effects of sexual arousal combined with the social norms of such communities contribute to the acceptance of hate speech and fringe ideologies while reducing the extent to which audiences evaluate rational arguments and competing points of view. Thus, sexual appeals paired with messaging or imagery that promotes fringe points of view, which can be more freely expressed in small online groups than in other contexts, are more likely to result in intended attitudinal and behavioral changes—in other words, extremism.
|
2023 |
Britt, B.C. |
View
Publisher
|
|
Politicization and Right-Wing Normalization on YouTube: A Topic-Based Analysis of the “Alternative Influence Network”
View Abstract
Scholarship has highlighted the rise of political influencer networks on YouTube, raising concerns about the platform’s propensity to spread and even incentivize politically extreme content. While many studies have focused on YouTube’s algorithmic infrastructure, limited research exists on the actual content in these networks. Building on Lewis’s (2018) classification of an “alternative influencer” network, we apply structural topic modeling across all text-based autocaptions from her study’s sample to identify common topics featured on these channels. This allows us to gauge which topics appear together and to trace politicization over time. Through network analysis, we determine channel similarities and evaluate whether deplatformed channels influenced topic shifts. We find that political topics increasingly dominate the focus of all analyzed channels. The convergence of culture and politics occurs mostly about identity-driven issues. Furthermore, more extreme channels do not form distinct clusters but blend into the larger content-based network. Our findings illustrate how political topics may function as connective ties across an initially more diverse network of YouTube influencer channels.
|
2023 |
Knüpfer, C.B., Schwemmer, C. and Heft, A. |
View
Publisher
|
Chapter |
The Future of Counterspeech: Effective Framing, Targeting, and Evaluation
View Abstract
Approaches for strategically countering or providing alternatives to hate speech and extremism online have evolved substantively in the last ten years. Technological advancement and a generation of young activists who have been socialized as digital natives have facilitated a maelstrom of both hate-based extremist content and attempts to counter this material in different guises and through diverse channels. The rate and pace of change within the tech sector, and social media growth in particular, have meant that although counterspeech is now more prevalent than ever before, it requires greater guidance and more robust public–private partnerships to effectively prevent and counter extremism online. The chapter embraces a cross-platform and international overview of some of the best practices within efforts to prevent and counter violent extremism online and discusses the future of counterspeech with recommendations for expanded innovation and partnership models.
|
2023 |
Saltman, E. and Zamir, M. |
View
Publisher
|
Journal Article |
Auditing Elon Musk’s Impact on Hate Speech and Bots
View Abstract
On October 27th, 2022, Elon Musk purchased Twitter, becoming its new CEO and firing many top executives in the process. Musk listed fewer restrictions on content moderation and removal of spam bots among his goals for the platform. Given findings of prior research on moderation and hate speech in online communities, the promise of less strict content moderation poses the concern that hate will rise on Twitter. We examine the levels of hate speech and prevalence of bots before and after Musk’s acquisition of the platform. We find that hate speech rose dramatically upon Musk purchasing Twitter and the prevalence of most types of bots increased, while the prevalence of astroturf bots decreased.
|
2023 |
Hickey, D., Schmitz, M., Fessler, D., Smaldino, P.E., Muric, G. and Burghardt, K. |
View
Publisher
|
Journal Article |
Moderating borderline content while respecting fundamental values
View Abstract
As efforts to identify and remove online terrorist and violent extremist content have intensified, concern has also grown about so‐called lawful but awful content. Various options have been touted for reducing the visibility of this borderline content, including removing it from search and recommendation algorithms, downranking it and redirecting those who search for it. This article contributes to this discussion by considering the moderation of such content, in terms of three sets of values. First, definitional clarity. This is necessary to provide users with fair warning of what content is liable to moderation and to place limits on the discretion of content moderators. Yet, at present, definitions of borderline content are vague and imprecise. Second, necessity and proportionality. While downranking and removal from search and recommender algorithms should be distinguished from deplatforming, tech companies’ efforts to deamplify borderline content give rise to many of the same concerns as content removal and account shutdowns. Third, transparency. While a number of platforms now publish their content moderation policies and transparency data reports, these largely focus on violative, not borderline content. Moreover, there remain questions around access to data for independent researchers and transparency at the level of the individual user.
|
2023 |
Macdonald, S. and Vaughan, K. |
View
Publisher
|