Skip to main content
PLOS One logoLink to PLOS One
. 2023 Jan 25;18(1):e0278511. doi: 10.1371/journal.pone.0278511

Offline events and online hate

Yonatan Lupu 1,2,*, Richard Sear 3, Nicolas Velásquez 4, Rhys Leahy 1,2, Nicholas Johnson Restrepo 2, Beth Goldberg 5, Neil F Johnson 2,3
Editor: Hossein Kermani6
PMCID: PMC9876356  PMID: 36696388

Abstract

Online hate speech is a critical and worsening problem, with extremists using social media platforms to radicalize recruits and coordinate offline violent events. While much progress has been made in analyzing online hate speech, no study to date has classified multiple types of hate speech across both mainstream and fringe platforms. We conduct a supervised machine learning analysis of 7 types of online hate speech on 6 interconnected online platforms. We find that offline trigger events, such as protests and elections, are often followed by increases in types of online hate speech that bear seemingly little connection to the underlying event. This occurs on both mainstream and fringe platforms, despite moderation efforts, raising new research questions about the relationship between offline events and online speech, as well as implications for online content moderation.

Introduction

Online hate and extremism are critical problems across the globe. Online, groups of like-minded, prejudiced individuals create communities on social media platforms, where they recruit new members and sometimes coordinate offline behavior. Users’ exposure to hate speech can have important negative consequences [1]. At times, this online activity can fuel high-profile offline violence, such as the January 6th attack on the U.S. Capitol. In turn, highly charged offline trigger events tend to be followed by sharp increases in online hate speech [2, 3]. Such spikes in online hate speech have, further, been shown to predict similar spikes in offline violent hate crimes [49].

Insights from literatures on social movements can be applied to online extremist communities. These movements typically focus on the creation of a group identity, particularly framed to the exclusion of others [10, 11]. Hate speech plays a particularly important role by intensifying feelings of in-group cohesion and negative feelings toward the out-group [12]. By communicating with each other on social media, individuals affirm their own identities as well as those of the group, often by using specific language, symbols, and images [13, 14]. They use social media to communicate, mobilize, and coordinate collective action [15, 16]. These technologies may facilitate such collective action and make violence more likely [17, 18]. Elites, including the media, play a particularly important role in the process. Like other individuals, members of online hate communities take cues from media and other elites regarding both which issues and events to focus on and the opinions they might hold on those issues and events [1924]. By making some offline events more salient, elites can thus focus the attention of online hate communities, so the relationship between offline trigger events and online hate speech is likely mediated in part by these actors. There is therefore a complex relationship between offline trigger events, elite/media cues and salience, online hate speech, and ensuing extremist violence that scholars across disciplines have sought to analyze and understand.

We study the connection between online hate speech and offline trigger events across the online ecology of moderated and unmoderated platforms. We focus on the following questions: How does online hate speech react to offline trigger events? Do these reactions differ across mainstream platforms that remove some hateful speech versus fringe platforms that are less likely to do so? Are offline events followed by online hate speech that is directly related to those events? Or, once hate speech has been kindled by an offline event, do members of online hate communities respond by using hate speech against seemingly unrelated targets?

Prior work has made significant progress on detecting and analyzing online hate speech. [Excellent surveys of this literature are available at [25, 26]]. With few exceptions [27, 28], existing work focuses on either classifying online hate speech as binary (i.e., hate or not hate) or classifying a small set of hate speech types [5, 2931]. Most prior work tends to draw data from 1 or 2 social media platforms, usually mainstream and moderated platforms [3136]. A recent survey found that Twitter was by far the most studied source in articles using automated detection of online hate speech [26]. Such approaches have two limitations we hope to mitigate. First, hate speech is likely to vary systematically between moderated and fringe platforms. Second, a more comprehensive classification of hate speech types is needed to better understand patterns of online hate speech following offline events.

We conduct a more comprehensive study of online hate speech by including a range of social media platforms and classifying a large set of hate speech types. We conduct a supervised machine learning analysis of 7 types of online hate speech using 59 million English-language posts drawn from 1,150 online hate communities on 6 social media platforms over the course of 19 months. This is a broad sample of online communities, particularly relative to prior work that tends to draw data from relatively few communities on moderated platforms. Our approach has several advantages. First, we include actively moderated mainstream platforms—Facebook, VKontakte, and Instagram—that have and enforce (to varying degrees) policies against hate speech, as well as the less-moderated platforms Gab, Telegram, and 4Chan. This allows us to capture online hate speech that would be filtered out of moderated platforms. Second, we collect data on posts on online hate communities across these platforms [3739], allowing us to capture online hate speech where it is most likely to arise both before and after offline trigger events. Finally, by classifying a broad range of types of hate speech by their targets, we are able to analyze the extent to which offline events are followed by different types of online hate speech.

Data

Our data collection effort begins by identifying online hate communities, which are online forums in which hate speech is most likely to be used. Many social media platforms offer individuals the means to join communities of like-minded users, such as fan pages on Facebook, channels on Telegram, and anonymous message boards on 4Chan. We identified candidate online communities by using the sampling procedure described in greater detail in the SI. Our team manually searched social media platforms for hate communities and search their content to identify new communities. We included a community in our data as a hate community if 2 or more of the 20 most recent posts on the community included hate speech. For this step in the process, we defined hate speech as either (a) content that would fall under the provisions of the United States Code regarding hate crimes or hate speech according to Department of Justice guidelines; or (b) content that supports or promotes Fascist ideologies or regime types (i.e., extreme nationalism and/or racial identitarianism). The determination of whether an online community is a hate community was made manually using these criteria. We identified 1150 online hate communities that are included in this study.

Between June 1, 2019, and December 31, 2020, we captured the posts on these hate communities using a combination of automated and manual methods. We removed non-English-language posts by using Google’s Compact Language Detector 2. Our data include the texts of 59 million posts, and do not include images, videos, and audio. The SI provides information on the distribution of posts and hate communities across the six platforms. We do not collect data on the geographic locations of users. A large share of them appears to be located in the United States based on their discussion of U.S. politics, but there are also frequent discussions of European issues and those of other English-speaking countries. We make no assumptions about the geographic locations of these users.

All data we collected were publicly posted on the Internet, and we had no contact with the users who posted the data. All data used in this study were collected and used in compliance with the terms of use of the online platforms. No personal identifying information was collected. No information about users was collected. User names were not collected. All the data are anonymous. We obtained advanced approval from the IRB.

Our study codes for 7 types of hate speech: speech targeting race, gender, religion, gender identity/sexual orientation (GI/SO), immigration, ethnicity/identitarian/nationalism (E/I/N), as well as anti-semitism. The SI provides additional information about how we define these types of hate speech. We chose these types of hate speech because, in our manual review of the hate communities, these types were both prevalent and distinguishable from each other (although they do co-occur).

Methods

To construct our training data, four annotators were trained to manually code whether or not a post contained each of the 7 types of hate speech. We trained the enumerators to distinguish between mere discussion of an issue and hate speech, e.g., to distinguish between non-hateful discourse about race and racial hate speech. A post could, in principle, be coded as containing between 0 and 7 types of hate speech. In total, the training data include 31,323 manually coded posts; the SI provides additional information about the coding procedure and training data.

Using the algorithm developed by [40], we split these data into 26,354 posts used to train the machine and 4,969 used to test performance. We compared the performance of 8 model architectures with differing classification and text representation methods. Using each model architecture, we trained 7 distinct models, one for each type of hate speech, each of which separately learns from the training data. Each of these 7 models provides a binary coding indicating whether or not a given post contained the applicable type of hate speech or not. We then compared the performance of these 7 models across the 8 model architectures in order to choose the best-performing architecture. The best performing model architecture used a neural network classifier with BERT embeddings [41]. The accuracy ranged from 91.7% to 98.3%, depending on the type of hate speech. We validated the machine results using human annotators and found that the results were highly reliable. The results below are based on the 7 models using this model architecture. The SI provides additional information regarding the 8 model architectures we tested, a comparison of their performance, a detailed description of the performance of the model architecture we chose, and detail regarding the validation exercise.

Results

We begin our analysis by focusing on the distribution of types of hate speech within the data. Panel A of Fig 1 shows how many posts contained each type of hate speech. By far the most common type of hate speech in the communities we tracked is racism, which appears in nearly 7 million posts. Most posts, despite originating from hate communities, did not contain any of the seven types of hate speech, as shown in Panel B. These communities often discuss unrelated issues, and some posts featured content other than text, which we did not analyze. About 20% of hate posts contained more than one type of hate speech. Panel C depicts how frequently each pair of hate speech types co-occurred within the same post. Most pairs are either weakly positively or weakly negatively correlated. The largest positive correlation is between race and E/I/N (0.2396), while the largest negative correlation is between gender and anti-semitism (-0.0226).

Fig 1. Posts, hate posts, and types of hate speech.

Fig 1

Panel A shows how many posts contain each type of hate speech. Panel B shows how many posts contain different quantities of hate speech types. Panel C shows the pairwise correlations between hate speech types within posts. Diagonal lines indicate negative correlations.

We next analyzed how hate speech changed over time. Fig 2 illustrates these changes in hate speech in these communities over time. The SI provides versions of these plots by platform. Panel A shows the temporal trend in the total count of daily posts across all 1150 communities. Activity in these communities differs systematically based on the day of the week, so here and elsewhere our results depict 7-day rolling averages. The total number of posts demonstrates an upward trend over the study period, increasing by about 67% from about 60,000 to 100,000 daily posts.

Fig 2. Temporal trends in posts and hate posts.

Fig 2

Panel A shows the daily posts in the hate communities we tracked over time. Panel B shows the daily hate posts in these communities over time. Panel C shows the daily percentage of posts that were hate posts over time. Panel D shows the mean number of hate speech types in hate posts over time.

There are two especially notable spikes in total activity: in late May/early June 2020, and in early November 2020. They key offline events discussed in the posts during these periods were (1) the May 25 killing of George Floyd by Minneapolis police officers and the ensuing protests organized by the Black Lives Matter (BLM) movement; and (2) the November 2020 U.S. elections. Panel B shows that not only were there far more posts than usual during these two events, there were also increases in hate posts. Most importantly, these results indicate that during periods of intense and contentious offline activity, the activity in online hate communities increases dramatically. In part this be due to the heightened salience of these events more generally, i.e., these events were likely being discussed not only in these hate communities, but elsewhere online, leading to greater online activity in general, including hate speech activity.

Panel C, however, indicates that these two events had differing effects on the relative frequency of hate speech within these communities. During the early days of the BLM protests, the percentage of posts that contained hate speech increase, meaning that hate posts increased not only in absolute terms (as shown in Panel B) but also in relative terms compared to other posts. During the period around the U.S. elections, however, Panel C shows that the percentage of posts that contained hate speech decreased. There was an increase in hate posts (Panel B), but in relative terms their frequency decreased, i.e., posts increased overall at a greater rate than hate posts. Finally, Panel D shows that the mean number of hate speech types contained in hate posts has remained quite consistent over time. Two relatively small decreases occur during the events mentioned above.

We continue analyzing how online hate speech changed over time by focusing on each type of hate speech separately. Fig 3 shows the temporal trends in each type of hate speech. The SI provides versions of these plots by platform. Many events, occurring both offline and online, can affect online hate speech; here, we focus on the largest changes in the time-series during the study period. The largest spikes in online hate activity occurred around the following dates in 2020: (1) January 3 with respect to religion and antisemitism; (2) February 29 with respect to immigration; (3) late May, with respect to race, gender, GI/SO, E/I/N, and antisemitism; and (4) early November, with respect to religion, gender, GI/SO, and immigration. Structural break analysis confirms that these changes in the data were not in line with prior patterns of hate speech (p<0.001) and are the largest breaks in the time-series.

Fig 3. Temporal trends in each type of hate speech.

Fig 3

Each panel shows the total daily posts that contain a given type of hate speech over time. Some types of hate speech appear much more often than others (Fig 1A), so the differences in y-axis scales (i.e., number of posts) should be noted.

We manually reviewed random samples of 500 hate posts from each of the periods surrounding these spikes and concluded they are associated with the following offline events: (1) the assassination of Iranian General Qasem Soleimani; (2) a crisis at the border of Turkey and Greece that gained publicity when a large number of mostly Syrian refugees sought entry to the European Union; (3) the George Floyd murder and ensuing protests; and (4) the November 2020 U.S. elections. Events 1 and 2 were less highly publicized than the latter two, and their effects on our data may be less intuitive. We therefore supplemented our manual analysis with an automated topic analysis to confirm the spikes in hate posts were related to those events. See the SI for details.

We focus our next analysis on these events. Each panel in Fig 4 shows the changes in online hate speech following one of these events. Because different types of hate speech are used at different rates, we normalize the level of each type of hate speech at 100 based on the seven-day rolling average prior to the event. Fig 4 therefore shows the relative changes in the frequency of each type of hate speech after each event. Following the Soleimani assassination, several types of hate speech increased significantly, with the largest increases in religion (90%) and anti-semitism (65%). Interestingly, the religious hate expressed in these posts was largely Islamophobic, whereas the rise in anti-semitism focused on speculation that Israel was behind the assassination. The change in hate speech activity during the Turkey-Greece border crisis is the narrowest of the events we studied: a large increase in anti-immigration hate speech, with relatively little change in other types.

Fig 4. Offline events and online hate.

Fig 4

Each panel shows the changes in the relative levels of each type of hate speech following a trigger event.

The largest increases in hate activity during the study period followed the death of George Floyd and continued during the ensuing protests. The rate of racial hate speech increased by 250% by early June and, by the end of the year, remained about double what it had been prior to this event. This is perhaps not surprising given the negative reaction to the protests in some segments of the public. What is surprising, however, is that the frequency of most other types of hate speech also increased dramatically at this time, especially GI/SO (75%), E/I/N (60%), gender (50%), and anti-semitism (40%). While Floyd’s death and the protests were largely framed as being about racial issues, it is not intuitively clear what relationship these events have to issues of, for example, gender and sexual orientation.

The 2020 U.S. elections also present an interesting case. Unlike the other events we have examined, the election was scheduled, so online communities might react in anticipation of it. Thus, in the fourth subfigure of Fig 4, we normalize hate speech levels a month prior to the election rather than a few days prior. Immediately before the election there was large increase in religion hate, most of which is Islamophobic; this appears to be in response to several events in Europe and not related to the election. By Election Day (November 3) and continuing through the declaration of Joe Biden as winner by the Associated Press (November 7), there were large increases in GI/SO hate speech (over 100%) and immigration hate speech (50%); a wave of gender hate speech soon followed. In some ways, these waves of hate speech are surprising. Issues of gender identity and sexual orientation were not prominent during the campaign; the ensuing wave of associated online hate speech appears to be largely a case of users employing anti-LGBTQ slurs in a generalized manner to malign a wide range of political targets, such as candidates, parties, and voters. Likewise, immigration was a less prominent campaign issue than it had been in 2016, but after Biden’s victory online hate speech increased due to speculation that he would moderate U.S. immigration policy. The ensuing wave of gender hate speech appears to be at least partly targeted toward Vice President Kamala Harris based on our manual review of the relevant posts. Interestingly, although racial issues were at the forefront during the campaign, online race hate speech changed relatively little before and after the election. This may be because the rate of race hate speech, despite having dropped from its high in June, remained elevated throughout this period as compared to the period prior to George Floyd’s death (see Fig 3).

Because hate activity increased to the greatest extent after Floyd’s death, we focus in more detail on this event. Fig 5 shows results similar to those of Fig 4, but broken down by platform. Instagram is excluded from Fig 5 because there were not enough daily hate speech posts during this period to make meaningful inferences about changes in activity. The pattern on 4Chan largely follows the overall pattern across all the platforms because 4Chan makes up the majority of our data. More surprising is the relatively large increase in hate speech on Facebook. On Facebook, the daily average of posts containing race hate speech increased by a factor of 10 within days of Floyd’s death and remained highly elevated (5x) a month later. Other types of hate speech, especially anti-semitism, gender, and GI/SO, also increased by large amounts on Facebook during these offline events. We observe similar patterns on Gab and Telegram: a very large increase in race hate speech along with smaller, but nonetheless notable, increases in several other types of hate that endure several weeks after the event. On VKontakte, however, there was relatively little reaction to these events, perhaps because users there are less attuned to U.S. politics. The increase in GI/SO hate activity on VKontakte in mid-June appears to be unrelated to these events.

Fig 5. Hate speech by platform after George Floyd’s death.

Fig 5

Each panel shows the changes in the levels of each type of hate speech following the death of George Floyd on a given platform.

Discussion

Our findings have several implications for researchers and practitioners. The first is that online hate speech is a large and growing problem that affects both mainstream and fringe platforms. Over the study period, the daily volume of hate speech posts began at an average of 20,000, peaked at around 50,000, and ended at around 25,000. Some prior work has found that the volume of extremist content and activity declined on YouTube, a moderated platform [42, 43]. Yet we find that, on the platforms we studied, including those that are moderated, hate speech increased despite elevated efforts to curtail it by moderated platforms, particularly during the second half of 2020. Although some platforms are increasingly committed to removing hate speech and other malicious content, hate content persists on such platforms (see, for example, Fig 9 in S1 File). Despite efforts by mainstream platforms to remove hate speech, it has grown in volume on fringe platforms such as 4Chan and Telegram (see Figs 8 and 10 in S1 File). Online extremists are known to migrate to less-moderated platforms after being removed from mainstream platforms [38, 44], which may be less susceptible to public and policymaker pressure.

It is therefore crucial to continue to analyze how mainstream platforms’ content moderation affects the overall ecology of online hate speech. In particular, our findings raise several research questions. First, given the prior findings of decreases of extremist content on YouTube and our finding of increases in hate speech on other moderated platforms, we need a better of understanding of the conditions under which platforms can be more effective in moderating content. Key factors appear to include the type of content (e.g., hate speech as compared to other extremist content), the format of the content (e.g., text versus video), and the varying responsiveness of malicious content to offline events. In addition, given how quickly users are known to migrate across platforms, a key issue for future research is the overall impact of this migration on malicious content. When extremists and others who post malicious content are deplatformed from mainstream platforms, how many of them migrate to fringe platforms? Do only the most invested and extreme users do so, or do they bring a more casual following along with them? After such migration, does the nature of malicious content, such as hate speech, change? Does this content become more extreme, hateful, or otherwise malicious, or not?

Secondly, the spikes in online hate speech following offline events are not limited to fringe platforms. We know from prior work that online hate speech can react strongly following offline events [2], but our data allow us to compare the sizes of these increases across platforms and types of hate speech, which yields important new results. After the death of George Floyd, for example, the largest increase in online hate speech in absolute terms (i.e., the number of hate speech posts) was on 4Chan, but in percentage terms the largest increase was on Facebook (see Fig 5). This is a particularly striking result given Facebook’s increased attention to content moderation. The implication of this for moderated platforms is that they should be especially watchful during such periods, but also that increases in hate speech on other, fringe platforms may be predictive of potentially similar increases on their own platforms, a point we hope to explore in future research.

Finally, and perhaps most importantly, we find that online hate speech reacts to differing offline events in different and sometimes counter-intuitive ways. We might conceptualize at least three types of offline events in terms of the patterns of hate speech that follow them. First would be offline events that could plausibly be followed by significant increases in hate speech (i.e., events involving frequent targets of hate speech), but that draw little reaction from online hate communities. For example, on March 13, 2020, Louisville, Kentucky, police fatally shot an African-American woman named Breonna Taylor in her home. The killing was followed by a relatively small discussion in the online hate communities we tracked–in contrast to the large reaction to the murder of George Floyd two months later. A second type of offline event is one that is followed by a discrete increase in online hate speech that appears to be directly related to the offline event, and then dissipates relatively quickly. The assassination of Suleimani and the Turkish-Greek border crisis appear to be events of this type. A third type of offline event is followed by sharp increases in and broad cascades of online hate speech that build on, and potentially transform, the meaning and interpretation of the initial offline event and target groups in an extensive and seemingly indiscriminate manner. Events of this type can have longer-lasting effects on patterns and volume of online hate speech. In our study period, the murder of George Floyd and the U.S. Elections appear to be events of this type. It may be intuitive, for example, that online hate speech regarding race would increase during the BLM protests given the way those events were centered on the subject of race. Yet most other types of online hate speech also increased during that time, including, for example, GI/SO hate, which, on the surface, would appear to have an indirect connection to issues of police brutality and race.

Conclusions

What factors explain the different relationships between individual offline events and online hate speech? A detailed analysis of this question is beyond the scope of this paper, but media and elite attention likely play an important role, as discussed in the introduction. Both the volume and variety of online reactions to offline events depend, in part, on the salience of those events in other media. For example, the killing of Taylor received little attention in the U.S. national media in the weeks that followed it (perhaps because it occurred during the time when many U.S. jurisdictions were announcing the first lockdowns of the COVID-19 pandemic), and became significantly more salient during the protests that followed the Floyd murder. The Soleimani assassination and EU border crisis received media attention, but for a shorter time span than the BLM protests and U.S. election. These observations suggest that media attention may be one of the key factors that explains why the patterns of online hate speech following individual offline events differ so sharply.

Yet this factor likely does not explain the range of our findings. For example, if online racist hate speech was, in part, activated by media attention to the BLM protests that made racial identity a more salient issue, why did other forms of hate speech also increase dramatically at that time? Why does this broad activation of online hate speech occur following some offline events but not others? We hope future work will investigate these questions. A particularly important question for future work is to investigate which types of offline events are more likely to be followed by broad and seemingly indiscriminate cascades of online hate speech. Likewise, our results also raise the question of which types of online hate speech, once triggered by such events, are more likely to be followed by such a cascade into other types of hate speech.

Our study faces two important limitations. It is not possible to identify the entire population of online hate communities and sample from it. The online ecology of social media platforms is vast and constantly changing. Tracking and capturing data on these platforms is also limited by terms-of-service and privacy concerns. It is therefore not possible to capture the entire ecology of online speech nor to confidently claim that any sample of such speech is representative of the ecology. Our list of online hate communities is nonetheless the most comprehensive such list we are aware of. After several months of search by several members of our team, we are not aware of any other communities on the 6 platforms we studied that meet our definition of hate communities. We have captured many of the most prominent and active online communities on which hate speech is regularly used. Our findings and conclusions are limited to the public communities we identified. We can infer from our results the relationship between offline events and the use of hate speech on our sample of prominent online hate communities, but we cannot infer from our results, nor from those of prior work, whether or not this relationship holds across the online ecology.

A second limitation of our work is that our analysis does not allow for causal inference between observed offline and online phenomena. We reviewed the posts associated with the spikes in online hate speech discussed above and concluded they are substantively related to the trigger events we described. Following the death of George Floyd protests, for example, the large increase in posts that contained race hate speech (see Fig 4) included many posts that referenced Floyd, the circumstances surrounding his death, and the ensuing protests, leading us to conclude the increase in online hate speech was related to these offline events.

We conclude by noting that much work remains to be done to better understand the relationship between offline events and online hate speech. We have conducted a wide-ranging analysis of online hate speech–across types of hate speech, many online communities, and a diverse set of platforms. We have provided evidence that suggests complex relationships between offline events and online hate, between different types of online hate, and between moderated and fringe platforms. These results raise important questions for scholars across multiple disciplines and suggest several strategies for mitigating online hate speech. First, content moderators should be increasingly watchful of offline events, and efforts may need to be augmented following highly salient events. Second, the effects of an offline event on online hate speech may be difficult to predict, and may quickly evolve and broaden, which further suggests a broad and rapid mitigation effort in response. Because online hate speech levels are predictive of offline hate crimes [49], the potential exists for an offline trigger event to be followed by a broad cascade of online hate speech that, in turn, precedes hate crimes of a nature unrelated to the initial event. Third, because the effects of an offline event on different types of platforms are often similar, content moderators on mainstream platform could advance their efforts by systematically observing ongoing developments on other, fringe platforms.

Supporting information

S1 File. Supplementary material to the manuscript.

(PDF)

Data Availability

We have created an online depository for our replication data. This is available at the Harvard Dataverse here: https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/KKYTUN.

Funding Statement

YL and NFJ received funding from the Air Force Office of Scientific Research (FA9550-20-1-0382 and FA9550-20-1-0383) and National Science Foundation (SES-2030694). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Oksanen A., Hawdon J., Holkeri E., Näsi M., & Räsänen P. (2014). Exposure to online hate among young social media users. In Soul of society: a focus on the lives of children & youth. Emerald Group Publishing Limited. [Google Scholar]
  • 2.Burnap P., & Williams M. L. (2016). Us and them: identifying cyber hate on Twitter across multiple protected characteristics. EPJ Data science, 5, 1–15. doi: 10.1140/epjds/s13688-016-0072-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Siegel, A., Tucker, J., Nagler, J., & Bonneau, R. (2018). Socially mediated sectarianism. Unpublished manuscript. https://alexandra-siegel.com/wp-content/uploads/2019/05/Siegel_Sectarianism_January2017.pdf
  • 4.Müller, K., & Schwarz, C. (2020). From hashtag to hate crime: Twitter and anti-minority sentiment. SSRN 3149103.
  • 5.Williams M. L., Burnap P., Javed A., Liu H., & Ozalp S. (2020). Hate in the machine: Anti-Black and anti-Muslim social media posts as predictors of offline racially and religiously aggravated crime. The British Journal of Criminology, 60(1), 93–117. [Google Scholar]
  • 6.Benesch S., Buerger C., Glavinic T., Manion S., & Bateyko D. (2018). Dangerous speech: A practical guide. Dangerous Speech Project. [Google Scholar]
  • 7.Müller K., & Schwarz C. (2021). Fanning the flames of hate: Social media and hate crime. Journal of the European Economic Association, 19(4), 2131–2167. [Google Scholar]
  • 8.Karell D., Linke A., and Holland E. C. 2021. "Right-Wing Social Media and Political Unrest", SocArXiv. [Google Scholar]
  • 9.Olteanu, A., Castillo, C., Boy, J., & Varshney, K. (2018, June). The effect of extremist violence on hateful speech online. In Proceedings of the international AAAI conference on web and social media (Vol. 12, No. 1).
  • 10.Hunt S. A., & Benford R. D. (1994). Identity talk in the peace and justice movement. Journal of Contemporary Ethnography, 22(4), 488–517. [Google Scholar]
  • 11.Snow, D. (2001). Collective identity and expressive forms. https://escholarship.org/uc/item/2zn1t7bj;jsessionid&.
  • 12.Vollhardt J., Coutin M., Staub E., Weiss G., & Deflander J. (2006). Deconstructing hate speech in the DRC: A psychological media sensitization campaign. J. Hate Stud., 5, 15. [Google Scholar]
  • 13.Hine, G. E., Onaolapo, J., De Cristofaro, E., Kourtellis, N., Leontiadis, I., Samaras, R., et al. (2017, May). Kek, cucks, and god emperor trump: A measurement study of 4chan’s politically incorrect forum and its effects on the web. In Eleventh International AAAI Conference on Web and Social Media.
  • 14.Klein O., Spears R., & Reicher S. (2007). Social identity performance: extending the strategic side of SIDE. Personality and Social Psychology Review, 11(1), 28–45. doi: 10.1177/1088868306294588 [DOI] [PubMed] [Google Scholar]
  • 15.Ackland R., & O’Neil M. (2011). Online collective identity: the case of the environmental movement. Social Networks, 33(3), 177–190. [Google Scholar]
  • 16.Polletta F., & Jasper J. M. (2001). Collective identity and social movements. Annual Review of Sociology, 27(1), 283–305. [Google Scholar]
  • 17.Pierskalla J. H., & Hollenbach F. M. (2013). Technology and collective action: The effect of cell phone coverage on political violence in Africa. American Political Science Review, 107(2), 207–224. [Google Scholar]
  • 18.Weidmann N. B. (2015). Communication networks and the transnational spread of ethnic conflict. Journal of Peace Research, 52(3), 285–296. [Google Scholar]
  • 19.Zaller J. R. (1992). The nature and origins of mass opinion. Cambridge university press. [Google Scholar]
  • 20.Druckman James N. 2001. “The Implications of Framing Effects for Citizen Competence.” Political Behavior 23 (3): 225–56 [Google Scholar]
  • 21.Siegel A. A., & Badaan V. (2020). #No2Sectarianism: Experimental approaches to reducing sectarian hate speech online. American Political Science Review, 114(3), 837–855. [Google Scholar]
  • 22.Del Vicario M., Vivaldo G., Bessi A., Zollo F., Scala A., Caldarelli G., et al. 2016. Echo chambers: Emotional Contagion and Group Polarization on Facebook. Scientific Reports, 6(1), 37825. doi: 10.1038/srep37825 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.McCombs M. E., & Shaw D. L. (1972). The agenda-setting function of mass media. Public opinion quarterly, 36(2), 176–187. [Google Scholar]
  • 24.Salminen J., Sengün S., Corporan J., Jung S. G., & Jansen B. J. (2020). Topic-driven toxicity: Exploring the relationship between online toxicity and news topics. PloS one, 15(2), e0228723. doi: 10.1371/journal.pone.0228723 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Schmidt, A., & Wiegand, M. (2017, April). A survey on hate speech detection using natural language processing. In Proceedings of the fifth international workshop on natural language processing for social media (pp. 1–10).
  • 26.Fortuna P., & Nunes S. (2018). A survey on automatic detection of hate speech in text. ACM Computing Surveys (CSUR), 51(4), 1–30. [Google Scholar]
  • 27.Silva, Leandro, Mainack Mondal, Denzil Correa, Fabricio Benevenuto, and Ingmar Weber. 2016. “Analyzing the Targets of Hate in Online Social Media.” ArXiv:1603.07709 [Cs], March. http://arxiv.org/abs/1603.07709.
  • 28.Alorainy, Wafa, Pete Burnap, Han Liu, and Matthew Williams. 2018. “The Enemy Among Us: Detecting Hate Speech with Threats Based ‘Othering’ Language Embeddings.” ArXiv:1801.07495 [Cs], March. http://arxiv.org/abs/1801.07495.
  • 29.Nemanja Djuric, Jing Zhou, Robin Morris, Mihajlo Grbovic, Vladan Radosavljevic, and Narayan Bhamidipati. 2015. Hate speech detection with comment embeddings. In Proceedings of the 24th International Conference on World Wide Web Companion. International World Wide Web Conferences Steering Committee, pages 29–30.
  • 30.Chikashi Nobata, Joel Tetreault, Achint Thomas, Yashar Mehdad, and Yi Chang. 2016. Abusive Language Detection in Online User Content. In Proceedings of the 25th International Conference on World Wide Web. International World Wide Web Conferences Steering Committee, pages 145–153.
  • 31.Burnap P., & Williams M. L. (2015). Cyber hate speech on twitter: An application of machine classification and statistical modeling for policy and decision making. Policy & internet, 7(2), 223–242. [Google Scholar]
  • 32.Guang Xiang, Bin Fan, Ling Wang, Jason Hong, and Carolyn Rose. 2012. Detecting offensive tweets via topical feature discovery over a large scale twitter corpus. In Proceedings of the 21st ACM international conference on Information and knowledge management, pages 1980–1984, Maui, HI, USA. ACM.
  • 33.Zhang Z., & Luo L. (2019). Hate speech detection: A solved problem? the challenging case of long tail on twitter. Semantic Web, 10(5), 925–945. [Google Scholar]
  • 34.Del Vigna, F., Cimino, A., Dell’Orletta, F., Petrocchi, M., & Tesconi, M. (2017). Hate me, hate me not: Hate speech detection on facebook. In Proceedings of the First Italian Conference on Cybersecurity (ITASEC17) (pp. 86–95).
  • 35.Berglind, T., Pelzer, B., & Kaati, L. (2019, August). Levels of hate in online environments. In Proceedings of the 2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (pp. 842–847).
  • 36.Siegel A. A., Nikitin E., Barberá P., Sterling J., Pullen B., Bonneau R., et al. (2021). Trumping hate on Twitter? Online hate speech in the 2016 US election campaign and its aftermath. Quarterly Journal of Political Science, 16(1), 71–104. [Google Scholar]
  • 37.Johnson N. F., Zheng M., Vorobyeva Y., Gabriel A., Qi H., Velásquez N., et al. (2016). New online ecology of adversarial aggregates: ISIS and beyond. Science, 352(6292), 1459–1463. doi: 10.1126/science.aaf0675 [DOI] [PubMed] [Google Scholar]
  • 38.Johnson N. F., Leahy R., Restrepo N. J., Velasquez N., Zheng M., Manrique P., Devkota P. & Wuchty S. (2019). Hidden resilience and adaptive dynamics of the global online hate ecology. Nature, 573(7773), 261–265. doi: 10.1038/s41586-019-1494-7 [DOI] [PubMed] [Google Scholar]
  • 39.Johnson N. F., Velásquez N., Restrepo N. J., Leahy R., Gabriel N., El Oud S., et al. (2020). The online competition between pro-and anti-vaccination views. Nature, 582(7811), 230–233. doi: 10.1038/s41586-020-2281-1 [DOI] [PubMed] [Google Scholar]
  • 40.Sechidis, K., Tsoumakas, G., & Vlahavas, I. (2011, September). On the stratification of multi-label data. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases (pp. 145–158). Springer, Berlin, Heidelberg.
  • 41.Devlin, J., Chang, M. W., Lee, K., & Toutanova, K. (2018). Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
  • 42.Munger K., & Phillips J. (2020). Right-Wing YouTube: A supply and demand perspective. The International Journal of Press/Politics, 1940161220964767. [Google Scholar]
  • 43.Hosseinmardi H., Ghasemian A., Clauset A., Mobius M., Rothschild D. M., & Watts D. J. (2021). Examining the consumption of radical content on YouTube. Proceedings of the National Academy of Sciences, 118(32). doi: 10.1073/pnas.2101967118 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Mitts T. (2021). Countering Violent Extremism and Radical Rhetoric. International Organization, 1–22. [Google Scholar]

Decision Letter 0

Stefano Cresci

1 May 2022

PONE-D-22-01701Offline Events and Online Hate PLOS ONE

Dear Dr. Lupu,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Specifically, pay particular attention to the doubts that all reviewers raised about the methodology and to the need to better frame and motivate the manuscript. The former point includes providing thorough information on how the data was chosen, collected, and labeled, and providing adequate evidence of the classifiers' accuracy.

Please submit your revised manuscript by Jun 15 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Stefano Cresci

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please provide additional details regarding participant consent. In the Methods section, please ensure that you have specified (1) whether consent was informed and (2) what type you obtained (for instance, written or verbal). If the need for consent was waived by the ethics committee, please include this information.

*In line with our policy regarding third-party data use (https://journals.plos.org/plosone/s/submission-guidelines#loc-personal-data-from-third-party-sources), please provide a copy of the protocol that your IRB reviewed and approved, and which specifies what types of data you collected and how data were anonymized. Please submit this as an ‘Other’ file, as this will be only for our records and will not be published."

- please also ping me and the Data team at PRTC to discuss the data sharing/availability options for this ms

3. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide

Additional Editor Comments (if provided):

Thank you for giving us the chance to evaluate this manuscript. All reviewers found merit in this submission. However, they all also highlighted several major shortcomings that need to be addressed. Among them, pay particular attention to the doubts that all reviewers raised about the methodology and to the need to better frame and motivate the manuscript. The former point includes providing thorough information on how the data was chosen, collected, and labeled, and providing adequate evidence of the classifiers' accuracy. Good luck for the revision.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

Reviewer #3: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This manuscript addresses an important topic: how hate speech responds to offline events. It also engages in measurement of cross-platform hate speech directed at multiple targets. While I am generally very supportive of such efforts to systematically measure and track online speech, I have several concerns, which I detail below.

First, there are several aspects of the authors’ sampling approach that they should do more to explain. Why do the authors look at only one community on 4chan? This seems particularly odd given that the vast majority of the authors’ training data comes from 4chan. How might this bias the authors’ ability to detect hate speech across platforms? The majority of their training data also focuses on racist hate speech in particular, but is this an artifact of using almost entirely data from a single 4chan community? This likely limits the authors ability to detect other hate speech types in their data, particularly across platforms. More generally, what is the advantage in focusing on hate groups rather than representative samples of posts across each platform (see Siegel et al. 2021). The authors need to do more to justify this approach.

Second, given how rare many types of hate speech are in the authors’ data, it seems the classifier performance metrics (often in the high 90s) are likely inflated. I would like to see balanced accuracy metrics to better understand whether high classifier performance is simply an artifact of having so many 0s in the data. The authors should also conduct systematic validation exercises, taking random samples of their data and using human coding to determine how accurate their classifier performance is in practice.

Third, the authors focus on measuring changes in volume over time makes it very difficult to interpret their results. In the aftermath of an event that makes a particular group identity more salient (eg George Floyd protests), we should expect to see large increases in discussions of that group (eg African Americans) online. This would lead to an increase in hate speech volume simply because the group is discussed more often and a certain percentage of posts about that group will always include hate speech online. Indeed, when the authors do look at their results in percentage terms, they yield very different findings. The authors should do more to describe changes in absolute vs relative terms in their data and interpret any results referencing absolute changes with caution.

Reviewer #2: This paper advances the state-of-the-art on Hate-Speech characterization by analysing seven types of online hate speech on a set of six social media which includes both moderated and unmoderated platforms; moreover, it investigates the connection between offline and online hate speech.

The research questions are very interesting, and the findings are effectively supported by plots and the written discussion. Also, the data collection is very accurate. Though I have got some stylistic suggestions to give and doubts about the classification methods to solve, I believe that this work is extremely valuable.

Stylistic suggestions:

1) As far as it concerns the paper’s structure, the ‘Introduction’ title should be written (maybe it’s an oversight), and the Conclusion section should be added; also, it would be great to separate the data collection and the methods, making it clear that the methods are about the multi-label classification of posts.

2) Move the discourse of lines [67-69] right before “Second, by classifying …” at line 65, because you’re talking about the six platforms again.

3) You could start the ‘Classification Method’ section including lines [106-121]. Choose an appropriate section title.

4) Before stating which is the best classifier model at line 115, explain in more details the goal of that model:

4.1) Generally speaking, you built 6 models whose goal is classifying the 50k posts according to the seven hate-speech types, or assigning the ‘no hate speech is present’ label. You should explicit this goal.

4.2) Clearly explain how the model is used to derive the 7 machine learning algorithms; additionally, it’s not quite understandable what are those 7 machine learning algorithms.

4.3) Finally, if for each hate-speech type you used a binary classifier telling whether a post does or does not contain a given type of hate speech, how do you label a post as ‘hate-speech not detected’?

5) Figures:

5.1) Provide color-blind friendly figures. For example, use different markers, line sizes, grey shades.

5.2) Provide higher quality images following PLOS ONE’s guidelines.

5.3) Fig 1, panel B: it needs a better title, like “Total posts containing N hate speech typologies”; add the x-axis label, something like “N Hate Speech Typologies” or “How Many Hate Speech Typologies”

5.4) Fig 1, panel C: you cannot clearly detect the ‘-‘ sign before the ‘0.1’ at the end of the scale. Besides, the negative values (negative correlation) should be well separated from the positive ones, otherwise the correlation between Gender and Religion (to cite an example) is doubtful. Moreover, what does the dots’ size mean?

6) Line 212: specify that you are talking about the fourth subfigure in Fig. 4

7) Supporting Information document:

7.1) Fig 1, panel B: same suggestions as for the main paper

7.2) Maybe the ‘Machine Learning Methods and Model Selection’ section could benefit from a figure showing how the generic model is used to classify each post with hate-speech types. I did not understand how the ‘weights’ are assigned to perform binary classification (at the end of the first paragraph of this section)

7.3) The accuracy should be used only when the dataset is balanced, hence Fig. 2 may show the F1 score instead of the accuracy.

Classification method doubts:

1) if for each hate-speech type you used a binary classifier telling whether a post does or does not contain a given type of hate-speech, how do you label a post as ‘hate-speech not detected’? From what I understand, a post is labelled as ‘hate-speech not detected’ if for all seven binary classifiers that post is classified as ‘hate-speech not detected’, right?

2) Inside the Supporting Information: “[…] whose output is used for 7 separate densely-connected layers […]”, there are 7 separated layers because each one is used to detect a hate speech type, right?

Reviewer #3: The paper "Offline Events and Online Hate" is a well written manuscript that deals with a highly significant issue in social science research by using data science techniques. I would stress two main merits of this research: (i) the chosen approach that looks into the relationship between "real" events and the online discursive world, which is one of the main challenges of social science research for the next years, given the fact that these two stages are fully interconnected but we are not sure yet how this imbrication works; (ii) the huge collected data with an unprecedented survey of online hate communities across different platforms (and not only Twitter!) and the specificity of hate speech (divided into 7 categories).

However, the manuscript has some important shortfalls that the authors should address:

1. The title is ambitious and the empirical findings do not provide enough evidence to explain the relationship between offline events and online hate (the authors themselves write this is beyond the the scope of the paper). This initial consideration about the title also affects the general narrative of the manuscript.

2. The rationale of the paper is not only short but weak. There is not mention to any theoretical background that help to understand the relationship between offline events and online discourses, which in turn would provide a better justification of the RQ's. Which social science theoretical approaches can support the ideas or hypotheses included in this research? Even with a data-driven approach this paper should reflect on this important matter: How can we theoretically model the connections between offline events - media - hate speech - hate crimes? In fact, the authors stressed the importance of what they call "media attention" to understand the connection between the first three elements, but there not a single mention to well-grounded theoretical approaches that already explain part of this process, such as mediatization or, more specifically, the Agenda-Setting hypothesis. If offline events affect the amount of hate speech through the media, some attention should be payed to explain this relationship. Interestingly, the seminal research also suggest that hate speech derives in more hate crimes, which is a offline event again but of a different nature. Probably, including the theoretical understanding of this (recursive) pipeline would be of significant help for the scientific community.

3. The manual classification for the training corpus (made by 4 coders) should be better explained (in SI or in the paper). First, which kind of training received these coders?, were they supervised (i.e. by choosing and double-checking a random sample of the messages they coded)?; Second, and more important, what is the reliability of the manual annotation?, is there an inter-coder agreement test that guarantees the reliability of the coding process? Were some messages coded simultaneously by 2 or more coders?, and if it was the case, what happened to those messages in which there was not agreement? This may seem a small detail in the method, but I consider that the first step to obtain trust in the ML models is to have a really good training corpus and to be sure the the manual classification was of high quality.

4. Regarding the "manual review" of the hate posts, the authors should better explain how it was done (qualitatively, I guess) (and how many of them were reviewed before going to topic modeling) and specially how they linked the topics to the offline events. Did they build a list of offline events? Or how they decided to mention the set offline events included in the manuscript and not some others. Could they bring some agenda-setting studies to get the events with more media attention during the studied period? Moreover, when linking the spikes of hate speech to the events using the main topic detected with LDA, I wonder which other topics were found (I didn't see it in the SI) and if they could be meaningful in some way.

5. When choosing the ML models, I wonder why specific NN were not tested, even when they are state-of-the-art in supervised text classification. Specifically, I am referring to RNN and CNN, which had been used in many previous studies of hate speech detection.

6. Some minors details should be addressed such as the correct references to figures in line 266 (p. 17)

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2023 Jan 25;18(1):e0278511. doi: 10.1371/journal.pone.0278511.r002

Author response to Decision Letter 0


9 Nov 2022

Please see the response memo at the end of this pdf for a detailed response to the editor and reviewer comments. Thank you.

Attachment

Submitted filename: Plos One Response Memo.docx

Decision Letter 1

Hossein Kermani

18 Nov 2022

Offline Events and Online Hate

PONE-D-22-01701R1

Dear Dr. Lupu,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Hossein Kermani

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Congratulations! 

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #2: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #2: Yes

Reviewer #3: (No Response)

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #2: Yes

Reviewer #3: (No Response)

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: No

Reviewer #3: (No Response)

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #2: Yes

Reviewer #3: (No Response)

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #2: The authors addressed all comments and requests by better structuring the manuscript and detailing the data gathering and methodology in an exhaustive way. The analysis has been performed appropriately, leading to interesting results and discussions.

As for the data availability, the authors decided to share them only after the acceptance of the manuscript.

Reviewer #3: All my comments have been addressed. The paper is now ready for publication. The manuscript constitutes a contribution to the field.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: No

Reviewer #3: Yes: Carlos Arcila Calderón

**********

Acceptance letter

Hossein Kermani

23 Dec 2022

PONE-D-22-01701R1

Offline Events and Online Hate

Dear Dr. Lupu:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Hossein Kermani

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. Supplementary material to the manuscript.

    (PDF)

    Attachment

    Submitted filename: Plos One Response Memo.docx

    Data Availability Statement

    We have created an online depository for our replication data. This is available at the Harvard Dataverse here: https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/KKYTUN.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES