Skip to main content
PLOS One logoLink to PLOS One
. 2022 Dec 14;17(12):e0277292. doi: 10.1371/journal.pone.0277292

Polarization and trust in the evolution of vaccine discourse on Twitter during COVID-19

Ignacio Ojea Quintana 1,*, Ritsaart Reimann 2, Marc Cheong 3, Mark Alfano 2, Colin Klein 1
Editor: Hossein Kermani4
PMCID: PMC9749990  PMID: 36516117

Abstract

Trust in vaccination is eroding, and attitudes about vaccination have become more polarized. This is an observational study of Twitter analyzing the impact that COVID-19 had on vaccine discourse. We identify the actors, the language they use, how their language changed, and what can explain this change. First, we find that authors cluster into several large, interpretable groups, and that the discourse was greatly affected by American partisan politics. Over the course of our study, both Republicans and Democrats entered the vaccine conversation in large numbers, forming coalitions with Antivaxxers and public health organizations, respectively. After the pandemic was officially declared, the interactions between these groups increased. Second, we show that the moral and non-moral language used by the various communities converged in interesting and informative ways. Finally, vector autoregression analysis indicates that differential responses to public health measures are likely part of what drove this convergence. Taken together, our results suggest that polarization around vaccination discourse in the context of COVID-19 was ultimately driven by a trust-first dynamic of political engagement.

Introduction

Trust in vaccination remains high, but is eroding in many parts of the world [1, 2]. Decreased confidence in the safety, efficacy, and importance of vaccination may manifest as open skepticism and conspiracy theorizing. It may also manifest more subtly in vaccine hesitancy, which leads to questioning the need for vaccination. Hesitancy comes in degrees: some people are ‘accepters’ while others are ‘fence-sitters’ or ‘rejecters’ [3]. Confidence is impacted by lack of information and access to misinformation, and by distrust of medical and government sources [46].

When people lose trust in medical experts and public health officials, they tend to turn to other sources, including social media. Social media sites optimize for engagement, rather than other measures such as information veracity or epistemic well-being [79]. This is concerning because exposure to negative opinions about vaccines on social media has been shown to be among the strongest predictors both of expressing such opinions oneself [10] and of failure to vaccinate [11].

Expressions of vaccine hesitancy on social media, and in particular on Twitter, have been shown to co-vary with offline expressions of the same sentiment [12]. Furthermore, a pair of recent, pre-COVID-19 studies found that the English-language discourse about vaccines on Twitter is highly polarized, that the anti-vaccine camp has greater reach and receptivity, and that discussants tend to rely on and amplify just a few, non-independent sources [13, 14].

Most of the pre-pandemic discourse about vaccination revolved around well-established immunizations to well-understood childhood diseases, seasonal influenza, and human papillomavirus. Well-established vaccines suffer from the ‘curse of success’ because their widespread administration has in many cases reduced incidence levels of the relevant diseases to near-zero. By contrast, in the early part of 2020 COVID-19 was not well understood, and vaccines against it were months away. Debates about vaccine efficacy and side-effects were therefore conducted in the absence of empirical evidence.

To shed light on the evolution of social media discourse around vaccines in the first few months of the pandemic, we conducted an observational study of Twitter discourse around vaccine topics in general between December 27th 2019, and May 5th 2020. This time frame spans 75 days prior to the World Health Organization’s March 11th 2020 pandemic declaration through to 75 days after. The WHO’s declaration was not merely symbolic; it also triggered a series of institutional responses, some of which (we show) significantly affected vaccine discourse.

Previous work on Reddit has demonstrated the utility of mixed methods for understanding the evolution of online engagement in complex domains [15]. We set out to use a combination of methods to answer the following three research questions:

  • RQ1. Which groups are most important in the English-language discourse around vaccines on Twitter?

  • RQ2. How did vaccination-related engagement and discourse change over the first five months (12/2019–05/2020) of the pandemic?

  • RQ3. What social forces might help in explaining observed changes in engagement?

As explained in the coming sections, we used modularity clustering to answer the first question, Linguistic Inquiry and Word Count (LIWC) to answer the second, and vector autoregression analysis (VAR) for the third. We found that authors cluster into five interpretable groups, and that the discourse was greatly affected by American partisan politics. Since we did not limit data collection to the U.S. nor any other territory, the prevalence of American politics within our analysis is partly due to our focus on English-language discourse; partly due to the fact that the vast majority of Twitters English speaking user-base is located in the United States; and partly a reflection of the extent to which American political cleavages define global online discussions, at least when those discussions are carried out in English. Important to note therefore is that even though we distinguish ‘Democratic’ and ‘Republican’ clusters of users, it is not the case that all users in our data set are based in the United States. This is especially clear with respect to the coalitions that emerge between ‘Democrats’ and ‘Public Health’ on the one hand and ‘Republicans’ and ‘Antivaxxers’ on the other, for neither health institutions nor Antivaxxers are unique to U.S. discourse.

Finally, since ‘Antivaxxer’ is a loaded and much contested concept, it bears spelling out that our use of the term is descriptive rather than normative and, more than anything else, reflects our interpretation of the top accounts (e.g., StopVaxTyranny) and most popular hashtags (#VaccineRoullete) that connect members of this cluster. Concomitantly, we operationalize this label simply as a set of social relations around anti-vaccination attitudes. This definition is deliberately open-ended, for tracking how these attitudes and social relations developed over the course of the pandemic is a core objective of the current paper.

With this in mind, our linguistic analysis zeros in on both the moral and non-moral language used by the various communities. In so doing, we identify significant patterns of thematic converge and divergence within and between both sets of coalitions as the pandemic unfolds. Finally, we find that these patterns can be partially explained at hand of VAR analysis, which shows that different groups responded differently to public health interventions. This result, combined with the observed changes in language use, leads us to the conclusion that polarization in the context of COVID-19 can be explained by a trust-first dynamic of political engagement: the realignment of interests over the course of the pandemic was driven less by shared information or shared values, and more by a proclivity of individuals to trust those with whom they already had some previous pattern of interaction.

The next section describes the methodology with respect to data collection, network construction, community detection, linguistic study, and time series analysis. The section after presents the results as organised by the three guiding research questions. The final two sections include a discussion of the results, statement of limitations, and a conclusion with some policy suggestions.

Materials and methods

Data collection

We queried the Twitter Streaming API with a series of vaccination-related keywords, hashtags, and short expressions between December 2019 and June 2020. Some examples include: ‘vax’, ‘vaxxed’, ‘vaccine’, ‘vaccination’, ‘antivax’, ‘anti-vax’, ‘anti vax’, #vaxsafety’, ‘#vaccineswork’, ‘#novax’, ‘#antivax’. The choice of words was done following similar literature on vaccination discourse on Twitter [16], and with the goal of trying to capture a wide spectrum vaccine related attitudes. See S1 Appendix for a complete list. Taking as a reference point the date the World Health Organization declared COVID-19 a pandemic (March 11th 2020), we divided our data set into a symmetric time-span of tweets comprised of approximately 1.3 million original tweets and 18 million retweets between December 27th 2019 and May 26th 2020.

Because we were interested in the interactions between users and groups, we focused on retweets rather than tweets with original content. The phenomenon of ‘signal-boosting’ members of one’s own group is familiar to anyone who has spent time on Twitter, where people tend to retweet messages published by those they view as co-partisans or allies [17]. Retweets generally signal endorsement of content and an attempt to signal-boost. More specifically, retweets serve three purposes: to spread tweets, to start a conversation, and to draw attention to the originating user [18, 19]. Retweets thus play an important community-building role.

A tweet is either wholly original content, a quote tweet (which retweets and adds commentary), or a retweet of either an original or quote tweet. We considered retweets of both original and quote tweets in our analyses. We also examined retweeted content. If an original tweet was retweeted, we considered the content and author to be that of the original tweet. If a quote tweet (or a series of quote tweets) was retweeted, then we considered the retweeted content and author to be that of the most recent comment. This preserves the endorsement flavor of retweets: if x says something, y quotes to disagree, and z retweets y’s disagreement, it is likely that z also disagrees with x. Note that the Twitter API functions in such a way that intermediate retweets are not stored: if y retweeted x’s tweet T, and z retweets y’s retweet, the data will show only a retweet of x by z (omitting y).

Network construction and community detection

We generated a retweet network, a weighted directed network where nodes are authors and the weight of an edge from node u to node v represents the number of times that user v retweeted user u. Self-retweeting was discarded. Users that only retweeted but never authored an original tweet were discarded. Retweet networks have been used before to study community engagement and the spread of fake news [13, 14, 20]. We considered only the principal weakly connected component of the network. The full network has ∼ 380K nodes and ∼ 3.6M edges. To test for biases in our data, we did a power law analysis and found that the network follows a statistically significant power law distribution (see S1 Appendix for details).

Using time-lag analysis for users we identified around 500 bots in our data set (see S1 Appendix for details). We did not remove the bots for two reasons. First, the overall small number of bots (500 over ∼ 380K total users identified) gives us confidence that the linguistic results will not be seriously skewed by any oddity in bot content. Second, we note that we have treated bots and non-bot users uniformly, in the sense that bots are themselves only included if they make original tweets as well. These are likely to be unusual bots (or, perhaps, hybrid human/bot accounts), and the fact that they both retweet and post original tweets is reason to believe that they would play the same functional role in bringing together communities as would a similarly situated human user.

Modularity optimization is an unsupervised method used for community detection. The modularity of a network measures the strength with which a network can be divided into groups. The measure works by computing the fraction of edges that fall within a given community minus the expected fraction if edges were distributed at random, but keeping the same (weighted) degree distribution. Therefore, a high modularity indicates that members of a community are unexpectedly bound to each other, holding node centrality fixed and having randomness as a baseline.

The most well studied and standard algorithm for modularity maximization is Louvain, developed by [21]. We used Gephi’s implementation of it because it can handle weighted and directed networks like our retweet network, and allows for different resolutions as developed by [22]. We ran the community detection algorithm including randomization and edge weight. Furthermore, we repeated the implementation multiple times and with different resolution values and results reported in the next section remained consistent.

To characterize the communities we considered the top verified accounts in each cluster, as well as the typical hashtags used by the communities. To confirm that the communities extracted by our network modularity analysis were also thematically unified, we ensured that standard machine learning classifiers, using a variety of approaches, could classify users on content at a level well above chance (see S1 Appendix for detailed results). We then restricted further analyses to users in the top 5 identified communities.

Corpus-based analysis of retweets

To study the corpora of tweets per group, we preprocessed the data to exclude non-English words, characters and symbols, as well as English stop-words. The analysis employed a frequency-based approach modeled on Linguistic Inquiry and Word Count (LIWC) [23, 24], which has proven useful in other recent analyses of COVID-related social media discourse [25, 26] In the interest of open science, we used the R package LIWCalike [27], which imitates and expands the functionality of LIWC. As is standard in LIWC analyses, we did not perform any stemming/lemmatizing of the corpora.

An advantage LIWC is the ability to create and share custom dictionaries for categories of interest. Recent interdisciplinary work has shown that it is possible to extract a moral signal from natural language using various tools [28]. For this analysis, we decided to use the custom Moral Foundations Dictionaries (MFD), which are keyed to various moral concerns. Details of these dictionaries are available here (MFD).

The moral foundations dictionaries measure the number of words in a text associated with care (versus harm), fairness (versus unfairness), authority (versus insubordination), loyalty (versus disloyalty), and sanctity (versus corruption). These domains or ‘foundations’ feature in Moral Foundations Theory [29] and are conceived of as topics towards which individuals are differentially sensitive. MFD includes two sub-dictionaries (one related to virtues, the other to vices) for each foundation.

Time series analysis

To further examine the relationship between groups and tweets over the course of the studied period, we performed a vector autoregression (VAR) analysis of retweets. VAR is an extension of multiple regression that attempts to fit the value of variables at time t using their value at time tl, where l is a chosen lag. VAR is widely used in econometrics [30] and has been used in the study of online time series data to, e.g., investigate the relationship between mass shootings and online interest in gun control and gun purchasing [31]. We used the Python package statsmodels (v0.12.2) for all VAR-related analyses [32].

For the endogenous variable we used retweets of tweets by a particular cluster indexed to the day of retweeting, which is a measure of the influence of each cluster. We chose an a priori lag of 1 day, as the influence of tweets tends to fade quickly. Unlike (e.g.) raw tweet counts or retweet activity—both of which have obvious trend increases across our time series—influence is stationary in each group (AD Fuller test, p ≤ 0.05 uncorrected).

We examined VAR coefficients at lag 1 for the influence of each of the five groups. Using the generated model, we also performed tests for Granger causality between groups. To examine the effect of public health interventions on the relationship between groups, we re-ran our analyses using a publicly available data set of health measures [33], using the aggregated US response as our model.

Results

RQ1: Community characterization and classification

Community characterization

We obtained a network modularity score of 0.608 for our retweet network. The implementation found 231 total communities, with the top five largest communities comprising ∼80% of the population and the top eight largest comprising ∼95%. Manual inspection showed that similar community partitions emerged in repeated implementations of the algorithm, and with different resolution values.

We focused on the top five communities. These not only contain ∼ 80% of nodes, but are responsible for ∼ 90% of retweets. Communities beyond the top five also tended to cluster around non-English-language accounts, which limits the utility of our dictionary-based tools. For all further analyses, we considered the subgraph of our full graph containing members of these five groups, and only content originated and retweeted by those members.

Using representative nodes and popular hashtags for each group, we were able to interpret each cluster and add descriptive labels. Note that in adding these labels, we neither attempt nor purport to offer rigid definitions. Instead, we pick out broad patterns within and between communities that can help orient our thinking and guide the ensuing discussion. Worth reiterating here is that even though we did not limit data collection to the United States, the vast majority of users are U.S. based. With this in mind, the ‘Democratic’ cluster (Blue) includes Democratic politicians, center-left media, and self-identified Democratic partisans. The ‘Republican’ group (Red) similarly contains many Republican politicians (including then-President Donald Trump), self-identified Republican partisans, and right-leaning media outlets, at least some of which contributed to the spread of COVID-related misinformation [34, 35]. ‘Public Health’ (Yellow) is largely made up of independent health professional and public health institutions, including the World Health Organization, the Centre for Disease Control and Prevention, and various European health agencies. This cluster is most clearly distinguished from the ‘Antivax’ community (Black), which features traditional anti-vaccination accounts such as StopVaxTyranny and promotes hashtags like #VaccineRoullete. Note again that we use this label in a descriptive rather than normative sense. Note as well that even though most of the accounts in this cluster are U.S. based, it also features users from Europe and Africa. Finally, and in addition to being the only community that is not dominated by U.S. accounts, the ‘Unorthodox’ group (Green) is considerably more heterogeneous than the previous four; we return to characterizing it further below.

As Fig 1 shows, the retweet network was highly polarized before and after the declaration, and features two distinct alliances. At one pole, we see that Democrats and Public Health organizations are already closely connected and that these connections increase as time passes; likewise, at the other pole, Antivaxxers and Republicans appear to interact a lot initially and even more so after the pandemic declaration. This topography is consonant with various recent studies showing that attitudes toward vaccination and other preventive measures aimed at reducing the spread of COVID-19 are strongly sorted along partisan lines [3638]. In fact, political polarization has been found to be the primary driver of opposition to public health interventions [3941], partially explaining the alliance between Antivaxxers and right-leaning users. Finally, we see that the Unorthodox community sits somewhere between these two poles, but is pulled slightly more toward the Antivax-Republican coalition. Fig 1 also shows a significant increase in engagement with vaccine discourse after the declaration; we discuss this in detail later in Table 2.

Fig 1. Visualisation of the retweet networks before and after the WHO declaration, color-coded by community.

Fig 1

To get a better sense of these communities, Table 1 provides summary statistics for each group. Note that the contribution of each group dissociates somewhat from its size. Despite accounting for only ∼ 8% of nodes in the network, Antivaxxers contributed ∼ 22% of retweets. Similarly, Republicans made up just ∼ 18% of the network but were responsible for ∼ 35% of retweet activity. These results are consistent with findings by [13, 14] and suggest that these groups consist of extremely active and vocal individuals.

Table 1. Summary statistics for the top five communities.
Community name % of nodes % of retweets % of verified users Popular Hashtags Representative Nodes
Democrats ∼ 24% ∼ 20% ∼ 10% #moronpresident, #trumpslump, #gopvirus, #trumpgenocideforprofit, #trumpburialpits JoeBiden, KamalaHarris, SenWarren, BillGates, CNN, nytimes, washingtonpost, ABC, businessinsider, MSNBC, guardian (theguardian), TIME, BBCWorld
Republicans ∼ 18% ∼ 35% ∼ 2% #kungflu, #notest, #boycottchina, #trumpCOVIDgate, #illuminati realDonaldTrump, mikepence, RealCandaceO, WhiteHouse, FoxNews
Unorthodox ∼ 16% ∼ 6% ∼ 3% #AfricansAreNotLabRats, #AfricansAreNotGuineaPigs, #listentotheexperts, #locksouthafricadown, #coronavirusghana BernieSanders, Trevornoah, spectatorindex, jacobinmag, NaomiAKlein, BBCAfrica, DrTedros, News24 (African Media)
Public Health ∼ 13% ∼ 7% ∼ 9% #epidemic, #hepatitisa, #immunoonc, #whatwedoinpharmacy, #scteenvax CDCgov, WHO, EU_Health, UniofOxford, UNICEF, ProfPCDoherty (Nobel Laureate Immunology), VaccinesToday, CDCFlu, newscientist, CEPIvaccines, gavi
Antivaxxers ∼ 8% ∼ 22% ∼ 0.8% #illuminati, #praybig, #notest, #mykidsmychoice, #vaccineroulette stopvaccinating*, StopVaxTyranny*, EpigeneticWhisp*, vaxxplained*, JustSayNo2Vax*, va_shiva, Jimcorrsays

Table notes: To protect user privacy, only public figures (with a ‘verified account’ badge) and suspended/deleted accounts are listed as examples within. (i) % of retweets: These account for both retweeting and being retweeted within the network. (ii) Popular Hashtags: All of these are within the top 15 hashtags that each community used, after preprocessing and using a term frequency-inverse document frequency (tf-idf). (iii) Representative Nodes: Users marked with asterisks (*) have either been suspended or deleted by Twitter at time of writing. The typical reason for suspension is violation of the Twitter Terms of Service.

The first four clearly definable groups are consistent with extant research on automated community detection and characterisation on US-specific tweets during the dawning of the COVID-19 pandemic, circa early 2020 [42]. However, since our data-set covers a longer time frame, and we did not filter non-US tweets, we were able to identity a fifth community we called the ‘Unorthodox’ group. In addition to featuring accounts as diverse as those of Bernie Sanders, Trevor Noah, and BBC Africa; its top hashtags reflect both pro- and anti-vaccination attitudes. To make sense of these mixed impressions, we conducted a more careful inspection of this group’s posts around day 100 of our data set, during which time their activity spiked. We found that this community contains many Africa-based users who became active in response to the suggestion that COVID-19 vaccines should be trialled in Africa [43].

For all groups, the vast majority of accounts were created before December 2019 [Democrats: 96%, Republicans: 88%, Unorthodox: 95%, Health: 96%, Antivaxxers: 86%]. If we compute engagement time of an author as the distance in days between their first and last tweet in our data set, Democrats were active for an average of 16 days, Republicans 17, Unorthodox 8, Public Health 18, and Antivaxxers 23. As one might expect, Antivaxxers and Public Health institutions engaged for a longer time span than the other communities, since they were often tweeting about vaccination issues before the pandemic started. It is also noticeable in Fig 2 that by the end, the conversation is dominated by Democrats and Republicans, hinting at how politicised the issue became (see also [34, 38]).

Fig 2. Daily word count by community.

Fig 2

WHO pandemic declaration at the center.

In addition to differences between groups in total numbers of retweets, there are also substantial variations in who each group retweets. Table 2 below shows the change in absolute numbers and ratios of post- to pre-declaration retweets. Unsurprisingly, the largest increases were in groups signal-boosting their own members. This depended in part on how active the groups were pre-pandemic: Antivaxxers and Public Health increased at the lowest rate, while Republicans retweeted Republicans at a vastly higher rate. There were also notable cross-group interactions, particularly between Antivaxxers, Republicans, and the Unorthodox community. In addition to frequently boosting each other’s signal, both Republicans and Antivaxxers retweeted the Unorthodox more than that the Unorthodox retweeted either of these two groups. One possible explanation for this derives from our earlier observation that the Unorthodox community was specifically concerned with vaccine trials in Africa. Hence, despite expressing some hesitancy, it stands to reason that they only partially endorsed the general scepticism expressed by Antivaxxers and Republicans. While it is difficult to assess the authenticity of Republicans’ and Antivaxxers’ support for the Unorthodox, this asymmetry suggests that Antivaxxers and Republicans were more willing to promote the Unorthodox community’s concerns than vice-versa.

Table 2. Changes in absolute numbers of retweets, expressed in 1000s of tweets, with ratio in parentheses.
↓ \ → Democrats Republicans Unorthodox Public Health Antivaxxers
Democrats 452 (3.4) 11 (3.1) 20 (4.1) 29 (3.7) 5 (3.4)
Republicans 7 (4.5) 1151 (8.8) 7 (9.2) 2 (5.5) 129 (9.6)
Unorthodox 11 (2.4) 4 (6.1) 142 (5.2) 3 (2.2) 7 (7.3)
Public Health 18 (2.6) 2 (3.0) 4 (3.7) 108 (2.4) 2 (2.7)
Antivaxxers 2 (3.3) 110 (4.8) 8 (9.8) 3 (2.9) 101 (1.3)

Table notes: Rows (↓) correspond to the retweeting community, while columns (→) correspond to the retweeted community.

WHO pandemic declaration as a threshold

We constructed our data set to be symmetric around the WHO pandemic declaration on March 11th 2020. To further justify this choice, we note that daily word count shows increased engagement across all communities over the course of the pandemic, and begins to spike around the declaration (Fig 2). In addition, we can see that Democrats, Republicans, and the Unorthodox begin to account for a larger share of the conversation, post-declaration. This is suggestive of a significant shift in discourse dynamics.

We also used time series methods to look for a structural break in the data (see S1 Appendix for details). We found evidence of a significant structural break about 5 days after the declaration of the pandemic. Part of this difference might have been due to a lagged response of tweets to the event. However, a more plausible candidate for the cause of the break is the ramp-up of public health measures in response to the pandemic itself.

RQ2: The evolution of vaccine discourse

Our second research question concerned the evolution of vaccine discourse over time. We were interested in whether the moral and non-moral language used by the various communities reflects the patterns of polarization and alliance formation identified in the retweet network.

Linguistic evolution

There were 10 total corpora (5 communities × 2 periods, i.e., pre- and post-declaration). The corpus associated with each group more than doubled in size from pre- to post-pandemic declaration. The largest increase was among Republicans, whose corpus went from 533,620 words to 4,509,393 words, suggesting a surge in interest in a topic that previously had been of relatively little concern to these users.

We use hierarchical clustering to show similarities between the language used by groups before and after the pandemic declaration. For this part of our analysis, each community was associated with a vector corresponding to the tf-idf [44] score for each word.

As Fig 3 suggests, before the pandemic declaration, there were two thematically-unified discourses: one about politics, carried out by Republicans and Democrats, and another about health, carried out by Public Health and Antivaxxers. This doesn’t mean that Democrats and Republicans agreed: rather, they were talking about the same issues in a broadly similar way. Likewise, Public Health and Antivaxxers debated using much the same language.

Fig 3. Dendrogram showing hierarchical clustering for pre- and post-declaration tf-idf vectors.

Fig 3

After the pandemic declaration, however, we observe a reshuffling of the discourses: Democrats and Public Health start using the same language, while Republicans, Antivaxxers, and to some extent the Unorthodox become more linguistically similar to one another as well.

Standard LIWC dictionaries

Using standard LIWC dictionaries, we can shed further light on exactly how the language of these different communities shifted.

The language-based dendrograms in Fig 4 show that patterns in both moral and non-moral discourse evolved to match the patterns of social connection. Initially, the moral and non-moral language used by Antivaxxers and Public Health were most similar to one another, and the moral and non-moral language used by Republicans, Democrats, and Unorthodox were most similar to one another. However, after the pandemic declaration we find Republicans and Antivaxxers expressing the same moral and non-moral concerns, Democrats and Public Health expressing the same moral and non-moral concerns, and the Unorthodox again playing an ambivalent role in between the two polarities.

Fig 4. Dendrogram for LIWC-Specific content.

Fig 4

At a more granular level, Fig 5 shows 15 LIWC components that changed the most from pre- to post-declaration, shedding light on how the discourse evolved during the first few months of the pandemic. Across the board, we see decreases in ‘Female’, ‘Family’, ‘Risk’, ‘Sexual’, and ‘Health’. The first three might be explained by a shift from a vaccine discourse traditionally centred around parental vaccination of children and the perceived risk thereof to one that focuses on the broader context of vaccination. The decrease in ‘Sexual’ is likely due to a comparative drop-off in discussions around human papillomavirus vaccination in particular. The drop-off in discussion of ‘Health’ is somewhat surprising; one possible explanation is that increased polarization means that the issue of vaccination was increasingly framed in political terms.

Fig 5. Changes on selected LIWC components pre- and post-declaration.

Fig 5

X axis shows top 15 absolute score changes. Y axis shows percentage of max pre/post score. Arrow shows direction of change.

In contrast to these convergent trends, Democrats and Public Health score much lower than Antivaxxers and Republicans on the ‘Anger’, ‘Body’, and ‘Feel’ dictionaries post-declaration; suggesting a shift towards a more neutral, dispassionate mode of discussion. Perhaps more strikingly, we see a dramatic increase in discussions surrounding ‘Home’, ‘Money’, and ‘Masculinity’, especially among Republicans and Antivaxxers. The increase in ‘Money’ is likely related to rising unemployment, increasing economic uncertainty, and the NYSE’s March 18th decision to close Wall Street [45, 46]. With respect to ‘Home’, we note that California became the first state to issue stay-at-home orders on the 19th of March and that 15 further states followed suite over the next five days [47]. In response to these restrictions, a cascade of anti-lockdown protests swept across the United States [48].

MFT dictionaries

Using custom Moral Foundations Theory dictionaries, we can further examine the specifically normative discourse each group shows around vaccines. We here discuss selected columns in Fig 6. The full table is available here, or upon request. As with the LIWC dictionaries, the dendrograms in Fig 7 show a linguistic realignment over the course of the study.

Fig 6. Changes in score for moral foundations dictionaries pre- and post-declaration.

Fig 6

Components on X axis absolute score changes. Y axis shows percentage of max pre/post score across all groups (for raw scores, see SM §2.2). Arrow shows direction of change.

Fig 7. Dendrogram for custom MFT virtue dictionaries.

Fig 7

Moving along the five moral foundations from pre- to post-declaration, we observe two patterns that reflect the consolidation of an Antivax-Republican alliance at one end, and a Democrat-Public Health partnership at the other. All groups demonstrate a decrease in the proportion of care-related words, with the largest decrease among Antivaxxers, who end up converging with Republicans in placing the least emphasis on care. On the loyalty foundation, we see that Democrats and Public Health converge in placing additional weight on the virtue and little emphasis on the vice, while Republicans and Antivaxxers move in the opposite direction and place less emphasis on the virtue and more on the vice. Republicans and Antivaxxers also see a relatively large uptick in emphasis on the vices of authority. Taken together, these results suggest that whereas Democrats and Public Health became increasingly concerned with questions of collective responsibility, Antivaxxers and Republicans were less concerned with collective well-being and became progressively more antagonistic towards state and federal authorities. This interpretation dovetails a recent study by [49], which finds that conservatism and anti-vaccination attitudes are both strongly correlated with opposition to public health interventions.

Whereas most communities changed relatively little along the fairness foundation, it is worth noting that Republicans and Antivaxxers again show considerable convergence on both the vice and virtue dimension of this foundation post-declaration. Finally, with respect to sanctity, the virtue and vice dictionaries tend to move in opposite directions. Only the Public Health community increases for both virtue and vice.

In sum, these trends evidence a willingness among but not across both sets of communities to adjust their initial moral emphasis so as to accommodate the values of their closest interlocutors; indicating that the discourse evolved in line with pre-existing patterns of social connection. Interestingly, these dynamics of moral (dis)agreement depart somewhat from traditional models of (de)polarization. While it is true that communities situated at opposing ends of the spectrum move further apart, those who initially emphasize just a few shared foundations seem ready to negotiate their remaining moral differences. Thus, while polarization between poles persists, within each pole, depolarization takes place. On the one hand, this pattern suggests that moral compromise is possible; on the other, it also suggests that local compromise with your closest allies can lead to increased global disagreement [50].

RQ3: Social mechanisms for polarization

The reasons for convergence are likely to be complex and various: all groups are responding both to one another and to ongoing offline events as the pandemic unfolds. The use of VAR sheds light on one set of patterns.

The VAR analysis showed several significant coefficients at p ≤ 0.05 (uncorrected), pictured on the left side of Fig 8. Each coefficient represents the predicted increase in influence of a group at time t given a 1-retweet increase by some group at time t − 1. Tests for Granger causality indicate that all and only the depicted arrows are Granger-causal relationships. Since the number of tweets and influential retweets varies substantially across the groups, the right side of Fig 8 shows the coefficients multiplied by the total influence of the inbound group across the whole time series, presented on a log scale for comparison.

Fig 8. Significant VAR coefficients for the simple lag-1 model.

Fig 8

The left side shows the raw coefficients, the right side shows the log-normalized influence taking into account total influence of each group (see text for details).

The VAR model without exogenous variables shows several interesting patterns of interaction. There are reciprocal interactions between both Republicans and Antivaxxers, and between Democrats and Public Health accounts. While some of the coefficients seem small, the normalized net influence shows that (e.g.) the influence of Antivaxxers on Republicans was comparable to that of Republican self-promotion. So a first important lesson from the time series seems to be that the convergence between groups was driven by the interaction between them, rather than by independent evolution.

Including public health measures as an exogenous variable adds further nuance to the picture. Fig 9 shows the results of incorporating public health measures as an exogenous variable. This is an aggregate measure, focusing on the United States, that includes things such as mask-mandates and stay-at-home orders that were implemented in the weeks following the pandemic declaration.

Fig 9.

Fig 9

Left: significant VAR coefficients for a model incorporating public health measure. Right: log-normalized influence.

Treating public health measures as an exogenous variable preserves some structural relationships—notably, the influence of Democratic users seems to be largely independent of public health measures. On the other hand, Republicans and Antivaxxers (and, less surprisingly, Public Health) gain in influence as public health measures get stronger. The influence of Public Health officials on Republicans and Antivaxxers also switches to strongly negative, suggesting that their messaging was effective in context. Finally, we note that the relationship between Antivaxxers and Republicans becomes asymmetric: once we take public health measures into account, Antivaxxers increase the influence of Republican messaging but not vice-versa.

Discussion

Summary of results

To answer RQ1 we used modularity clustering, an unsupervised method that consistently partitioned the communities into Democrats, Republicans, Public Health, Antivaxxers, and Unorthodox. We also justified the methodology and provided a detailed qualitative and quantitative description of each group. In order to answer RQ2 we used Linguistic Inquiry and Word Count (LIWC) to study the moral and non-moral language used by these communities. Our analysis of both the linguistic and network behavior of these communities shows that two distinct polarized axes—Democrat-Republican and Health-Antivax—converged into a single polarized discourse around vaccination. We call this ‘convergent polarization.’ This polarization is now firmly entrenched as part of the US political landscape [36, 38, 51, 52]. Our response to RQ3 partly explains why this might be.

Recall that to address RQ3 we used vector autoregression analysis (VAR) with the retweet rate of each of the communities as endogenous variables, and public health interventions as the exogenous variable. Our results show that responsiveness to public health measures varied across groups and can explain part of the polarization we observed. Yet the question remains as to why the groups responded differently, which requires reflecting on dynamics of political engagement. On the basis of our observations, we here examine the value-first and the trust-first dynamics, and offer evidence in support of the latter.

It is important to keep in mind that things could have been different. One could imagine (for example) both sides of the political spectrum coming together in opposition to antivaxxers; this was largely the experience in Australia, as well as several European countries [53, 54]. Polarization could have also gone the other way; anti-vaccination groups have, after all, often had a left-leaning component [55]. Given that Donald Trump was spearheading a vaccination push, it would not have been surprising for his followers to rally around him [34, 38]. Finally, it could have been the case for most groups to have intermediate and cross-cutting concerns, as appears to have happened with the Unorthodox community. Yet none of this happened. Why?

One possibility is that shifts in signal-boosting are driven by a values-first dynamic, in which a group is treated as an ally worthy of signal-boosting if and only if they tend to publish information that expresses values shared with your group. In previous work, Haidt and colleagues have found that more politically conservative people and communities tend to place greater emphasis than liberals or leftists do on the so-called ‘binding’ foundations of loyalty, authority, and sanctity [56]. In addition, political conservatives are typically found to be higher in dispositional disgust-sensitivity, which is associated with the sanctity foundation [57] (though see [58] for a dissenting view).

When it comes to COVID-19 and vaccination in particular, it has been hypothesized that normative health behaviors during the pandemic may be partially explained by individual differences in moral foundations. Americans who score high on the care and fairness domains are more likely to report staying at home, support wearing face-masks, and respect social distancing, while those who score high on the sanctity domain are more likely to report wearing face-masks and comply with social distancing, but less likely to limit their movement [59]. Parents who score high on the sanctity foundation are especially likely to be fence-sitters or rejecters when it comes to vaccinating their children [60]. Parents high on the sanctity foundation, low on the authority foundation, and high on the care foundation are more likely to be rejecters [3].

Our data offers only equivocal support for these hypotheses. At least in the context of vaccine discourse, Democrats and Public health appear to score high on (e.g.) sanctity, while Republicans became much more concerned with fairness over the course of the pandemic. Indeed, one of the striking findings of our work is the degree to which groups were willing to change both their moral and non-moral language-use around vaccines over the course of the pandemic. Values, at least to the extent that they can be extracted by linguistic analysis, changed with the emergence of COVID-19 and did not constitute the fix point that explains the polarization.

An alternative hypothesis is that signal-boosting is driven by a trust-first dynamic: groups are treated as allies worthy of signal-boosting if and only if your group has reasons to trust them, regardless of informational veracity. On a trust-first dynamic, information from trusted sources can push one to update what you believe and your commitment to particular values, precisely because you trust your sources [61]. Trust-first dynamics are susceptible to what Begby [62] calls ‘evidential pre-emption’, which occurs when a trusted source also warns that one is likely to encounter misleading contrary evidence. Trust has been theorized as an unquestioning attitude towards testimony [63]. While not intrinsically bad, evidential pre-emption arguably plays an important role in spreading conspiracy theories [15, 64] and supporting online echo chambers [65].

Our data are consistent with a trust-first dynamic. One possible force pulling together Republicans and Antivaxxers is a shared distrust of scientific expertise [51, 52]. Republican mistrust of the scientific establishment dates back at least as far as the 1980s [66], steadily increased in the context of climate change [67], and has become even more pronounced in response to COVID-19 [51, 68]. Furthermore, while conservative self-identification is correlated with vaccine hesitancy, this appears to be mediated by distrust of scientific expertise, vanishing when distrust is controlled for [5]. In contrast to these pre-existing patterns of distrust, Democrats have historically scored high on trust in science [69]. Moreover, and as we would expect if our trust-first hypothesis is correct, Democrats have become more confident in medical experts over the course of the pandemic [70].

Patterns of trust might also shed some light on the intermediate position of the Unorthodox group. Recall that a primary concern of this group was unease at vaccine trials in Africa. There is a well-established pattern of distrust of medical experimentation on Blacks, even on the left, stemming from historical abuses such as the Tuskegee syphilis experiment [71]. It is important to bear in mind that there is no evidence that this community opposes vaccines tout court, or subscribes to a conservative world view. Instead, their distrust for vaccination seems rooted in concerns with issues of colonial and racial injustice. So while partisan attitudes about trust in science might be overall entrenched, the Unorthodox group shows that specific patterns of distrust can cross-cut these broader patterns.

Our explanation coheres with the hypothesis advanced by DiResta & Lotan [72, 73], who argue that Twitter’s content moderation aimed at medical misinformation unified Antivaxxers and Republicans because it prompted Antivaxxers to reframe their message. Rather than straightforward medical misinformation (e.g., ‘Vaccines cause autism’), Antivaxxers began to seek political cover by allying with Republicans and reframing their message in political terms (e.g., ‘Mandatory vaccination is tyranny’). Within the context of COVID-19 therefore, both Twitter and Public Health authorities represented a shared focus of distrust, against which resistance would seem appropriate for some groups.

Limitations and future work

As with all observational work on Twitter, our data collection was limited by what Twitter makes available. Roughly 1% of all tweets [18] are available to researchers without a commercial contract. Our corpus is also mostly in the English language, which we estimate represents 68% of Twitter chatter worldwide pertaining to COVID-19. Future work to address these issues may include augmenting our corpus with multi-lingual and more recent COVID-19 Twitter data sets. Recent work on estimating the distribution of missing tweets from the API stream [74] may also offer a more nuanced picture of any gaps.

As a purely observational study, we are limited in the sorts of causal conclusions we can draw. That said, our work does suggest avenues for future research. We used a proxy measure for public health interventions that aggregated at the level of the entire US. Since many public health measures were implemented state-by-state, coordinating twitter’s (sparse) geolocation data with data about local interventions might shed further light on the influence of local public health measures. Indeed, a recent study by [75] shows that there are important regional differences in partisan responsiveness to public health mandates.

Our data set was limited to tweets which discussed vaccines. An important open question therefore is whether the dynamics of trust that we identify extend beyond vaccine-related discourse. Looking for similar dynamics in different domains, or coordinating distinct actors across different topics, might shed further light on these questions.

Finally, we relied on dictionaries specific to Moral Foundations Theory to make the case for a shift in values. Despite enjoying widespread adoption within academia and even crossing the mainstream into popular culture (e.g. [76, 77]), Moral Foundations Theory has also attracted sustained criticism [78, 79]. Future work might consider other taxonomies of moral reasons, such as Morality as Cooperation [80], or consider using data-driven approaches (such as sentiment analysis) to tease apart the valence of different responses.

Conclusion

In this paper, we showed how the online conversation about vaccines underwent a political realignment as the COVID-19 pandemic progressed. By analyzing networks, language use, and time series data, we demonstrated a realignment of parties around a familiar, politically polarized, set of axes. While these dynamics are largely entrenched by now, they were in flux in the early months of the pandemic. Hence, our research provides a unique window into their formation.

We have argued that this realignment shows that the dynamics of online discourse are driven by social connection and trust, rather than by underlying static values. If so, then trusted informants might be uniquely well positioned to change peoples beliefs. Hence, rather than target hesitant communities with evidence that they are likely to ignore, political resources could be used to identify individuals who are trusted by the target group. Recent reports suggest that this strategy is already gaining traction, at least in the United States, where the Biden administration has partnered with social media ‘influencers’ in an effort to promote vaccination among the countries younger population [81]. Previous work has suggested that personal accounts play an important role in both pro- and anti-vaccination groups on twitter [10]. Furthermore, although reframing was part of the problem, it could be used to signal pro-vaccination messages in terms that are attractive to dissenters: ‘vaccines are good for business’, or ‘protect your body with vaccines’. More recent work on moral and political persuasion has also suggested that personal narratives are more convincing than purely factual accounts precisely because they help build interpersonal trust [82]. Vaccination is a critical public health issue, and disseminating accurate information is vital. Whether accurate information is believed and acted upon, however, is a matter of interpersonal trust—and it is here, we suggest, that there remains substantial work to be done.

Supporting information

S1 Appendix. Supplementary information on the data and methods.

This includes data collection considerations (incl. hashtags), analysis of bots in the dataset, biases in data, classification tasks, and structural break analyses.

(ZIP)

Data Availability

Data are available from Open Science Framework: https://osf.io/b65uc/.

Funding Statement

This paper was supported by Australian Research Council Grant DP190101507 (to Colin Klein). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Larson H, Clarke R, Jarrett C, Eckersberger E, Levine Z, Schulz WS, et al. Measuring trust in vaccination: A systematic review. Human Vaccines & Immunotherapeutics. 2018;14(7):1599–1609. doi: 10.1080/21645515.2018.1459252 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. de Figueiredo A, Simas C, Karafillakis E, Paterson P, Larson H. Mapping global trends in vaccine confidence and investigating barriers to vaccine uptake: A large-scale retrospective temporal modelling study. Lancet. 2020;396(10255):898–908. doi: 10.1016/S0140-6736(20)31558-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Rossen I, Hurlstone M, Dunlop P, Lawrence C. Accepters, fence sitters, or rejecters: Moral profiles of vaccination attitudes. Social Science & Medicine. 2019;224:23–27. doi: 10.1016/j.socscimed.2019.01.038 [DOI] [PubMed] [Google Scholar]
  • 4. Yaqub O, Castle-Clarke S, Sevdalis N, Chataway J. Attitudes to vaccination: A critical review. Social Science & Medicine. 2014;112:1–11. doi: 10.1016/j.socscimed.2014.04.018 [DOI] [PubMed] [Google Scholar]
  • 5. Mesch GS, Schwirian KP. Confidence in government and vaccination willingness in the USA. Health Promotion International. 2015;30(2):213–221. doi: 10.1093/heapro/dau094 [DOI] [PubMed] [Google Scholar]
  • 6.Goldstein DA, Wiedemann J. Who do you trust? The consequences of political and social trust for public responsiveness to COVID-19 orders. The Consequences of Political and Social Trust for Public Responsiveness to COVID-19 Orders (April 19, 2020). 2020.
  • 7. Alfano M, Carter JA, Cheong M. Technological seduction and self-radicalization. Journal of the American Philosophical Association. 2018;4(3):298–322. doi: 10.1017/apa.2018.27 [DOI] [Google Scholar]
  • 8. Alfano M, Carter JA, Ebrahimi Fard A, Clutton P, Klein C. Technologically scaffolded atypical cognition: The case of YouTube’s recommender system. Synthese. 2020;(1–2):1–24. [Google Scholar]
  • 9. Burr C, Cristianini N, Ladyman J. An analysis of the interaction between intelligent software agents and human users. Minds and Machines. 2018;28:735–774. doi: 10.1007/s11023-018-9479-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Dunn A, Leask J, Zhou X, Mandl K, Coiera E. Associations between exposure to and expression of negative opinions about human papillomavirus vaccines on social media: An observational study. Journal of Medical Internet Research. 2015;17(6):e144. doi: 10.2196/jmir.4343 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Dunn A, Surian D, Leask J, Dey A, Mandl K, Coiera E. Mapping information exposure on social media to explain differences in HPV vaccine coverage in the United States. Vaccine. 2017;35(23):3033–3040. doi: 10.1016/j.vaccine.2017.04.060 [DOI] [PubMed] [Google Scholar]
  • 12. Nowak SA, Chen C, Parker AM, Gidengil CA, Matthews LJ. Comparing covariation among vaccine hesitancy and broader beliefs within Twitter and survey data. PLOS ONE. 2020;15(10):1–16. doi: 10.1371/journal.pone.0239826 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Sullivan E, Sondag M, Rutter I, Meulemans W, Cunningham S, Speckmann B, et al. Can real social epistemic networks deliver the wisdom of crowds? In: Lombrozo T, Knobe J, Nichols S, editors. Oxford Studies in Experimental Philosophy. Oxford University Press; 2020. [Google Scholar]
  • 14. Sullivan E, Sondag M, Rutter I, Meulemans W, Cunningham S, Speckmann B, et al. Vulnerability in social epistemic networks. International Journal of Philosophical Studies. 2020;28(5):731–753. doi: 10.1080/09672559.2020.1782562 [DOI] [Google Scholar]
  • 15. Klein C, Clutton P, Dunn AG. Pathways to conspiracy: The social and linguistic precursors of involvement in Reddit’s conspiracy theory forum. PLOS ONE. 2019;14(11):1–23. doi: 10.1371/journal.pone.0225098 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Shah Z, Surian D, Dyda A, Coiera E, Mandl K, Dunn A. Automatically appraising the credibility of vaccine-related web pages shared on social media: A Twitter surveillance study. Journal of Medical Internet Research. 2019;21(11). doi: 10.2196/14007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Heilig LL. Signal Boost!: Hashtags as performative writing and social action; 2015.
  • 18.Cheong M. Inferring social behavior and interaction on Twitter by combining metadata about users & messages; 2013.
  • 19.Boyd D, Golder S, Lotan G. Tweet, tweet, retweet: Conversational aspects of retweeting on Twitter. In: Proceedings of the 43rd Hawaii International Conference on System Sciences; 2010.
  • 20. Bovet A, Makse H. Influence of fake news in Twitter during the 2016 US presidential election. Nature Communications. 2019;10(7). doi: 10.1038/s41467-018-07761-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Blondel VD, Guillaume JL, Lambiotte R, Lefebvre E. Fast unfolding of communities in large networks. Journal of Statistical Mechanics: Theory and Experiment. 2008;(10):P10008. doi: 10.1088/1742-5468/2008/10/P10008 [DOI] [Google Scholar]
  • 22. Lambiotte R, Delvenne JC, Barahona M. Laplacian Dynamics and Multiscale Modular Structure in Networks. Physics and Society. 2009;1(2):76–90. [Google Scholar]
  • 23. Pennebaker J. The secret life of pronouns: What our words say about us. Bloomsbury: Bloomsbury Press; 2011. [Google Scholar]
  • 24.Pennebaker J, Boyd R, Jordan K, Blackburn K. The development and psychometric properties of LIWC2015. University of Texas at Austin; 2015.
  • 25. Ashokkumar A, Pennebaker JW. Social media conversations reveal large psychological shifts caused by COVID-19’s onset across US cities. Science advances. 2021;7(39):eabg7843. doi: 10.1126/sciadv.abg7843 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Negri A, Andreoli G, Barazzetti A, Zamin C, Christian C. Linguistic markers of the emotion elaboration surrounding the confinement period in the Italian epicenter of COVID-19 outbreak. Frontiers in Psychology. 2020; p. 2464. doi: 10.3389/fpsyg.2020.568281 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Benoit K, Watanabe K, Wang H, Nulty P, Obeng A, Mueller S, et al. quanteda: An R package for the quantitative analysis of textual data. Journal of Open Source Software. 2018;3(30):774. doi: 10.21105/joss.00774 [DOI] [Google Scholar]
  • 28. Alfano M, Higgins A, Levernier J. Identifying virtues and values through obituary data-mining. Journal of Value Inquiry. 2018;52(1):59–79. doi: 10.1007/s10790-017-9602-0 [DOI] [Google Scholar]
  • 29. Graham J, Haidt J, Koleva S, Motyl M, Iyer R, Wojcik S, et al. Moral foundations theory: The pragmatic validity of moral pluralism. Advances in Social Psychology. 2013;47:55–130. doi: 10.1016/B978-0-12-407236-7.00002-4 [DOI] [Google Scholar]
  • 30. Stock JH, Watson M. Introduction to econometrics. Pearson; 2020. [Google Scholar]
  • 31. Gunn LH, Ter Horst E, Markossian TW, Molina G. Online interest regarding violent attacks, gun control, and gun purchase: a causal analysis. PLoS one. 2018;13(11):e0207924. doi: 10.1371/journal.pone.0207924 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Seabold S, Perktold J. Statsmodels: Econometric and statistical modeling with python. In: 9th Python in Science Conference. Austin, TX; 2010.
  • 33. Porcher S. A novel dataset of governments’ responses to COVID-19 all around the world. Scientific Data. 2020;7(423). doi: 10.1038/s41597-020-00757-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Hart PS, Chinn S, Soroka S. Politicization and polarization in COVID-19 news coverage. Science Communication. 2020;42(5):679–697. doi: 10.1177/1075547020950735 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Motta M, Stecula D, Farhart C. How right-leaning media coverage of COVID-19 facilitated the spread of misinformation in the early stages of the pandemic in the US. Canadian Journal of Political Science/Revue canadienne de science politique. 2020;53(2):335–342. doi: 10.1017/S0008423920000396 [DOI] [Google Scholar]
  • 36. Allcott H, Boxell L, Conway J, Gentzkow M, Thaler M, Yang D. Polarization and public health: Partisan differences in social distancing during the coronavirus pandemic. Journal of public economics. 2020;191:104254. doi: 10.1016/j.jpubeco.2020.104254 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Grossman G, Kim S, Rexer JM, Thirumurthy H. Political partisanship influences behavioral responses to governors’ recommendations for COVID-19 prevention in the United States. Proceedings of the National Academy of Sciences. 2020;117(39):24144–24153. doi: 10.1073/pnas.2007835117 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Druckman JN, Klar S, Krupnikov Y, Levendusky M, Ryan JB. How affective polarization shapes Americans’ political beliefs: A study of response to the COVID-19 pandemic. Journal of Experimental Political Science. 2021;8(3):223–234. doi: 10.1017/XPS.2020.28 [DOI] [Google Scholar]
  • 39. Gadarian SK, Goodman SW, Pepinsky TB. Partisanship, health behavior, and policy attitudes in the early stages of the COVID-19 pandemic. Plos one. 2021;16(4):e0249596. doi: 10.1371/journal.pone.0249596 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Makridis C, Rothwell JT. The real cost of political polarization: Evidence from the COVID-19 pandemic. Available at SSRN 3638373. 2020.
  • 41. Wu JD, Huber GA. Partisan differences in social distancing may originate in norms and beliefs: Results from novel data. Social Science Quarterly. 2021;102(5):2251–2265. doi: 10.1111/ssqu.12947 [DOI] [Google Scholar]
  • 42. Jiang J, Chen E, Lerman K, Ferrara E. Political Polarization Drives Online Conversations About COVID-19 in the United States. Hum Behavior and Emerging Technology. 2020;2(3):200–211. doi: 10.1002/hbe2.202 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Rosman R. Racism row as French doctors suggest virus vaccine test in Africa. Aljazeera. 2020;(Retrieved from https://www.aljazeera.com/news/2020/4/4/racism-row-as-french-doctors-suggest-virus-vaccine-test-in-africa).
  • 44. Salton G, McGill MJ. Introduction to modern information retrieval. McGraw-Hill; 1983. [Google Scholar]
  • 45.Jackson L. Wall Street plunges to worst level in 12 years. Reuters. 2020;(Retrieved from https://www.nst.com.my/business/2020/03/576662/wall-street-plunges-worst-level-12-years).
  • 46.Muccari R, Chow D, Murphy J. Coronavirus timeline: Tracking the critical moments of Covid-19. NBC News. 2020;(Retrieved from https://www.nbcnews.com/health/health-news/coronavirus-timeline-tracking-critical-moments-covid-19-n1154341).
  • 47.Neuman S. California Issues’Stay At Home’ Order As Coronavirus Infections Rise. NPR. 2020;(Retrieved from https://www.npr.org/2020/03/20/818764136/california-issues-stay-at-home-order-as-coronavirus-infections-rise).
  • 48.Brennan E. Coronavirus anti-lockdown movement surges in the US after Donald Trump’s’Liberate’ tweet. ABC News. 2020;(Retrieved from https://www.abc.net.au/news/2020-05-27/coronavirus-us-protests-on-the-rise/12288686).
  • 49. Perry SL, Whitehead AL, Grubbs JB. Save the economy, liberty, and yourself: Christian nationalism and Americans’ views on government COVID-19 restrictions. Sociology of Religion. 2021;82(4):426–446. doi: 10.1093/socrel/sraa047 [DOI] [Google Scholar]
  • 50. Uslaner E. The Moral Foundations of Trust. Cambridge: Cambridge University Press; 2002. [Google Scholar]
  • 51. Whitehead A, Perry S. How Culture Wars Delay Herd Immunity: Christian Nationalism and Anti-vaccine Attitudes. Socius: Sociological Research for a Dynamic World. 2020;6(1):1–12. [Google Scholar]
  • 52. Kerr J, Panagopoulos C, van der Linden S. Political polarization on COVID-19 pandemic response in the United States. Personality and Individual Differences. 2021;179:110892. doi: 10.1016/j.paid.2021.110892 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53. Bernacer J, García-Manglano J, Camina E, Güell F. Polarization of beliefs as a consequence of the COVID-19 pandemic: The case of Spain. PloS one. 2021;16(7):e0254511. doi: 10.1371/journal.pone.0254511 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54. Jungkunz S. Political polarization during the COVID-19 pandemic. Frontiers in Political Science. 2021;3:622512. doi: 10.3389/fpos.2021.622512 [DOI] [Google Scholar]
  • 55. McCoy CA. The social characteristics of Americans opposed to vaccination: Beliefs about vaccine safety versus views of US vaccination policy. Critical Public Health. 2020;30(1):4–15. doi: 10.1080/09581596.2018.1501467 [DOI] [Google Scholar]
  • 56. Graham J, Haidt J, Nosek B. Liberals and conservatives rely on different sets of moral foundations. Journal of Personality and Social Psychology. 2009;96(5):1029–1046. doi: 10.1037/a0015141 [DOI] [PubMed] [Google Scholar]
  • 57. Inbar Y, Pizarro D, Bloom P. Conservatives are more easily disgusted than liberals. Cognition and Emotion. 2009;23(4):714–725. doi: 10.1080/02699930802110007 [DOI] [Google Scholar]
  • 58. Elad-Strenger J, Proch J, Kessler T. Is disgust a “conservative” emotion? Personality and Social Psychology Bulletin. 2020;46(6):896–912. doi: 10.1177/0146167219880191 [DOI] [PubMed] [Google Scholar]
  • 59. Chan E. Moral foundations underlying behavioral compliance during the COVID-19 pandemic. Personality and Individual Differences. 2020;171:110463. doi: 10.1016/j.paid.2020.110463 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60. Amin A, Bednarczyk R, Ray CE, Melchiori K, Graham J, Huntsinger J, et al. Association of moral values with vaccine hesitancy. Nature Human Behavior. 2017;1:873–880. doi: 10.1038/s41562-017-0256-5 [DOI] [PubMed] [Google Scholar]
  • 61. Levy N, Alfano M. Knowledge from vice: Deeply social epistemology. Mind. 2019;129(515):887–915. doi: 10.1093/mind/fzz017 [DOI] [Google Scholar]
  • 62. Begby E. Evidential preemption. Philosophy and Phenomenological Research. 2021;102(3):515–530. doi: 10.1111/phpr.12654 [DOI] [Google Scholar]
  • 63. Nguyen CT. Trust as an unquestioning attitude. In: Gendler T, Hawthorne J, editors. Oxford Studies in Epistemology. Oxford University Press; forthcoming. [Google Scholar]
  • 64. Sunstein CR, Vermeule A. Conspiracy theories: Causes and cures. Journal of Political Philosophy. 2009;17(2):202–227. doi: 10.1111/j.1467-9760.2008.00325.x [DOI] [Google Scholar]
  • 65. Nguyen CT. Echo chambers and epistemic bubbles. Episteme. 2020;17(2):141–161. doi: 10.1017/epi.2018.32 [DOI] [Google Scholar]
  • 66. Herman E, Chomsky N. Manufacturing Consent: Political Economy of the Mass Media. New York: Pantheon Books; 1988. [Google Scholar]
  • 67. Hamilton L, Hartter J, Saito K. Trust in Scientists on Climate Change and Vaccines. Sage Open. 2015;5(3):1–13. doi: 10.1177/2158244015602752 [DOI] [Google Scholar]
  • 68. Motta M. Republicans, Not Democrats, Are More Likely to Endorse Anti-Vaccine Misinformation. American Politics Research. 2021;49(5):428–438. doi: 10.1177/1532673X211022639 [DOI] [Google Scholar]
  • 69. Gauchat G. Politicization of Science in the Public Sphere: A Study of Public Trust in the United States, 1974 to 2010. American Sociological Review. 2012;77(2):167–187. doi: 10.1177/0003122412438225 [DOI] [Google Scholar]
  • 70.Funk C, Kennedy B, Johnson C. Trust in medical scientists has grown in US, but mainly among democrats. 2020.
  • 71.Jones JH. Bad blood: The Tuskegee syphilis experiment; 1993.
  • 72.DiResta R, Lotan G. Anti-vaxxers are using Twitter to manipulate a vaccine bill. Wired Magazine. 2015;(Retrieved from https://www.wired.com/2015/06/antivaxxers-influencing-legislation/).
  • 73.DiResta R. Anti-vaxxers think this is their moment. The Atlantic. 2020;(Retrieved from https://www.theatlantic.com/ideas/archive/2020/12/campaign-against-vaccines-already-under-way/617443/).
  • 74.Wu S, Rizoiu MA, Xie L. Variation across Scales: Measurement Fidelity under Twitter Data Sampling. In: International AAAI Conference on Web and Social Media (ICWSM’20); 2020.
  • 75. Druckman JN, Klar S, Krupnikov Y, Levendusky M, Ryan JB. Affective polarization, local contexts and public opinion in America. Nature human behaviour. 2021;5(1):28–38. doi: 10.1038/s41562-020-01012-5 [DOI] [PubMed] [Google Scholar]
  • 76.Brooks D. The end of philosophy. The New York Times. 2009;(Retrieved from https://www.nytimes.com/2009/04/07/opinion/07Brooks.html).
  • 77.Wade N. Is “do unto others” written into our genes? The New York Times. 2007;(Retrieved from https://www.nytimes.com/2007/09/18/science/18mora.html).
  • 78. Suhler CL, Churchland P. Can Innate, Modular “Foundations” Explain Morality? Challenges for Haidt’s Moral Foundations Theory. Journal of Cognitive Neuroscience. 2011;29(3):2103–2016. doi: 10.1162/jocn.2011.21637 [DOI] [PubMed] [Google Scholar]
  • 79. Gray K, Keeney JE. Disconfirming Moral Foundations Theory on Its Own Terms: Reply to Graham. Social Psychological and Personality Science. 2015;6(8):874–877. doi: 10.1177/1948550615592243 [DOI] [Google Scholar]
  • 80. Curry O, Chesters M, Van Lissa C. Mapping morality with a compass: Testing the theory of ‘morality-as-cooperation’ with a new questionnaire. Journal of Research in Personality. 2019;78:106–124. doi: 10.1016/j.jrp.2018.10.008 [DOI] [Google Scholar]
  • 81.Kelly J. Thousand-Dollar Cash Payments And A TikTok ‘Influencer Army’ Are Part Of The Campaign To Get People Vaccinated. Forbes. 2021;(Retrieved from https://www.forbes.com/sites/jackkelly/2021/08/04/thousand-dollar-cash-payments-and-tiktok-influencer-army-are-part-of-the-campaign-to-get-people-vaccinated/?sh=486b1d624684).
  • 82. Kubin E, Puryear C, Schein C, Gray K. Personal experiences bridge moral and political divides better than facts. Proceedings of the National Academy of Sciences. 2021;118(6):e2008389118. doi: 10.1073/pnas.2008389118 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Tingshao Zhu

23 May 2022

PONE-D-21-40016Polarization and trust in the evolution of vaccine discourse on Twitter during COVID-19PLOS ONE

Dear Dr. Ojea Quintana,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jul 07 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Tingshao Zhu

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf  and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please ensure that you refer to Figure 6 in your text as, if accepted, production will need this reference to link the reader to the figure.

3. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments (if provided):

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The article has been written meticulously and addresses an important public health issue of developed and developing nations alike using an innovative approach. In my opinion all the findings and interpretations are scientifically sound as presented in the paper.

Reviewer #2: The manuscript is written in very prominent contemporary debate. Scientifically written in logical flows. "We have argued that this realignment shows that the dynamics of online discourse are driven by social connection and trust, rather than by underlying static values." Agreed.

Reviewer #3: The authors describe an analysis of the debate about COVID-19 on Twitter over the first five months of the pandemic. The final aim is to study English-language discourse around vaccines, vaccination-related engagement and discourse, and the causes of the changes in engagement. They describe the data collected, their approach, and the observations derived from the analyses. The authors focus their research on the top five communities they extracted (as they represent most of the English-language data they collected), summarising their statistics, and focusing all of their analyses on these groups. The authors provide detailed descriptions of the outcomes of their study by answering their research questions and properly discussing the limitations of their approach.

The article is well organised, and the outcomes are discussed thoroughly. I advise providing additional context and explanations for the following points.

- In the data collection part, when mentioning “...a series of vaccination-related keywords, hashtags, and short expressions…”, it would be good to have a few examples of such words/expressions (even if they are included in the Appendix) with a small discussion on how and why such words were chosen (maybe you could pick the most relevant and/or complex ones).

- The authors discarded “Users that only retweeted but never authored an original tweet” while keeping “...a very small number of bots”. Both users with retweets only and bots may play a signal-boosting function. Discussing the reasons behind the decision of keeping one category while discarding the other may improve the soundness of the article. Moreover, how different would the network be without bots’ data? Do bots really change the shape of the network or the interactions between the communities?

- The authors considered only the top five communities, mainly because they share content using the English language while representing 80% of the nodes and 90% of the tweets. How and why did the authors choose only the top five? What is the choice criterion? Was it only because of the language used within the identified communities?

- The authors state “Communities beyond the top five also tended to cluster around non-English-language accounts, which limit the utility of our dictionary-based tools”. Did the authors discard all the non-English content the top five communities shared? I advise discussing such a point and carrying out a quick analysis of the percentages of English content VS non-English for the top five (or even 10) groups to cover such an aspect, also providing further support for the discussion about the choice of the top five communities.

Furthermore, the authors mention the Appendix many different times within the text, although no Appendix is provided within the article.

Reviewer #4: 1. The article is very clear and well defined. But still some corrections are presents.

2. Authors need to focus the novelty in abstract section and expected output with actual output.

3. introduction section is too short to intro the proposed work motto.

4. Result section is very organized but images are unclear. Need to upgrade with high resolution images.

5. Try to use some recent papers as related work (2020-2022).

6. There are some grammatical correction, please recheck and correct.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Dr. Aftab Ahmad, MD (Community Medicine)

Reviewer #2: No

Reviewer #3: Yes: Andrea Tocchetti

Reviewer #4: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Dec 14;17(12):e0277292. doi: 10.1371/journal.pone.0277292.r002

Author response to Decision Letter 0


4 Jun 2022

Please see the attached letter in response to reviewers, where we address each of the helpful comments in detail.

Attachment

Submitted filename: Response to Reviewers - PLOS ONE.pdf

Decision Letter 1

Tingshao Zhu

29 Jul 2022

PONE-D-21-40016R1Polarization and trust in the evolution of vaccine discourse on Twitter during COVID-19PLOS ONE

Dear Dr. Ojea Quintana,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please see reviewer comments below.

Please submit your revised manuscript by Sep 11 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Hanna Landenmark

Staff Editor, PLOS ONE

on behalf of 

Tingshao Zhu

Academic Editor, PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments (if provided):

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #3: All comments have been addressed

Reviewer #4: All comments have been addressed

Reviewer #5: All comments have been addressed

Reviewer #6: (No Response)

Reviewer #7: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #3: Yes

Reviewer #4: Yes

Reviewer #5: No

Reviewer #6: Yes

Reviewer #7: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #3: Yes

Reviewer #4: Yes

Reviewer #5: No

Reviewer #6: Yes

Reviewer #7: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #3: Yes

Reviewer #4: Yes

Reviewer #5: No

Reviewer #6: Yes

Reviewer #7: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #3: Yes

Reviewer #4: Yes

Reviewer #5: No

Reviewer #6: Yes

Reviewer #7: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #3: The authors presented a study analysing the impact of COVID-19 on the vaccine discourse over the first 5 months of pandemic on Twitter. They clearly stated the research questions they address and provide detailed analyses covering such questions. The authors properly describe the methods, data, and procedures employed in the process.

Reviewer #4: Authors are successfully addressed all of the comments from reviewer. No further corrections are identified.

Reviewer #5: The topic presented by the authors is very interesting and timely. However, there are several major issues that need to be addressed.

1. The abstract is very vage and not very clear. Please revise in a way the reader knows exactly what to expect from the paper.

2. I like highlighting RQ1 - RQ3 on page 2. However, I suggest to be more specific.

RQ1. Which groups are most important in the English-language discourse around 40 vaccines on Twitter? I do not think that is what the authors do. Instead, the focus is on specific countries using the English language. Which countries are these? Add to RQ1!

RQ2. How did vaccination-related engagement and discourse change over the first 42 five months of the pandemic? Add years, ie. 02/2000 - 07/2000 or similar.

RQ3. What social forces might help explain observed changes in engagement? should be

RQ3. What social forces might help in explaining observed changes in engagement?

In general, the language needs to be improved of the paper. In its current form it is very poor language at many places and I cannot list all of them.

As explained in the sections coming (just another example of poor language).

3.To identify different communities, we used Gephi’s implementation of the Louvain 110 modularity algorithm developed by [22] This is unclear. Please discuss community/module detection algorithms in details, e.g.,

https://link.springer.com/article/10.1186/s12859-016-0979-8

4. Discussion: From the beginning it is unclear to me how to define 'Antivaxxers'? I searched the paper but could not find a proper definition (it should pop out immediately).

Question: Is the population of a country with a low vaccination rate (e.g. some African countries) in the category Antivaxxers? I is unclear to me if this is a negative term or neutral and if one has a choice to be in this group or is this possible determined by a third party.

In this context, it would be good to add some English-speaking African countries with a low vaccination rate and low death rate which I think exist. If the focus is only on the US a discussion of this would be sufficient.

Considering this the distinction Health-Antivax seem incorrect. I would suggest a revision of wording to make the presentation more factual.

4. From line 407 to end: I find the connection between the numerical results and the provided discussion unsatisfactory. The main problem is a lack of connection. The authors do not use the numerical results to provide a discussion but the discussion seems not well grounded at all. At least I could not see such a connection.

It is very important that the discussion focuses exclusively on the interpretation of the numerical results. At the end of the discussion, a wider perspective, which could be even speculative, could be provided.

I suggest to rewrite the entire discussion section.

5. Our corpus is also mostly in the English language, 481 which we estimate represents 68% of Twitter chatter worldwide pertaining to COVID. How can this information be relevant when Democrat-Republican are limited to the US only?

Similar to the point about Health-Antivax (see 4) also

Democrat-Republican is not well defined. Definitions of all 4 terms should be added to the methods section.

6. The conclusion section is similar unclear and lacks a connection to the numerical results.

Reviewer #6: This paper presents an original and as of yet unpublished study. The authors have clearly and effectively described their analyses and drawn appropriate (and quite interesting) conclusions from the analyses presented. The article is also well written and understandable, with one exception which stood out to me while reading the manuscript.

The dataset obtained from the Twitter Streaming API appears to be global, with no mention of any geographic restrictions from users, and several non-US users are listed, with the particular grouping of the "unorthodox" being non-American in nature. However, two of the communities are defined in terms of American political divisions, the public health measure is US-centered, and geolocation data are mentioned in the discussion section when addressing future work. I think it may be helpful to clarify briefly but explicitly either in the data collection subsection or in the community characterization subsection that, even though the labels "Republican" and "Democrats" are used for two communities, the data extend globally. As a secondary, optional, matter I would be interested to know why American political cleavages may define a global discussion space.

Reviewer #7: This paper reports an interesting analysis of the discourse on vaccines around the onset of the COVID-19 pandemic. In particular, it shows how the discourse changed after the WHO officially declared the pandemic in March 2021. Overall, I think that it is well done and makes a significant contribution.

Although I review the paper for the first time, I see that it is a resubmission. Therefore, I limit my critical comments to points that can be reasonably addressed at this stage of the review process.

First, I encourage the authors to clarify the geographic scope of the analysis. The dataset includes only tweets written in English, but is not restricted to particular countries. At the same time, the analysis relies on the categories of US politics, particularly the distinction between Republicans and Democrats. This tension should be clarified as early as possible in the paper.

Second, the link between vaccines and COVID was a bit confusing to me, given the time frame of the analysis (December 2019-June 2020). In that period, the emphasis was not on COVID vaccines, but more on the nature of the disease and mitigation strategies such as masks and containment. The authors do frame the study as an analysis of changes of the vaccination discourse following the emergence of COVID, but they could be more explicit that the discourse is on vaccines _in general_. Currently, readers may have the wrong expectation that the paper is about COVID vaccines.

Third, the pre-post comparisons are the most interesting parts of the analysis, but they could be discussed more in depth. For example, we see in Figure 1 that the structure of the retweet network changes significantly, but the point is barely elaborated in the text.

Finally, Figures 5 and 7 are not legible, partly because of the low resolution but also due to the lack of a legend and the small size of the points. These figures should be improved.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #3: Yes: Andrea Tocchetti

Reviewer #4: Yes: F.M. Javed Mehedi Shamrat

Reviewer #5: No

Reviewer #6: No

Reviewer #7: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Dec 14;17(12):e0277292. doi: 10.1371/journal.pone.0277292.r004

Author response to Decision Letter 1


15 Aug 2022

Please find attached a version of the document copied below.

Dr Ignacio Ojea Quintana and coauthors

Australian National University

Canberra

ACT 2600

Australia

August 10th, 2022.

Re: Response to Reviewers, PLOS ONE / PONE-D-21-40016

Dear Hanna Landenmark and Tingshao Zhu, on behalf of the Editors and Reviewers,

We thank you, the Editorial Board, and the panel of Reviewers for your generous time in the review process of our paper and for providing us with helpful comments on our manuscript, “Polarization and trust in the evolution of vaccine discourse on Twitter during COVID-19”.

On the following pages, please find our detailed responses to address the comments raised by the reviewer panel.

Thank you for your time and consideration.

Kind regards,

Dr Ignacio Ojea Quintana

on behalf of all coauthors

(Ignacio Ojea Quintana, Ritsaart Reimann, Marc Cheong, Mark Alfano, Colin Klein)

Encl. (Responses to the Review Report)

-----------------------------------------------------------------------

Responses to the Review Report

-----------------------------------------------------------------------

Reviewer #3:

The authors presented a study analysing the impact of COVID-19 on the vaccine discourse over the first 5 months of pandemic on Twitter. They clearly stated the research questions they address and provide detailed analyses covering such questions. The authors properly describe the methods, data, and procedures employed in the process.

- We thank the reviewer for their charitable comments.

------------------------------------------------------------------------

Reviewer #4:

Authors are successfully addressed all of the comments from reviewer. No further corrections are identified. (sic.)

- We are glad to have addressed the comments, and we thank the reviewer for them since they genuinely improved the essay.

--------------------------------------------------------------------------

Reviewer #5:

The topic presented by the authors is very interesting and timely. However, there are several major issues that need to be addressed.

We thank the reviewer for their comments.

In the next paragraphs we will address them one by one.

1. The abstract is very vague and not very clear. Please revise in a way the reader knows exactly what to expect from the paper.

- The abstract was modified to improve clarity. It now reads:

“Trust in vaccination is eroding, and attitudes about vaccination have become more polarized. This is an observational study of Twitter analyzing the impact that COVID-19 had on vaccine discourse. We identify the actors, the language they use,how their language changed, and what can explain this change.

First, we find that authors cluster into several large, interpretable groups, and that the discourse was greatly affected by American partisan politics. Over the course of our study, both Republicans and Democrats entered the vaccine conversation in large numbers, forming coalitions with Antivaxxers and public health organizations, respectively. After the pandemic was officially declared, the interactions between these groups increased. Second, we show how the moral and non-moral language used by the various communities converged in interesting and informative ways. Finally, vector autoregression analysis indicates that differential responses to public health measures are likely part of what drove this convergence. Taken together, our results suggest that polarization around vaccination discourse in the context of COVID-19 was ultimately driven by a trust-first dynamic of political engagement.”

The abstract states our three main contributions: (i) an observational study of which communities were identified, (ii) an analysis of how their language changed, and (iii) an explanatory hypothesis using vector autoregression, complemented with another hypothesis about what drove the dynamics.

In the body of the paper we also made an effort to be more precise in the description of the techniques used and the numerical results. But since all other reviewers found the original abstract sufficiently clear, we do not want to make substantial changes to the original version.

2. I like highlighting RQ1 - RQ3 on page 2. However, I suggest to be more specific.

RQ1. Which groups are most important in the English-language discourse around vaccines on Twitter? I do not think that is what the authors do. Instead, the focus is on specific countries using the English language. Which countries are these? Add to RQ1!

RQ2. How did vaccination-related engagement and discourse change over the first five months of the pandemic? Add years, ie. 02/2000 - 07/2000 or similar.

RQ3. What social forces might help explain observed changes in engagement? should be

RQ3. What social forces might help in explaining observed changes in engagement?

- We thank the reviewer for asking us to sharpen our research questions. With respect to RQ2 and RQ3, the suggested changes have been incorporated. With respect to RQ1, we have added some comments to clarify that we did not limit data collection to the United States nor any other territory, and that the prevalence of specific regions within our dataset is an artifact of the interaction between our methodology and the distribution of Twitter’s user-base. In particular, we point out that the prevalence of U.S-based users within our analysis is partly due to our focus on English-language discourse; partly due to the fact that the vast majority of Twitters English speaking user-base is located in the United States; and partly a reflection of the extent to which American discourse defines global online discussions, at least when those discussions are carried out in English.

In general, the language needs to be improved of the paper. In its current form it is very poor language at many places and I cannot list all of them.

- As explained in the sections coming (just another example of poor language).

In the next few responses we explain thow we clarified the language used and the structure of the essay in a way that is clearly tied with the observations we made.

3.To identify different communities, we used Gephi’s implementation of the Louvain 110 modularity algorithm developed by [22] This is unclear. Please discuss community/module detection algorithms in details.

- Many thanks for this comment.

The Materials and Methods section now includes a better explanation of what modularity is and the algorithm used.

In fact, in the original version of the essay we provided some description but we decided to remove it for the first R&R, so we are sympathetic to the comment of the reviewer. We now included a description in some detail of how the unsupervised method works. Nevertheless, we decided not to present the mathematical and algorithmic aspects in excessive detail because we regard the method as standard in the literature, and we do not want the readers to get stuck in unnecessary details.

4. Discussion: From the beginning it is unclear to me how to define 'Antivaxxers'? I searched the paper but could not find a proper definition (it should pop out immediately).

Question: Is the population of a country with a low vaccination rate (e.g. some African countries) in the category Antivaxxers? I is unclear to me if this is a negative term or neutral and if one has a choice to be in this group or is this possible determined by a third party.

In this context, it would be good to add some English-speaking African countries with a low vaccination rate and low death rate which I think exist. If the focus is only on the US a discussion of this would be sufficient.

Considering this the distinction Health-Antivax seem incorrect. I would suggest a revision of wording to make the presentation more factual.

Similar to the point about Health-Antivax also, Democrat-Republican is not well defined. Definitions of all 4 terms should be added to the methods section.

- We thank the reviewer for asking us to more carefully define each group, and antivaxxers in particular. To avoid confusion, we have added a definition of ‘Antivaxxers’ to the introduction, specifying that we use this term in a descriptive rather than normative sense: it is our interpretation of a community of users whose top hashtags and accounts display anti-vaccination attitudes. We also note that whether or not users belong to this cluster is not contingent on their geographic location; for as explained in the Network construction and community clustering section, community membership is determined by social interaction (e.g., retweeting).

The distinctions between Democrats and Republicans on the one hand and Antivaxxers and Public Health organizations on the other are addressed in more detail in the ‘community characterization and classification’ section. We emphasize that since we are dealing with large and diverse populations of users about whom we only have limited information, our definitions are informed by broad patterns of similarities and differences within and between the various communities.

With respect to the U.S focused nature of our analysis, we have added a section to clarify that we did not limit data collection to the United States, and that the prevalence of American politics within our analysis is an artifact of the interaction between our methodology and the distribution of Twitter’s English-speaking user-base.

Finally, with respect to the reviewers request to add English-speaking African countries, we note that these are already included in the original paper. In particular, our analysis of the unorthodox community makes it clear that there are African-based users involved in the discourse, and that these users express both pro- and anti-vaccination attitudes.

5. From line 407 to end: I find the connection between the numerical results and the provided discussion unsatisfactory. The main problem is a lack of connection. The authors do not use the numerical results to provide a discussion but the discussion seems not well grounded at all. At least I could not see such a connection.

It is very important that the discussion focuses exclusively on the interpretation of the numerical results. At the end of the discussion, a wider perspective, which could be even speculative, could be provided. I suggest to rewrite the entire discussion section.

- The discussion section is now reorganized and rewritten in light of the reviewer’s comments. Two notes on this.

First, we sympathize with the reviewer's objection that some of what we said there was not transparently tied with the observational results that we presented before. We now organize the material in a way that is justified by the results. In particular, our discussion of the values-first vs trust-first dynamics is warranted by the observations we did in our linguistic analysis, in particular the use of moral language by different communities.

Second, the purpose of the discussion section is to provide a broader theoretical background in order to explain the observations made. For this reason, that section builds on and discusses some contemporary literature.

6. Our corpus is also mostly in the English language, which we estimate represents 68% of Twitter chatter worldwide pertaining to COVID. How can this information be relevant when Democrat-Republican are limited to the US only?

- We now clarify this important point in the essay.

The prevalence of U.S-based users within our analysis is partly due to our focus on English-language discourse; partly due to the fact that the vast majority of Twitter’s English speaking user-base is located in the United States (68%); and partly a reflection of the extent to which American discourse defines global online discussions, at least when those discussions are carried out in English. This explains the brute fact that random sampling using English words will most likely give an over-representation of USA discourse and dynamics, which is what we found.

-----------------------------------------------------------------------------------

Reviewer #6:

This paper presents an original and as of yet unpublished study. The authors have clearly and effectively described their analyses and drawn appropriate (and quite interesting) conclusions from the analyses presented. The article is also well written and understandable, with one exception which stood out to me while reading the manuscript.

The dataset obtained from the Twitter Streaming API appears to be global, with no mention of any geographic restrictions from users, and several non-US users are listed, with the particular grouping of the "unorthodox" being non-American in nature. However, two of the communities are defined in terms of American political divisions, the public health measure is US-centered, and geolocation data are mentioned in the discussion section when addressing future work. I think it may be helpful to clarify briefly but explicitly either in the data collection subsection or in the community characterization subsection that, even though the labels "Republican" and "Democrats" are used for two communities, the data extend globally. As a secondary, optional, matter I would be interested to know why American political cleavages may define a global discussion space.

- We thank the reviewer for raising this point, and have clarified that despite identifying two clearly U.S. based clusters of users, our data extend globally. This point is made briefly in the introduction, and then elaborated in the community characterization and classification section. We note that since we did not limit data collection to the U.S. nor any other territory, the prevalence of American politics within our analysis is partly due to our focus on English-language discourse; partly due to the fact that the vast majority of Twitters English speaking user-base is located in the United States; and partly a reflection of the extent to which American political cleavages define global online discussions, at least when those discussions are carried out in English. Hence, even though we distinguish 'Democratic' and 'Republican' clusters of users, it is not the case that all users in our data set are based in the United States, for neither Public Health Institutions nor Antivaxxers are unique to U.S. discourse.

-----------------------------------------------------------------

Reviewer #7:

This paper reports an interesting analysis of the discourse on vaccines around the onset of the COVID-19 pandemic. In particular, it shows how the discourse changed after the WHO officially declared the pandemic in March 2021. Overall, I think that it is well done and makes a significant contribution.

Although I review the paper for the first time, I see that it is a resubmission. Therefore, I limit my critical comments to points that can be reasonably addressed at this stage of the review process.

- We thank the reviewer for their comments.

First, I encourage the authors to clarify the geographic scope of the analysis. The dataset includes only tweets written in English, but is not restricted to particular countries. At the same time, the analysis relies on the categories of US politics, particularly the distinction between Republicans and Democrats. This tension should be clarified as early as possible in the paper.

- We thank the reviewer for this remark, and have added a section to the introduction to clarify the geographic scope of our analysis. This section also addresses the apparent tension between finding two distinctly U.S. based clusters and the global nature of our data set. In particular, we note that since we did not limit data collection to the U.S. nor any other territory, the prevalence of American politics within our analysis is partly due to our focus on English-language discourse; partly due to the fact that the vast majority of Twitter’s English speaking user-base is located in the United States; and partly a reflection of the extent to which American political cleavages define global online discussions, at least when those discussions are carried out in English. Hence, even though we distinguish 'Democratic' and 'Republican' clusters of users, it is not the case that all users in our data set are based in the United States, for neither Public Health Institutions nor Antivaxxers are unique to U.S. discourse.

Second, the link between vaccines and COVID was a bit confusing to me, given the time frame of the analysis (December 2019-June 2020). In that period, the emphasis was not on COVID vaccines, but more on the nature of the disease and mitigation strategies such as masks and containment. The authors do frame the study as an analysis of changes of the vaccination discourse following the emergence of COVID, but they could be more explicit that the discourse is on vaccines _in general_. Currently, readers may have the wrong expectation that the paper is about COVID vaccines.

- We thank the reviewer for this comment, it is important that the scope and purpose of the paper is clearly understood.

In order to clarify that the scope of the paper is how vaccine discourse in general changed due to Covid, we made modifications both to the Abstract and the Introduction, so that it becomes clear from the beginning.

Third, the pre-post comparisons are the most interesting parts of the analysis, but they could be discussed more in depth. For example, we see in Figure 1 that the structure of the retweet network changes significantly, but the point is barely elaborated in the text.

- There are two central points about the before and after networks that are emphasized in the text now. First, that even before the pandemic communities already exhibited some polarization. Second, that engagement increased substantially after the pandemic declaration. We develop the second point in more detail through the study, in Table 2, of the ratios and total numbers of retweet behavior.

Finally, Figures 5 and 7 are not legible, partly because of the low resolution but also due to the lack of a legend and the small size of the points. These figures should be improved.

- We now increased the resolution (dpi) of both figures, increased the font and point sizes in order to make those figures more legible. One problem is that image quality is greately reduced in the compilation of the revision document on this webpage. The images submitted have a greater dpi and are clearly legible. We do not know what else we can do in this regard.

Attachment

Submitted filename: Reviews 2nd R&R.pdf

Decision Letter 2

Hossein Kermani

25 Oct 2022

Polarization and trust in the evolution of vaccine discourse on Twitter during COVID-19

PONE-D-21-40016R2

Dear Dr. Ojea Quintana,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Hossein Kermani

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Congratulations!

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #4: All comments have been addressed

Reviewer #7: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #4: Yes

Reviewer #7: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #4: Yes

Reviewer #7: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #4: Yes

Reviewer #7: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #4: Yes

Reviewer #7: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #4: All comments have been addressed by the authors.

Still some grammatical errors exits, check and resolve the mistakes.

Reviewer #7: The authors have addressed my comments satisfactorily. I have no other comments.

Some additional words to meet word count.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #4: No

Reviewer #7: No

**********

Acceptance letter

Hossein Kermani

21 Nov 2022

PONE-D-21-40016R2

Polarization and trust in the evolution of vaccine discourse on Twitter during COVID-19

Dear Dr. Ojea Quintana:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Hossein Kermani

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix. Supplementary information on the data and methods.

    This includes data collection considerations (incl. hashtags), analysis of bots in the dataset, biases in data, classification tasks, and structural break analyses.

    (ZIP)

    Attachment

    Submitted filename: Response to Reviewers - PLOS ONE.pdf

    Attachment

    Submitted filename: Reviews 2nd R&R.pdf

    Data Availability Statement

    Data are available from Open Science Framework: https://osf.io/b65uc/.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES