Skip to main content
PLOS One logoLink to PLOS One
. 2024 Oct 23;19(10):e0305148. doi: 10.1371/journal.pone.0305148

News source bias and sentiment on social media

Brian Knutson 1,*, Tiffany W Hsu 1, Michael Ko 1, Jeanne L Tsai 1
Editor: Silvio Eduardo Alvarez Candido2
PMCID: PMC11498708  PMID: 39441772

Abstract

As social media becomes a key channel for news consumption and sharing, proliferating partisan and mainstream news sources must increasingly compete for users’ attention. While affective qualities of news content may promote engagement, it is not clear whether news source bias influences affective content production or virality, or whether any differences have changed over time. We analyzed the sentiment of ~30 million posts (on twitter.com) from 182 U.S. news sources that ranged from extreme left to right bias over the course of a decade (2011–2020). Biased news sources (on both left and right) produced more high arousal negative affective content than balanced sources. High arousal negative content also increased reposting for biased versus balanced sources. The combination of increased prevalence and virality for high arousal negative affective content was not evident for other types of affective content. Over a decade, the virality of high arousal negative affective content also increased, particularly in balanced news sources, and in posts about politics. Together, these findings reveal that high arousal negative affective content may promote the spread of news from biased sources, and conversely imply that sentiment analysis tools might help social media users to counteract these trends.

Introduction

Social media has changed how people consume and share news. Over half of adults in the United States report consuming at least some of their news online, either through news or social media sites (e.g., twitter.com, facebook.com) [1]. Social media news channels confer benefits, since the internet offers faster access to a broader range of content and platforms, which can promote sharing and discussion of information with others. These benefits may accompany costs, however, including the rapid dissemination and promotion of misinformation capable of sowing distress, division, and even death [2]. Thus, improved understanding of how news sources spread information and misinformation might eventually augment efforts to increase societal happiness, harmony, and health [3].

While some United States news sources offer more mainstream or balanced coverage, others feature more partisan or biased coverage. Concurrent with deepening political divisions, partisan news outlets have recently proliferated [4, 5]. Partisan news outlets are marked not only by subjective perceptions of bias from users, but also by objective evidence that they spread more misinformation [6, 7]. This misinformation may spread even faster and wider via news outlets’ social media accounts. Compared with traditional media, social media platforms can confer virality through sharing, engagement, and comments [8, 9]. Thus, as news sources on social media grow, so may the spread of misinformation.

Social media news content can vary with respect to several attributes which can promote spread. Specifically, semantics (what was said) and source (who said it) can promote spread [10]. Beyond semantics and source, however, emerging evidence suggests that sentiment (or how it was said) can also command users’ attention and engagement [5, 8, 9]. Which exact type of sentiment promotes the spread of online news content, however, remains unclear [1113]. Further, research has not examined whether the impact of sentiment on news virality varies across a continuum of balanced versus biased news sources over time.

In this research, we sought to address two gaps in existing work. First, although some studies of online news affective content and virality have separately focused on either mainstream or partisan news sources [9], fewer have explicitly compared the two [14, 15]. Such a comparison is critical for understanding how partisan versus mainstream news outlets use sentiment to compete for users’ attention online. While some studies of mainstream news outlets report that articles with high arousal affect (e.g., from The New York Times) are more likely to be reposted [11], other studies of partisan news content have reported that neither overall emotional strength [13] nor sensationalism (i.e., intentionally using extreme emotions and shock) predict engagement with news articles (e.g., on Facebook) [16]. One study found that negative emotional political messages were more likely to be reposted on social media, but did not distinguish between partisan and mainstream sources [17]. Another study reported that either positive or negative emotional language increased reposting about partisan topics within political ingroups (on Twitter), although the specific types of affect (e.g., positive or negative) which drove reposting depended on the topic [12]. To clarify and resolve these emerging findings, we directly compared affective content produced and spread by biased versus balanced news sources.

Second, while most research has focused on affect categories (e.g., positive versus negative sentiment, general emotional strength), more specific types of affective content that vary with respect to valence (positive or negative) as well as arousal (high or low arousal) may have distinct effects on the spread of content on social media. For example, some studies have reported an increased prevalence and virality of highly arousing negative emotional content (e.g., hate and outrage) relative to other types of emotional content in United States news content on social media [14, 18, 19] and online [20]. Another study demonstrated that misinformation (i.e., verified by fact-checking organizations) evoked more fearful, disgusted, and surprised reactions in social media users [9]. These findings imply that some types of negative aroused affective content may facilitate the faster and further spread of misinformation relative to information. Neuroeconomic research further implies that highly aroused positive and negative affect which typically occur during anticipation of uncertain outcomes promote subsequent attentional engagement and even action [21]. Since few studies have systematically compared all combinations of affective valence and arousal in social media news content, we examined whether the virality of affective content depends its valence and arousal. This approach [22] might help clarify inconsistent findings, while deepening mechanistic insights into which types of affective content are most prevalent and most powerfully drive engagement.

We addressed these questions by assessing which types of affective content were most prevalent and viral in social media posts of biased versus balanced news sources on social media [23]. To do so, we culled nearly 30 million posts from 182 U.S. news sources on a social media site (known as twitter.com during execution of this research) over the course of a decade (i.e., from 2011 to 2020). These brief posts typically contained headlines and summaries designed to entice users to read more extensive linked stories at affiliated news sites. News sources varied across a broad range of political orientations, from liberal/left-biased (e.g., Wonkette), to moderate/neutral (e.g., PBS News Hour), to conservative/right-biased (e.g., Breitbart) news sources. The social media platform under study (i.e., twitter.com) served as a news distribution platform and a forum for discussion, as well as a site of apparent ideological polarization [10] during this period. Using a modified version of the SentiStrength algorithm [19, 24], we quantified whether posts from these news sources contained four distinct types of affective content: (1) High Arousal Negative (or HAN; e.g., anxiety), (2) Low Arousal Negative (or LAN; e.g., dull), (3) Low Arousal Positive (or LAP; e.g., calm), and (4) High Arousal Positive (or HAP; e.g., excitement) affective content. Using this coded data, we addressed two key questions. First, does news source bias account for the prevalence of different types of affective content produced in posts? Second, is the affective content of a post associated with the extent to which it spreads (or is “viral” with respect to repost count), and do these associations vary for biased versus balanced news sources? We also explored whether these patterns generalized across time and news category (e.g., politics versus entertainment).

Given users’ high engagement with biased news sources and the suspected virality of arousing affective content [14, 15], we predicted that biased news sources would produce more high arousal negative affective content, and that high arousal negative affective content would be more viral than other types of affective content. Given increasing partisan polarization in the United States [4, 5], we further predicted that the influence of news source bias on affective content prevalence and spread might increase over time and generalize across news categories.

Materials and methods

Data collection

We collected data from 215 news sources that varied with respect to political bias based on scores obtained by the non-partisan company Ad Fontes Media (adfontesmedia.com). Ad Fontes Media recruited individuals who vary across the political spectrum to rate a set of articles published by each news source for political bias using a highly-resolved scoring system (–42 = extreme left to 0 = neutral to +42 = extreme right). These scores were then aggregated to create a general political bias score for each news source [25]. In this study, we used scores from the adfontes.com website obtained on February 9, 2021.

Since this work focuses on political bias in the United States (U.S.), we excluded nineteen news sources that were not based in the U.S. and eight news sources that did not explicitly address politics. We also excluded two news sources that were cable versions of news sources already in the database (i.e., CNN Cable TV News Network, Fox News TV Cable Network), two news sources that lacked accounts on twitter.com (i.e., Big League Politics, Rush Limbaugh), and two news sources that had been officially banned from the platform for providing false content (i.e., InfoWars, The Gateway Pundit). Thus, the final dataset for analysis included a total of 182 news sources. The mean bias across these news sources leaned slightly to the political left (Mean = –1.75, SD = 14.67, range = –28.24 to 33.78).

To convergently validate news source bias scores, we extracted independent bias scores for an available subsample of overlapping news sources (n = 126) from allsides.com (accession date: Nov 20, 2021). These bias scores (ranging from 1 or leftmost to 5 or rightmost) were coded using a combination of editorial reviews, blind bias surveys, independent reviews, and third-party research. Despite differences in the derivation of measures, the correlation between Ad Fontes and AllSides bias scores was robust (r(125) = 0.89, p < .001; ρ(125) = 0.88, p < .001), supporting convergent validity of the news sources bias scores (Fig 1).

Fig 1. Pairwise associations of AdFontes quality and bias scores with AllSides bias scores (parametric (r) and nonparametric (ρ) linear correlations; for 126 matched news sources from 182 total analyzed news sources).

Fig 1

Using the Twitter Application Programming Interface (API) Academic Research Track, which allowed full-archive access during the study period, we collected posts posted by each news source on its official Twitter account over ten years (January 1, 2011 to December 31, 2020) to examine annual change. Data collection and processing complied with the terms and conditions imposed by twitter.com (now x.com). Because the Twitter repost (or “retweet”) function was launched in late 2009, we excluded the year 2010 and began data collection in 2011 to ensure that the repost function had fully taken effect throughout the analyzed dataset. This yielded a total of 29,946,524 posts across 182 analyzed news sources.

Sentiment analysis and categorization

To assess the affective content of message sentiment [19], we used a modified version of the SentiStrength algorithm [24]. SentiStrength relies on an English dictionary set that includes terms labeled with sentiment strength (valence and intensity), as well as semantically relevant terms including booster words (e.g., extremely, somewhat, quite), negating words (e.g. not, don’t, couldn’t), question words (e.g., why, how), emoticons (e.g., Inline graphic, (^ ^), Inline graphic), slang words (e.g., lol, bff), and domain-specific terms (e.g., “must watch” in the context of film). The program then optimizes the term labels with machine learning, developed on a set of human-labeled social media web texts (i.e., six types of social media, including posts) [26] (Fig 2).

Fig 2. Modified Sentistrength analysis flowchart (adapted from Hsu et al., 2021).

Fig 2

Raw posts were preprocessed, scored for positive and negative intensity, and assigned affective labels.

This version of the SentiStrength algorithm offered several advantages in coding sentiment relative to other algorithms. First, it is optimized to detect sentiment in short social media web texts and has been previously applied to content from Twitter [27] as well as news-related social media content [28]. Second, SentiStrength codes affective intensity separately from valence (i.e., each text is assigned a score of 1 = not positive to 5 = extremely positive and a score of 1 = not negative to 5 = extremely negative). Although intensity and arousal are not theoretically identical, they are often highly correlated in self-report measures, allowing SentiStrength to assume similarity. For instance, SentiStrength assigns calm and other low arousal positive states an intensity rating of 2, whereas it assigns excitement and other high-arousal positive states an intensity rating of at least 3. Further, exclamation marks receive higher intensity ratings. This version of the SentiStrength algorithm also includes updated emojis [19]. Posts were initially transformed to lower-case characters to support recognition of affective words in capitalized headlines.

To derive measures of affective content that varied in terms of both valence and arousal, posts were coded as containing: “Low Arousal Positive [LAP]” content if they received a SentiStrength positivity score of 2; “High Arousal Positive [HAP]” content if they received SentiStrength positivity scores of 3, 4, or 5.; “Low Arousal Negative [LAN]” content if they received SentiStrength negativity scores of 2; and “High Arousal Negative [HAN]” content if they received SentiStrength negativity scores of 3, 4, and 5. Since words psychometrically associated with “high arousal” (e.g., “excitement”) were assigned a score of 3 by SentiStrength, we used scores of 3 and above to indicate high arousal affect. Thus, single posts could contain mixed affective content. Examples of posts containing different types of affect content include HAP (pos = 5, neg = 1): “‘Jaw-droppingly exciting’: Scientists discover seven Earth-like planets”; LAP (pos = 2, neg = 1): “Teaching kids to care, one book at a time”; LAN (pos = 1, neg = 2): “FDA reportedly planning stricter guidelines for release of COVID-19 vaccine”; and HAN (pos = 1, neg = 4): “Intruder shot to death with his own gun after barging into home, police say” (also see word clouds of post examples for each affect category in S1 Fig):

Results

Biased news sources produce different types of affective content

We first examined whether specific types of affective content were more prevalent overall. We calculated the overall prevalence of each affective type by first aggregating the percentage of posts that contained each type of affective content (%Affectk) for each news source by year (e.g., the percentage of posts posted by ABC in 2020 that contained HAN affective content), then aggregating across years for each news source (e.g., the percentage of posts posted by ABC across the decade that contained HAN affective content), and finally aggregating across news sources. To compare different affect types, we fit mixed linear regression models using affect type to predict percentage with random intercepts of news source and year.

Analyses revealed that news sources posted more negative content (46.55%) than positive content (25.90%) overall (with 37.72% of the posts being classified as neutral). In terms of specific affect types, news sources posted low arousal negative content the most (LAN content: 25.84%), followed by high arousal negative content (HAN content: 20.71%), low arousal positive content (LAP content: 16.55%), and high arousal positive content (HAP content: 9.35%) (all pairwise comparisons between affect types were significant, Ps < .001). Thus, news sources posted almost twice as much negative as positive affective content overall (see S1 File for all regression specifications and results).

Affective content prevalence by news source bias

To test for an association of news source bias with affect prevalence, we calculated the prevalence of each type of affective content (%Affectk) as described above for each news source aggregated across years. Then four linear models used the signed bias (Bias) and absolute bias (|Bias|) scores of each news source to predict the prevalence of each type of affective content, with a random intercept of year (i.e., one model for each k = {HAN, LAN, LAP, HAP}). We included the absolute value of political bias because political bias scores were bidirectional (i.e., negative = left-leaning, positive = right-leaning). The coefficient of the absolute bias score |Bias| quantified the effect of the magnitude of political bias in either direction on the prevalence of each affect type k (with a positive coefficient indicating that more political bias was associated with greater prevalence of affect type k). The coefficient of the signed bias score Bias tested whether bias direction mattered (with a positive coefficient indicating that more right-leaning political bias predicted greater prevalence of affect type k, and a negative coefficient indicating that more left-leaning political bias predicted greater prevalence of affect type k).

Analyses revealed that absolute bias (|Bias|) of the news source had the largest effect on the prevalence of affective content. News sources that were more biased in either direction (i.e., higher in |Bias|) posted more HAN content. On average, a 1.0% increase in political bias predicted a 0.378% increase in the prevalence of HAN content (b = 0.378, SE = 0.017, P < 0.001) (Fig 3A). This association indicated that the most biased news sources in both directions (most extreme left bias = –28.24, most extreme right bias = 33.78) posted HAN content in approximately 11% more posts than the most balanced news sources. More biased news sources also posted more of other types of affective content, including LAN (0.05%; b = 0.049, SE = 0.014, P = 0.001), HAP (0.14%; b = 0.145, SE = 0.014, P < 0.001), and LAP content (0.12%; b = 0.119, SE = 0.012, P < 0.001). All these effects were smaller, however, than that observed for HAN content.

Fig 3. Post affective content prevalence as a function of news source bias and year.

Fig 3

A) Effect of news source bias on the percentage of posts containing each affect type. Bands indicate 95% confidence intervals. Stars indicate significant coefficient for the absolute bias term in the model (** P < .01, *** P < .001). B) Effect of absolute political bias on the percentage of posts containing each affect type across years. Bars indicate 95% confidence intervals for linear fits. Stars indicate significant interaction between year and absolute bias (*** P < .001). Dashed lines represent the point at which absolute bias has no effect on the percentage of posts containing each affect type.

Analysis of signed political bias (Bias) indicated that more left-biased news sources posted more HAN (–0.07%; b = –0.068, SE = 0.001, P < 0.001), HAP (–0.04%; b = –0.042, SE = 0.008, P < 0.001), and LAP content (–0.02%; b = –0.024, SE = 0.007, P = 0.001), while more right-biased news sources posted more LAN content (0.02%; b = 0.019, SE = 0.008, P = 0.020), but these effects were again smaller in magnitude (by a half or less) than the effect of absolute bias of news sources on the percentage of posts containing HAN content.

Thus, more biased news sources on both the left and right posted more high arousal negative (HAN) content. To a lesser extent, more biased news sources also posted more of other types of affective content (i.e., LAN, HAP, and LAP), relative to more balanced news sources.

Affective content prevalence over time

Next, we examined whether these associations changed over time. Specifically, based on the association of high arousal negative affective content with political polarization in the United States [12, 15], we predicted that more politically biased news sources would post more HAN content over the course of a decade, but did not have predictions about other types of affective content.

To test for temporal trends, we modified the models above to replace the random intercept of year with a variable coded as Year since 2011 (year 2011 = 0, year 2012 = 1, …, year 2020 = 9) and interaction terms of Year with Bias and |Bias|. As predicted, a significant positive interaction between Year and |Bias| (b = 0.020, SE = 0.006, P < 0.001) emerged, indicating that more biased news sources posted more posts with HAN content over time (Fig 3B), relative to more balanced news sources. Importantly, this interaction was unique to HAN content, since the interaction of Year with |Bias| was not significant for the other three types of affective content (LAN: b = 0.007, SE = 0.005, P = 0.183; LAP: b = 0.001, SE = 0.004, P = 0.769; HAP: b = 0.002, SE = 0.005, P = 0.616). Thus, HAN content was the only type of affective content that biased news sources produced more of over time relative to balanced news sources. For signed political bias, there was an unpredicted and less significant positive interaction of Year with Bias and HAP content (b = 0.008, SE = 0.003, P = 0.004), suggesting that more right-biased news sources posted increasingly more HAP content over time.

These analyses revealed that more biased news sources posted more high arousal negative affective content compared to more balanced news sources over time (i.e., from 2011 to 2020; Fig 3A). Although overall more biased news sources had also posted more affective content of other types, those associations were less pronounced and did not change over time (Fig 3B).

Biased news sources spread different types of affective content

Repost rates were standardized by calculating repost rate of a given post relative to other posts from the same news source within the same year. For example, if ABC posted 4 posts with repost counts 10, 20, 30, and 40 in year 2011, the post with repost count 30 would have a repost count percentile of 3 (its rank) / 4 (4 posts total) = 0.75. Thus, the indexed post would have been reposted equal to or greater than 75% of all posts posted by ABC in 2020.

We first examined whether posts with different types of affective content were reposted more often than others by fitting four linear models (one for each type of affective content k = {HAN, LAN, LAP, HAP}) that used presence of each affect type (Containsk 0 = no, 1 = yes) to predict each post’s repost count percentile. Post news source and posted year were modeled as random intercepts. As in previous research, whether the post contained media (e.g., pictures, videos) or not was also included as a covariate, since inclusion of media tends to increase spread [12, 15]. For each model, the estimate associated with predictor Containsk captured the average difference in repost count percentile of posts that contained k versus posts that did not contain k. For example, if the estimate for HAN content was 2.00, posts containing HAN content were on average 2 percentiles higher in repost count than other posts.

Affective content spread by news source bias

To examine whether these effects varied as a function of news source bias, we fit linear mixed effects models for each type of affective content k = {HAN, LAN, LAP, HAP}. These predicted repost count percentile as a function of affect type (Containsk 0 = no, 1 = yes), news source bias (both signed Bias and absolute |Bias|), and their interaction (Containsk x Bias + Containsk x |Bias|). Random intercepts of the news source and of each post’s posted year were again included, along with the covariates described above. As in previous analyses, the signed coefficient represented the direction of bias, while the absolute coefficient represented the magnitude of bias.

Overall, posts that contained HAN content were reposted 3.27 percentiles more than posts that lacked HAN content (b = 3.272, SE = 0.013, P < .001), consistent with previous findings demonstrating a viral effect of HAN content in United States users (Fig 4) [19]. As predicted, posts with HAN content were reposted more for biased than for balanced news sources. Specifically, while posts containing HAN content were reposted 1.46 percentiles more among balanced news sources (simple effect of HAN at bias = 0 b = 1.457, SE = 0.022, P < .001), they were reposted by an additional 0.179 percentiles with each percent of increase in bias magnitude (ContainsHAN x |Bias| interaction: b = 0.179, SE = 0.002, P < .001). Also, with each percent decrease in signed bias (i.e., more leftward), posts containing HAN content were reposted by an additional 0.009 percentiles (ContainsHAN x Bias interaction b = –0.009, SE = 0.001, P < .001). Together, these effects imply that for the leftmost sources (bias = –28.24), posts containing HAN content were reposted 5.30 percentiles more than posts without HAN content, and for the rightmost sources (bias = 33.78), posts containing HAN content were reposted 5.73 percentiles more than posts without HAN content (Fig 4).

Fig 4. Post affective content spread as a function of news source bias and year.

Fig 4

A) Effect of news source bias on the average repost count percentile of posts containing each affect type. Bands indicate 95% confidence intervals for linear fits. Stars indicate significant coefficient for the absolute value term in the model (*** P < .001). B) Effect of absolute political bias in repost count percentile posts containing each affect type across years. Bars indicate 95% confidence intervals. Stars indicate significant interaction between year and absolute bias (*** P < .001). Dashed lines represent the point at which absolute bias has no effect on the percentage of posts containing each affect type.

Analyses also revealed some unexpected patterns for other types of affective content. For instance, posts containing LAN content were reposted 0.698 percentiles more than posts without LAN content (b = 0.698, SE = 0.020, P < .001), and these were also reposted more for biased news sources (ContainsLAN x |Bias| interaction: b = 0.114, SE = 0.002, P < .001). While similar to the effect of HAN content on reposting, the effect of LAN content on reposting was smaller, and more prominent for biased sources on the left (ContainsLAN x Bias interaction b = –0.028, SE = 0.001, P < .001). In contrast, posts containing LAP and HAP content were reposted less than posts containing other content (LAP: b = –1.340, SE = 0.024, P < .001; HAP: b = –1.600, SE = 0.032, P < .001). Posts containing LAP content were reposted more by biased media sources, particularly on the right (ContainsLAP x |Bias| interaction: b = 0.037, SE = 0.002, P < .001; ContainsLAP x Bias interaction: b = 0.008, SE = 0.001, P < .001). This pattern of findings might cohere with previous results suggesting that partisan news can trigger positive emotions supporting the ingroup to boost political engagement [29]. Posts containing HAP content, however, were instead reposted less by biased news sources (ContainsHAP x |Bias| interaction: b = –0.031, SE = 0.003, P < .001; ContainsHAP x Bias interaction: b = 0.019, SE = 0.002, P < .001).

Together, these findings supported predictions that posts containing HAN content would be reposted more than posts containing other types of affective content, and that this effect would be stronger for more biased news sources. Consistent with predictions, more biased news sources not only posted more HAN content, but also showed greater spread for posts containing HAN content. A similar but weaker pattern was apparent for posts containing LAN content.

Affective content spread over time

Based on the previously-documented association of negative emotional content with political polarization [12, 15, 18], we predicted that the effect of news source bias on reposts of posts containing HAN content might increase over time, but did not have specific predictions for other affect types.

To examine spread over time, we modified models of affective content to include a Year variable (coded as year since 2011: year 2011 = 0, year 2012 = 1, …, year 2020 = 9) that generated three-way interactions (with Containsk x Bias and Containsk x |Bias|). We therefore predicted a significant positive interaction for Year x ContainsHAN x |Bias|. While this interaction was significant, contrary to prediction, its coefficient weight was negative rather than positive (Year x ContainsHAN x |Bias| interaction: b = –0.027, SE = 0.000, P < .001), implying that the effect of news source bias on the repost count percentile of posts containing HAN content decreased over years. A similar but smaller interaction was also apparent for LAN content (Year x ContainsLAN x |Bias| interaction: b = –0.014, SE = 0.001, P < .001). This interaction was significantly positive for both LAP and HAP content, however, suggesting that the effect of news source bias on the repost count percentile of posts containing LAP or HAP content increased over years (LAP: Year x ContainsLAP x |Bias| interaction: b = 0.018, SE = 0.001, P < .001; HAP: Year x ContainsHAP x |Bias| interaction: b = 0.006, SE = 0.001, P < .001).

This negative effect of news bias on reposts of posts containing HAN content over time did not fit predictions. According to one account, reposts of posts containing HAN content might have decreased for biased news sources over time. According to an alternative account, however, reposts of posts containing HAN content instead might have instead increased for balanced news sources over time. To distinguish between these accounts, we split news sources into more “balanced” versus “biased” sets using a split on the median |Bias| score (i.e., 10.67). Next, we fit linear mixed effects models that predicted repost count percentile as a function of whether the post contained HAN content (ContainsHAN 0 = no, 1 = yes), year (coded as year since 2011: year 2011 = 0, year 2012 = 1, …, year 2020 = 9), and the interaction of affect content with year. Random intercepts for news sources were included, along with the covariates described above (Fig 4B).

Findings supported the second account, since the effect of HAN content on repost count percentile did not change over years for more biased news sources, but instead increased over years for more balanced news sources (biased: Year x ContainsHAN: b = –0.001, SE = 0.008, P = 0.896; balanced: Year x ContainsHAN: b = 0.402, SE = 0.007, P < .001) (Fig 4). For other affect types, LAN content also spread less for more biased news sources, but spread more for more balanced news sources (biased: Year x ContainsLAN: b = –0.153, SE = 0.008, P < .001; balanced: Year x ContainsLAN: b = 0.100, SE = 0.006, P < .001). Opposite patterns emerged, however, for the remaining affect types. Specifically, LAP content spread more for more biased news sources, but spread less for more balanced news sources (biased: Year x ContainsLAP: b = 0.135, SE = 0.009, P < .001; balanced: Year x ContainsLAP: b = –0.128, SE = 0.007, P < .001), and HAP content also spread more for more biased news sources but did not change over time for more balanced news sources (biased: Year x ContainsHAP: b = 0.120, SE = 0.011, P < .001; balanced: Year x ContainsHAP: b = –0.003, SE = 0.009, P = 0.723) (Fig 5).

Fig 5. Post affective content spread as a function of year grouped by biased (dashed) versus balanced (solid) news sources (split by median of absolute bias).

Fig 5

Bars indicate 95% confidence intervals for linear fits. Stars indicate significant interaction between year and absolute bias (*** P < .001). Gray dashed lines represent the point at which absolute bias has no effect on the percentage of posts containing each affect type.

Affect prevalence and spread by news category

To examine whether the association of news source bias with prevalence and spread of high arousal negative content generalized across different types of news, we modeled news content topics with a Latent Dirichlet Allocation (LDA) analysis [30]. To create semantically relevant topics, we incorporated priors into the LDA optimization by initializing topic distribution parameters with estimates from human coders. Specifically, we first divided the dataset into a smaller training sample of post hashtags that appeared most frequently (i.e., those that were ranked in the top 500 hashtags with respect to frequency). Next, we identified seven news categories along with short descriptions for coding their labels (i.e., “politics and policy,” “sports,” “health and wellness,” “business and economy,” “crime and safety,” “culture and lifestyle,” and “science and technology,” see S1 File for category label descriptions). Two trained raters then classified the 500 highest-frequency hashtag terms into these categories. The trained raters showed high inter-rater reliability (kappa = .743) in their categorical classifications of these hashtag terms. Labeled hashtags were then used to estimate the topic distribution of the training sample and the document-topic distribution in the LDA model. The LDA model was then optimized from the labelled training sample (using the Gensim 4.1.2 Python package) and generalized to the remainder of the dataset. This model then assigned the most likely news category label to each post, retaining label predictions that exceeded a 1% false positive rate.

Out of a total of 10,315,511 labeled posts, assigned percentages varied by label category (i.e., with the model assigning 51.0% to “politics and policy,” 15.7% to “sports,” 10.5% to “health and wellness,” 7.0% to “business and economics,” 6.3% to “crime and safety,” 5.7% to “culture and lifestyle,” and 3.8% to “science and technology”). Analyses of affective content and spread implemented for the entire sample of posts were then repeated specifically for posts within each of these categories. The association of news source bias with more prevalent as well as more viral high arousal negative affective content was most apparent in the largest news category of “politics and policy” (i.e., > 50% of posts) and less so in “business and economics”, but not in the other news categories (i.e., “sports,” “health and wellness,” “crime and safety,” “culture and lifestyle,” or “science and technology”; see S1 File for plotted LDA results).

Discussion

Sentiment analysis of nearly 30 million social media posts from 182 news sources varying in partisan bias over the course of a decade (January 1, 2011 to December 31, 2020) revealed that biased news sources produced more highly arousing negative affective content. Specifically, the most biased news sources (on both the left and the right) included approximately 12% more high arousal negative affective content in their posts than balanced news sources, and this trend increased over time. Further, posts that contained highly arousing negative affective content were more reposted, spreading more than those that did not. Critically, this virality effect was stronger for more biased news sources, since highly arousing negative affective content increased sharing (or reposts) by 5.30% in extreme left sources and by 5.73% in extreme right sources, compared to 1.46% in more balanced news sources. Unexpectedly, this difference between spread of high arousal negative content in biased versus balanced news sources appeared to diminish over time. Further analyses revealed, however, that this was not due to decreasing spread of high arousal negative content in biased news sources, but rather to increasing spread of high arousal negative content in balanced news sources over time.

The combination of increased prevalence and increased spread was specific to high arousal negative affective news content, but not other types of affective content. Although posts with low arousal negative affective content were also reposted more in biased news sources, biased news sources did not post more low arousal negative affective content than balanced news sources, and the association of low arousal negative content with spread was smaller than that for high arousal negative content. Biased news sources also posted more low arousal positive affective content and high arousal positive affective content than balanced news sources, but these effects were again smaller than for high arousal negative affective content, and posts containing low arousal positive affective content and high arousal positive affective content did differentially spread from biased versus balanced news sources. Exploratory analyses suggested that this pattern of higher prevalence and spread of high arousal negative content was primarily present in posts related to politics and policy, which comprised over half of the analyzed news posts (see S1 File).

This work provides an initial overview of the prevalence and spread of different types of affective content in social media posts by United States news sources that vary along a continuum of partisan bias over the course of a decade. By examining affective content that included all combinations of valence and arousal, we were able to isolate high arousal negative affective content as uniquely more prevalent and more viral in posts from biased news sources. This distinction between high and low arousal affect allowed us to test a prediction derived from an anticipatory affect model that highly arousing affective content would be more likely to capture attention and motivate behavior [21]. In the current findings, however, only high arousal negative affect showed a clear association with subsequent spread, extending previous findings associating negative affective content with news virality [15]. Notably, these patterns did not depend on the direction of news source bias, instead extending to biased news sources on both the left and the right [14]. These results add specificity to emerging findings which imply that negative affective content predominates and can drive virality in United States news media on both ends of the political spectrum [9, 12, 14, 15]. The prevalence and spread of high arousal negative content in biased sources was not limited to current developments or events, since it generally increased over ten years. Further, the virality of high arousal negative content did not decrease over time for biased news sources, but rather increased over time for balanced news sources.

Despite the fact that individual American social media users tend to produce more positive than negative affective content (i.e., ~50% more positive than negative affective content on twitter.com) [19], news sources instead produced more negative than positive affective content overall (i.e., ~80% more negative than positive content on twitter). This asymmetry in the production of affective content implies that part of the high arousal negative content circulating on social media may emanate not only from individual users, but even more from biased news media. This pattern of findings is also consistent with reports that social media algorithms may boost high arousal negative content [31]. Specifically, ranking algorithms on many social media platforms may learn from users’ engagement behavior (e.g., reposts) to prioritize and prominently present engaging content [32]. This algorithmic boost might create feedback loops that amplify the spread and reach of that engaging content. Since high arousal negative posts appear to elicit greater engagement, these posts may receive higher ranking in social media timelines, and so reach more users. To further maximize engagement in an increasingly competitive media landscape, news outlets might therefore be incentivized to produce even more high arousal negative affective content. Effects of such a feedback loop might then intensify in biased news sources since they post relatively more high arousal negative affective content and tend to show greater spread for those posts.

Excessive high arousal negative content might therefore generate a form of “affective pollution” by imposing unwanted costs on individual users, since social media users in the United States behaviorally appear to prefer producing positive content when left to their own devices [19]. Costs of this “affective pollution” could take several forms. First, negative arousing content might increase negative affect and decrease well-being in users, either over the short or the long term [33]. Second, and based on its prevalence in biased media, negative arousing content might boost attention to and subsequent acceptance of associated misinformation versus information [33, 34]. Third, and particularly in for outgroup targets, negative arousing content may increase political polarization and hostility towards others [15]. Thus, although negative affective content prevalence combined with spread may increase engagement and associated income for social media platforms [8, 9, 12], it may also impose a toxic downward spiral of unwanted negative costs on individual users.

Researchers have posited that the virality of high arousal negative affect on social media might be due to the salience of threat [12], an absence of channels for corrective feedback (e.g., a lack of retaliation for norm-violating behavior; [18]), or the surprisingness of affective content that violates cultural values [19]. Future research might examine the influence of these mechanisms on content from biased versus balanced news sources. This research focused on available data from only one social media platform (i.e., the former twitter.com). Another recent study of headline sentiment culled from a subset of these news sources across platforms on the open internet, however, reached similar conclusions about a general rise in negative emotional content from 2000–2020 [20]. Since the present research focused on news from the United States in social media, generalizability to other cultures also deserves investigation. For instance, the viral influence of high arousal negative affective content might not generalize to more interdependent East Asian cultures and so may fail to elicit similar engagement [19]. Comparing news bias and affective content in social media across cultures might also help to elaborate and refine global efforts to curb misinformation.

Future research might characterize the impact of affective content on other social media platforms and involving different groups in greater detail. With respect to different social media platforms, while the previous open accessibility of data (e.g., on twitter.com) made this large-scale analysis possible, generalizability should ideally be established to other closed social media platforms. Consistent with the present findings, a similar influence of negative affective content on post virality has been noted on at least one other prominent closed social media platform (i.e., Facebook) [15]. With respect to group identity, researchers could analyze whether affective content and spread varies as a function of the topic under discussion–particularly with respect to polarizing topics that denigrate outgroups or elevate the ingroup [15]. Consistent with a group identity account, exploratory analyses implied that the influence of negative affective content on virality was most pronounced for political news, which constituted the majority of analyzed posts. To a lesser extent than high arousal negative affective content, posts containing low arousal positive affective content also spread more in biased news sources. Further research might test whether virality is associated with positive affective content supporting the ingroup [29]. While this research focused on total reposts, future analyses might probe how affective content influences reposting across different user networks, both within and between different groups [12].

These findings coincide with a rise in scientific research on affective phenomena due to advances in measurement (including sentiment analysis), which allow researchers to test qualitative predictions with quantitative tools [35]. Recently, researchers have developed even more precise and granular tools for sentiment analysis (e.g., Large Language Models), which might be leveraged to test the specificity as well as the generalizability of these findings [36]. Measurement and quantification of affective content further implies novel possibilities for intervention. Beyond characterizing the prevalence and spread of affective content, precise interventions might be designed that can reduce negative aroused affective sentiment in an unbiased manner (i.e., independent of message source and semantics). The present findings suggest that interventions with the capacity to identify and filter negative aroused affective content might also decrease engagement with biased news content, possibly shifting user exposure towards more balanced and veridical news content. Since previous research suggests that American social media users preferentially produce positive rather than negative content, negative arousing content may “hijack” users’ preferences, leading to unintended and unwanted decreases in well-being, information quality, and social harmony. Digital literacy programs might therefore refer to the current findings to educate users about how their affect can be disrupted by biased news sources. Understanding how affective content influences engagement might also inform efforts to design more optimal algorithms that diminish rather than amplify the spread of biased news. Such algorithms might provide companies, governments, and especially individual users with tools capable of countering the rapid spread of damaging and divisive online content. Since this research focused on naturalistic data, we could only investigate mechanisms that might drive the observed effects, leaving the potential impact of relevant causal interventions open for future exploration and development [37, 38].

Supporting information

S1 Fig. Word prevalence clouds by post affect type (derived from 700 randomly-selected example posts per affect type category).

(TIF)

pone.0305148.s001.tif (2.2MB, tif)
S1 File. Supporting methods and results.

(DOCX)

pone.0305148.s002.docx (3.7MB, docx)

Acknowledgments

We thank Alexis Ceja for assistance with classification, Nicole Deng for assistance with analysis, and Yiwei Luo, Jonas Schöne, the Tsai lab, Spanlab, and two anonymous reviewers for feedback on previous drafts.

Data Availability

Summary data and analysis code supporting the reported findings are publicly available in an Open Science Framework repository (https://osf.io/63pzy/). Raw data, however, is owned by and available from x.com (formerly, twitter.com) through their Advanced Programming Interface (https://developer.x.com/en/products/x-api). The authors initially used this API to access raw data (without special access).

Funding Statement

This work was supported by NSF Grant 1732963, a Stanford Ethics, Society, and Technology Grant, and a Stanford HAI Faculty Seed Grant to JT and BK.

References

  • 1.Shearer E, Mitchell A. News use across social media platforms in 2020. Pew Res Cent. 2021; 18. [Google Scholar]
  • 2.Spring M. Coronavirus: The human cost of virus misinformation. BBC News. 2020. Available: www.bbc.com/news/stories-52731624 [Google Scholar]
  • 3.Van Bavel JJ, Rathje S, Harris E, Robertson C, Sternisko A. How social media shapes polarization. Trends Cogn Sci. 2021;25: 913–916. doi: 10.1016/j.tics.2021.07.013 [DOI] [PubMed] [Google Scholar]
  • 4.Iyengar S, Lelkes Y, Levendusky M, Malhotra N, Westwood SJ. The Origins and Consequences of Affective Polarization in the United States. Annu Rev Polit Sci. 2019;22: 129–146. doi: 10.1146/annurev-polisci-051117-073034 [DOI] [Google Scholar]
  • 5.Prior M. Media and Political Polarization. Annu Rev Polit Sci. 2013;16: 101–127. doi: 10.1146/annurev-polisci-100711-135242 [DOI] [Google Scholar]
  • 6.Jamieson KH, Albarracín D. The Relation between Media Consumption and Misinformation at the Outset of the SARS-CoV-2 Pandemic in the US. Harv Kennedy Sch Misinformation Rev. 2020;1: 1–21. doi: 10.37016/mr-2020-012 [DOI] [Google Scholar]
  • 7.Osmundsen M, Bor A, Vahlstrup PB, Bechmann A, Petersen MB. Partisan Polarization Is the Primary Psychological Motivation behind Political Fake News Sharing on Twitter. Am Polit Sci Rev. 2021;115: 999–1015. doi: 10.1017/S0003055421000290 [DOI] [Google Scholar]
  • 8.Hasell A, Weeks BE. Partisan Provocation: The Role of Partisan News Use and Emotional Responses in Political Information Sharing in Social Media: Partisan News, Emotions, and Information Sharing. Hum Commun Res. 2016;42: 641–661. doi: 10.1111/hcre.12092 [DOI] [Google Scholar]
  • 9.Vosoughi S, Roy D, Aral S. The spread of true and false news online. Science. 2018;359: 1146–1151. doi: 10.1126/science.aap9559 [DOI] [PubMed] [Google Scholar]
  • 10.Grinberg N, Joseph K, Friedland L, Swire-Thompson B, Lazer D. Fake news on Twitter during the 2016 U.S. presidential election. Science. 2019;363: 374–378. doi: 10.1126/science.aau2706 [DOI] [PubMed] [Google Scholar]
  • 11.Berger J, Milkman KL. What Makes Online Content Viral? J Mark Res. 2012;49: 192–205. doi: 10.1509/jmr.10.0353 [DOI] [Google Scholar]
  • 12.Brady WJ, Wills JA, Jost JT, Tucker JA, Van Bavel JJ. Emotion shapes the diffusion of moralized content in social networks. Proc Natl Acad Sci. 2017;114: 7313–7318. doi: 10.1073/pnas.1618923114 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Xu WW, Sang Y, Kim C. What Drives Hyper-Partisan News Sharing: Exploring the Role of Source, Style, and Content. Digit Journal. 2020;8: 486–505. doi: 10.1080/21670811.2020.1761264 [DOI] [Google Scholar]
  • 14.Bellovary AK, Young NA, Goldenberg A. Left- and Right-Leaning News Organizations Use Negative Emotional Content and Elicit User Engagement Similarly. Affect Sci. 2021. [cited 15 Nov 2021]. doi: 10.1007/s42761-021-00046-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Rathje S, Van Bavel JJ, van der Linden S. Out-group animosity drives engagement on social media. Proc Natl Acad Sci. 2021;118: e2024292118. doi: 10.1073/pnas.2024292118 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Mourão RR, Robertson CT. Fake News as Discursive Integration: An Analysis of Sites That Publish False, Misleading, Hyperpartisan and Sensational Information. Journal Stud. 2019;20: 2077–2095. doi: 10.1080/1461670X.2019.1566871 [DOI] [Google Scholar]
  • 17.Stieglitz S, Dang-Xuan L. Emotions and Information Diffusion in Social Media—Sentiment of Microblogs and Sharing Behavior. J Manag Inf Syst. 2013;29: 217–248. doi: 10.2753/MIS0742-1222290408 [DOI] [Google Scholar]
  • 18.Crockett MJ. Moral outrage in the digital age. Nat Hum Behav. 2017;1: 769–771. doi: 10.1038/s41562-017-0213-3 [DOI] [PubMed] [Google Scholar]
  • 19.Hsu TW, Niiya Y, Thelwall M, Ko M, Knutson B, Tsai JL. Social media users produce more affect that supports cultural values, but are more influenced by affect that violates cultural values. J Pers Soc Psychol. 2021. [cited 15 Nov 2021]. doi: 10.1037/pspa0000282 [DOI] [PubMed] [Google Scholar]
  • 20.Rozado D, Hughes R, Halberstadt J. Longitudinal analysis of sentiment and emotion in news media headlines using automated labelling with Transformer language models. Consoli S, editor. PLOS ONE. 2022;17: e0276367. doi: 10.1371/journal.pone.0276367 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Knutson B, Greer SM. Anticipatory affect: neural correlates and consequences for choice. Philos Trans R Soc B Biol Sci. 2008;363: 3771–3786. doi: 10.1098/rstb.2008.0155 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Watson D, Anna L, Tellegen A. Development and Validation of Brief Measures of Positive and Negative Affect: The PANAS Scales. [DOI] [PubMed] [Google Scholar]
  • 23.Weeks BE, Gil de Zúñiga H. What’s Next? Six Observations for the Future of Political Misinformation Research. Am Behav Sci. 2021;65: 277–289. doi: 10.1177/0002764219878236 [DOI] [Google Scholar]
  • 24.Thelwall M, Buckley K, Paltoglou G, Cai D, Kappas A. Sentiment strength detection in short informal text. J Am Soc Inf Sci Technol. 2010;61: 2544–2558. doi: 10.1002/asi.21416 [DOI] [Google Scholar]
  • 25.Otero V. Multi-analyst ratings project August 2019. AdFontes Media; 2019. Available: https://adfontesmedia.com/white-paper-multi-analyst-ratings-project-august-2019/?v=402f03a963ba [Google Scholar]
  • 26.Thelwall M, Street W, Wv W. Heart and Soul: Sentiment Strength Detection in the Social Web with SentiStrength.: 14. [Google Scholar]
  • 27.Ferrara E, Yang Z. Measuring Emotional Contagion in Social Media. Bauch CT, editor. PLOS ONE. 2015;10: e0142390. doi: 10.1371/journal.pone.0142390 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Thelwall M, Buckley K, Paltoglou G. Sentiment strength detection for the social web. J Am Soc Inf Sci Technol. 2012;63: 163–173. doi: 10.1002/asi.21662 [DOI] [Google Scholar]
  • 29.Wojcieszak M, Bimber B, Feldman L, Stroud NJ. Partisan News and Political Participation: Exploring Mediated Relationships. Polit Commun. 2016;33: 241–260. doi: 10.1080/10584609.2015.1051608 [DOI] [Google Scholar]
  • 30.Blei DM, Ng, Andrew Y, Jordan, Michael I. Latent Dirichlet Allocation. J Mach Learn Res. 2003; 993–1022. [Google Scholar]
  • 31.Merrill JB, Oremus W. Five points for anger, one for a ‘like’: How Facebook’s formula fostered rage and misinformation. The Washington Post. 26 Oct 2021. Available: https://www.washingtonpost.com/technology/2021/10/26/facebook-angry-emoji-algorithm/ [Google Scholar]
  • 32.Agichtein E, Brill E, Dumais S. Improving Web Search Ranking by Incorporating User Behavior Information. Proc 29th Annu Int ACM SIGIR Conf Res Dev Inf Retr. 2006; 19–26. [Google Scholar]
  • 33.Allcott H, Braghieri L, Eichmeyer S, Gentzkow M. The Welfare Effects of Social Media. Am Econ Rev. 2020;110: 629–676. doi: 10.1257/aer.20190658 [DOI] [Google Scholar]
  • 34.Weeks BE. Emotions, Partisanship, and Misperceptions: How Anger and Anxiety Moderate the Effect of Partisan Bias on Susceptibility to Political Misinformation: Emotions and Misperceptions. J Commun. 2015;65: 699–719. doi: 10.1111/jcom.12164 [DOI] [Google Scholar]
  • 35.Dukes D, Abrams K, Adolphs R, Ahmed ME, Beatty A, Berridge KC, et al. The rise of affectivism. Nat Hum Behav. 2021;5: 816–820. doi: 10.1038/s41562-021-01130-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Rathje S, Mirea D-M, Sucholutsky I, Marjieh R, Robertson CE, Van Bavel JJ. GPT is an effective tool for multilingual psychological text analysis. Proc Natl Acad Sci. 2024;121: e2308950121. doi: 10.1073/pnas.2308950121 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Martel C, Pennycook G, Rand DG. Reliance on emotion promotes belief in fake news. Cogn Res Princ Implic. 2020;5: 47. doi: 10.1186/s41235-020-00252-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Pennycook G, McPhetres J, Zhang Y, Lu JG, Rand DG. Fighting COVID-19 Misinformation on Social Media: Experimental Evidence for a Scalable Accuracy-Nudge Intervention. Psychol Sci. 2020;31: 770–779. doi: 10.1177/0956797620939054 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Silvio Eduardo Alvarez Candido

23 Jul 2024

PONE-D-24-21176News source bias and sentiment on social mediaPLOS ONE

Dear Dr. Knutson,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.While the two experts who reviewed your manuscript considered it to be strong and have great potential for publication, they also made suggestions regarding the contextualization and the illustration of your assessment and the verification of some methodological issues which I believe will help to improve your paper. 

Please submit your revised manuscript by Sep 06 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Silvio Eduardo Alvarez Candido

Academic Editor

PLOS ONE

Journal requirements:

1. When submitting your revision, we need you to address these additional requirements.

Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf.

2. (1) In your Methods section, please include additional information about your dataset and ensure that you have included a statement specifying whether the collection and analysis method complied with the terms and conditions for the source of the data.

(2) Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, we expect all author-generated code to be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. 

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

4. Thank you for stating the following financial disclosure: 

 [This work was supported by NSF Grant 1732963, a Stanford Ethics, Society, and Technology Grant, and a Stanford HAI Faculty Seed Grant to JT and BK. ].  

Please state what role the funders took in the study.  If the funders had no role, please state: ""The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript."" 

If this statement is not correct you must amend it as needed. 

Please include this amended Role of Funder statement in your cover letter; we will change the online submission form on your behalf.

5. Thank you for stating the following in the Acknowledgments Section of your manuscript: 

[We thank Alexis Ceja for assistance with coding and classification, as well as Yiwei Luo, Jonas Schöne, the Tsai lab, and Spanlab for feedback on previous drafts. This work was supported by NSF Grant 1732963, a Stanford Ethics, Society, and Technology Grant, and a Stanford HAI Faculty Seed Grant to JT and BK. ]

We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. 

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows: 

 [This work was supported by NSF Grant 1732963, a Stanford Ethics, Society, and Technology Grant, and a Stanford HAI Faculty Seed Grant to JT and BK. ]

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

6. We note that you have indicated that there are restrictions to data sharing for this study. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For more information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. 

Before we proceed with your manuscript, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially identifying or sensitive patient information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., a Research Ethics Committee or Institutional Review Board, etc.). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of recommended repositories, please see

https://journals.plos.org/plosone/s/recommended-repositories. You also have the option of uploading the data as Supporting Information files, but we would recommend depositing data directly to a data repository if possible.

We will update your Data Availability statement on your behalf to reflect the information you provide.

7. Please remove your figures from within your manuscript file, leaving only the individual TIFF/EPS image files, uploaded separately. These will be automatically included in the reviewers’ PDF.

8. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information. 

9. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This work investigates whether the bias of the news source affects the creation of emotional content and its spread, and whether these differences have evolved over time. The authors analyzed the sentiment of 30 million tweets from 182 U.S. news sources, ranging from extreme left to extreme right bias, over a decade (2011-2020). The results indicate that biased news sources, whether left or right, produced more high-arousal negative emotional content compared to balanced sources. This high-arousal negative content also increased the likelihood of reposting for biased sources as opposed to balanced ones (but this trend did not hold for other types of emotional content). In recent years, the virality of high-arousal negative content rose, especially in balanced news sources and posts about politics.

I would like to commend the authors for a methodologically robust work. The study is well thought out. The results are perhaps not too surprising and mostly consistent with the pre-existing literature but it is still important to do the quantitative documentation of such trends. The attention to detail and the robust methodology is visible in many places of the paper like for instance when the authors triangulate their initial choice of news media bias ratings (from Ad Fontes) with another provider (AllSides). The LDA topic modeling to characterize the news sources sections driving the results was another nice touch. The statistical analysis is solid and some of the somehow counterintuitive results (like the increased high arousal negative affective content in balanced news sources) are an important contribution to the existing literature. I barely have much criticism to utter and my recommendation is to accept this paper with only minor revisions.

I would also like to commend again the authors because as a reader of their manuscript they come across as very impartial in their work, as it should be!. This is unfortunately not always the case when reading these sorts of paper, and it feels like some authors have an agenda or some sort of viewpoint bias that they try to accommodate the results to. This is not the case at all in the current paper. It feels exquisitely neutral in that regard.

Only a bit of minor constructive criticism for the authors at the risk of being pedantic. I think the weakest link in the paper are the visualizations where there is definitely room for improvement. All the detailed reporting about the different linear models used for the analysis in the text is fine but sometimes a compelling visualization can convey the point across very efficiently.

The comments in the discussion section about the spread of negative affect news content through the population is highly relevant and important. Perhaps the authors should mention at least a sort of counter hypothesis: that people perhaps could be developing some sort of resistance or immunity after perhaps too much exposure to negative sentiment in their news diets and could be tuning out. I'm not saying that this is happening, but perhaps the authors should concede this as one of many potential hypotheses.

Finally, the usage of SentiStrength is well reasoned. I was just wondering whether an LLM like gpt-3.5-turbo would have a high degree of interrater agreement with the SentiStrength algorithm annotations, thus providing further evidence for the validity of the findings. But again, this is a bit pedantic.

To conclude, well done to the authors for a solid piece of academic work.

Reviewer #2: The paper contributes to fill empirical gaps though solid assessments. It deals with a very important issue of how political bias influence the diffusion of news. The findings are not so surprising, but clearly generate an empirical contribution. I have only a couple of suggestions to improve the work after publication.

While the empirical analysis of the paper is excellent, the paper is a bit atheoretical to my taste. I miss recourse to more general ideas and mechanisms in the contextualization regarding existing explanations about why would biased theory spread more. In my view this would be important to better contextualize the paper in nearby discussions. This would be particularly important for the introduction of the paper.

Another point refers to the illustration of the findings. First, the definition of some figures may be improved. Also, in some parts the authors describe textually findings that could be nicely visually represented.

I congratulate the authors for the excellent work.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Oct 23;19(10):e0305148. doi: 10.1371/journal.pone.0305148.r002

Author response to Decision Letter 0


10 Sep 2024

Journal requirements:

1. When submitting your revision, we need you to address these additional requirements.

Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf.

We have now updated our style to meet the PLOS ONE style requirements throughout.

2. (1) In your Methods section, please include additional information about your dataset and ensure that you have included a statement specifying whether the collection and analysis method complied with the terms and conditions for the source of the data.

We now state in the Methods that: “Data collection and processing complied with the terms and conditions imposed by twitter.com (now x.com).”

(2) Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, we expect all author-generated code to be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

We now provide our analysis code at: https://osf.io/63pzy/.

3. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match.

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

We have removed the updated funding information from the acknowledgments and now provide it in response to item 5 below.

4. Thank you for stating the following financial disclosure:

[This work was supported by NSF Grant 1732963, a Stanford Ethics, Society, and Technology Grant, and a Stanford HAI Faculty Seed Grant to JT and BK.].

Please state what role the funders took in the study. If the funders had no role, please state: ""The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.""

If this statement is not correct you must amend it as needed.

Please include this amended Role of Funder statement in your cover letter; we will change the online submission form on your behalf.

We now state in the Financial Disclosure that: “The authors have no conflicts of interest to declare. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.” (line 587)

5. Thank you for stating the following in the Acknowledgments Section of your manuscript:

[We thank Alexis Ceja for assistance with coding and classification, as well as Yiwei Luo, Jonas Schöne, the Tsai lab, and Spanlab for feedback on previous drafts. This work was supported by NSF Grant 1732963, a Stanford Ethics, Society, and Technology Grant, and a Stanford HAI Faculty Seed Grant to JT and BK.]

We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

[This work was supported by NSF Grant 1732963, a Stanford Ethics, Society, and Technology Grant, and a Stanford HAI Faculty Seed Grant to JT and BK.]

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

Thank you for this clarification -- we have removed this information from the Acknowledgments section and would like the Funding Statement to read, and now also mention this in our cover letter:

“This work was supported by NSF Grant 1732963, a Stanford Ethics, Society, and Technology Grant, and a Stanford HAI Faculty Seed Grant to JT and BK. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.”

6. We note that you have indicated that there are restrictions to data sharing for this study. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For more information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions.

Before we proceed with your manuscript, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially identifying or sensitive patient information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., a Research Ethics Committee or Institutional Review Board, etc.). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

We now provide summary data, but not raw data, which is prohibited by the current x.com data sharing policy. We also note this in the Data Availability statement (see below).

b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of recommended repositories, please see

https://journals.plos.org/plosone/s/recommended-repositories. You also have the option of uploading the data as Supporting Information files, but we would recommend depositing data directly to a data repository if possible.

We now provide summary data in an OSF repository, as noted in the Data Availability statement:

“Summary data and analysis code supporting the reported findings are publicly available in an Open Science Framework repository (https://osf.io/63pzy/). Restrictions now apply to the raw data based on the data policy of x.com (formerly, twitter.com), which were used under license and so are no longer publicly available.” (line 591)

We will update your Data Availability statement on your behalf to reflect the information you provide.

7. Please remove your figures from within your manuscript file, leaving only the individual TIFF/EPS image files, uploaded separately. These will be automatically included in the reviewers’ PDF.

We have removed the figures from the manuscript file and uploaded them separately.

8. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

We now include captions for our Supporting Information files at the end of the manuscript and have updated in-text citations accordingly.

9. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

We have reviewed and revised our references section to conform to PLOS ONE style.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Responses to Reviewers

Reviewer #1: This work investigates whether the bias of the news source affects the creation of emotional content and its spread, and whether these differences have evolved over time. The authors analyzed the sentiment of 30 million tweets from 182 U.S. news sources, ranging from extreme left to extreme right bias, over a decade (2011-2020). The results indicate that biased news sources, whether left or right, produced more high-arousal negative emotional content compared to balanced sources. This high-arousal negative content also increased the likelihood of reposting for biased sources as opposed to balanced ones (but this trend did not hold for other types of emotional content). In recent years, the virality of high-arousal negative content rose, especially in balanced news sources and posts about politics.

I would like to commend the authors for a methodologically robust work. The study is well thought out. The results are perhaps not too surprising and mostly consistent with the pre-existing literature but it is still important to do the quantitative documentation of such trends. The attention to detail and the robust methodology is visible in many places of the paper like for instance when the authors triangulate their initial choice of news media bias ratings (from Ad Fontes) with another provider (AllSides). The LDA topic modeling to characterize the news sources sections driving the results was another nice touch. The statistical analysis is solid and some of the somehow counterintuitive results (like the increased high arousal negative affective content in balanced news sources) are an important contribution to the existing literature. I barely have much criticism to utter and my recommendation is to accept this paper with only minor revisions.

I would also like to commend again the authors because as a reader of their manuscript they come across as very impartial in their work, as it should be! This is unfortunately not always the case when reading these sorts of paper, and it feels like some authors have an agenda or some sort of viewpoint bias that they try to accommodate the results to. This is not the case at all in the current paper. It feels exquisitely neutral in that regard.

We deeply appreciate this encouraging feedback (particularly given that this is a new line of inquiry for us)!

Only a bit of minor constructive criticism for the authors at the risk of being pedantic. I think the weakest link in the paper are the visualizations where there is definitely room for improvement. All the detailed reporting about the different linear models used for the analysis in the text is fine but sometimes a compelling visualization can convey the point across very efficiently.

Display more results graphically.

We appreciate this opportunity to improve and augment the figures (and apologize for their originally suboptimal resolution).

Specifically, we have now cleaned and re-rendered the existing figures, adding color to indicate affect type. Additionally, based on the feedback, we now have updated and integrated some of the figures from the Supplement into the manuscript (i.e., those depicting the validity of the bias ratings, a sentiment analysis flowchart, and a breakdown of the bias by time interaction).

The comments in the discussion section about the spread of negative affect news content through the population is highly relevant and important. Perhaps the authors should mention at least a sort of counter hypothesis: that people perhaps could be developing some sort of resistance or immunity after perhaps too much exposure to negative sentiment in their news diets and could be tuning out. I'm not saying that this is happening, but perhaps the authors should concede this as one of many potential hypotheses.

Address potential alternative hypotheses.

The notion that people habituate to high arousal negative content may have come from the figure showing that the difference in virality of high arousal negative content has narrowed over 10 years. This diminishing difference, however, is due to high arousal content becoming more viral for balanced sources rather than less viral for biased sources, increasing the overall virality of high arousal negative content over 10 years (which would be more consistent with sensitization than habituation). To clarify this increase, we now include a Figure from the Supplement (Figure 5) in the main manuscript which depicts this interaction. We also now clarify that both the overall production and virality of high arousal negative content increased over 10 years in the Discussion:

“The prevalence and spread of high arousal negative content in biased sources was not limited to current developments or events, since it generally increased over ten years. Further, the virality of high arousal negative content did not decrease over time for biased news sources, but rather increased over time for balanced news sources.” (line 487)

Finally, the usage of SentiStrength is well reasoned. I was just wondering whether an LLM like gpt-3.5-turbo would have a high degree of interrater agreement with the SentiStrength algorithm annotations, thus providing further evidence for the validity of the findings. But again, this is a bit pedantic.

Consider the application of large language models to sentiment coding.

We agree that these developments offer a promising set of methods for future inquiry (as further suggested by a new reference), and so now explicitly raise and recommend this research direction in the Discussion:

“These findings coincide with a rise in scientific research on affective phenomena due to advances in measurement (including sentiment analysis), which allow researchers to test qualitative predictions with quantitative tools [35]. Recently, researchers have developed even more precise and granular tools for sentiment analysis (e.g., Large Language Models), which might be leveraged to test the specificity as well as the generalizability of these findings [36].” (line 552)

To conclude, well done to the authors for a solid piece of aca

Attachment

Submitted filename: newsaff_responsetoreviewers_240901.docx

pone.0305148.s003.docx (36.3KB, docx)

Decision Letter 1

Silvio Eduardo Alvarez Candido

17 Sep 2024

News source bias and sentiment on social media

PONE-D-24-21176R1

Dear Dr. Knutson,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Silvio Eduardo Alvarez Candido

Academic Editor

PLOS ONE

Acceptance letter

Silvio Eduardo Alvarez Candido

20 Sep 2024

PONE-D-24-21176R1

PLOS ONE

Dear Dr. Knutson,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Silvio Eduardo Alvarez Candido

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig. Word prevalence clouds by post affect type (derived from 700 randomly-selected example posts per affect type category).

    (TIF)

    pone.0305148.s001.tif (2.2MB, tif)
    S1 File. Supporting methods and results.

    (DOCX)

    pone.0305148.s002.docx (3.7MB, docx)
    Attachment

    Submitted filename: newsaff_responsetoreviewers_240901.docx

    pone.0305148.s003.docx (36.3KB, docx)

    Data Availability Statement

    Summary data and analysis code supporting the reported findings are publicly available in an Open Science Framework repository (https://osf.io/63pzy/). Raw data, however, is owned by and available from x.com (formerly, twitter.com) through their Advanced Programming Interface (https://developer.x.com/en/products/x-api). The authors initially used this API to access raw data (without special access).


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES