Skip to main content
PLOS One logoLink to PLOS One
. 2023 Dec 20;18(12):e0295414. doi: 10.1371/journal.pone.0295414

Detecting nuance in conspiracy discourse: Advancing methods in infodemiology and communication science with machine learning and qualitative content coding

Michael Robert Haupt 1,2,*, Michelle Chiu 3, Joseline Chang 4, Zoe Li 2,5, Raphael Cuomo 6, Tim K Mackey 5,7
Editor: Stefano Cresci8
PMCID: PMC10732406  PMID: 38117843

Abstract

The spread of misinformation and conspiracies has been an ongoing issue since the early stages of the internet era, resulting in the emergence of the field of infodemiology (i.e., information epidemiology), which investigates the transmission of health-related information. Due to the high volume of online misinformation in recent years, there is a need to continue advancing methodologies in order to effectively identify narratives and themes. While machine learning models can be used to detect misinformation and conspiracies, these models are limited in their generalizability to other datasets and misinformation phenomenon, and are often unable to detect implicit meanings in text that require contextual knowledge. To rapidly detect evolving conspiracist narratives within high volume online discourse while identifying nuanced themes requiring the comprehension of subtext, this study describes a hybrid methodology that combines natural language processing (i.e., topic modeling and sentiment analysis) with qualitative content coding approaches to characterize conspiracy discourse related to 5G wireless technology and COVID-19 on Twitter (currently known as ‘X’). Discourse that focused on correcting 5G conspiracies was also analyzed for comparison. Sentiment analysis shows that conspiracy-related discourse was more likely to use language that was analytic, combative, past-oriented, referenced social status, and expressed negative emotions. Corrections discourse was more likely to use words reflecting cognitive processes, prosocial relations, health-related consequences, and future-oriented language. Inductive coding characterized conspiracist narratives related to global elites, anti-vax sentiment, medical authorities, religious figures, and false correlations between technology advancements and disease outbreaks. Further, the corrections discourse did not address many of the narratives prevalent in conspiracy conversations. This paper aims to further bridge the gap between computational and qualitative methodologies by demonstrating how both approaches can be used in tandem to emphasize the positive aspects of each methodology while minimizing their respective drawbacks.

1. Introduction

1.1 The emergence of infodemiology in the early internet era

Since the 2016 US presidential election, online misinformation spread had become a ubiquitous topic in public discourse with rising concerns related to the proliferation of “fake news” on social media platforms such as Facebook and Twitter before the election [13]. Public concern about misinformation spread has only grown since then, with the World Health Organization (WHO) declaring an ‘infodemic’ in reaction to the proliferation of misinformation related to the COVID-19 pandemic [4]. However, risk related to online misinformation has existed since the early stages of the internet, as shown by the emergence of the field of infodemiology during that era [5, 6]. Infodemiology, also known as “information epidemiology,” is the study of the determinants and distribution of health information and misinformation, and identifies areas where there is a knowledge translation gap between best available evidence and what most people do or believe, as well as markers for “high-quality” information [5, 7]. Infodemiological frameworks make distinctions between supply-based applications, such as analyzing what is being published on web sites and social media, and demand-based methods that examine search and navigation behavior on the internet [7]. Applications of infodemiology include the analysis of queries from internet search engines to predict disease outbreaks [6], monitoring peoples’ health status updates on platforms such as Twitter and Weibo for syndromic surveillance [810], identifying prevalent themes and discourse around health conditions and behaviors (including COVID-19) [7, 11, 12], and studies examining online mobilization of social media users to influence health policy outcomes [13, 14].

While social media platforms are an incredible tool for staying connected and communicating with others, they can also simultaneously be a source of uncertainty and fear, which are often accompanied by increased dissemination of unverified rumors, misinformation, and fringe or conspiracy theories [15, 16]. To minimize the spread of misinformation, it is crucial to rapidly identify and then systematically sift through the high volume of posts and comments that often accompanies related online narratives, such as claims stating false ties between COVID-19 and 5G wireless technology. In response to this growing need for rapid content characterization, particularly in the context of health emergencies, supervised and unsupervised machine learning approaches, such as natural language processing (NLP) and supervised learning trained on fact-checked data, have been used to identify and classify misinformation [1720]. Additionally, large language models (LLM) have shown high accuracy scores for identifying misinformation [21, 22], making popular models like chatGPT another promising tool for future infodemiological research. Despite the fact that various machine learning approaches are effective at classifying information and are crucial for adequately addressing the ongoing proliferation of misinformation that is overwhelming the social media ecosystem, these methods can be limited by their associated lexicon or dictionary and the semantic shifts that naturally occur in language over time and across contexts, including changing evidence about COVID-19 and other public health issues that may arise. Thus, for emerging issues that would not be recorded within training datasets, maintaining the use of human review and annotation remains necessary for properly detecting and contextualizing novel situations and narratives.

To gain a more thorough understanding of the nuanced patterns and dynamics underlying the spread of misinformation, our study introduces a hybrid analytic approach using metadata from 5G-COVID conspiracy discourse on Twitter (currently known as ‘X’ but referred to as Twitter for this paper) aimed at leveraging both the efficiency of NLP techniques and the qualitative schema afforded by human coders. More specifically, this study used topic modeling and sentiment analysis to identify influential posts and characterize the discourse, and then used both inductive and deductive coding to detect context-specific narratives that would not be measured by standard sentiment dictionaries. While NLP and qualitative coding methods have been widely used throughout the literature for identifying and characterizing misinformation (see the following for examples: [13, 2325]), this paper combines these methods into a streamlined approach that can be utilized for rapidly characterizing nuanced themes and emerging narratives within large scale online discourses that requires significantly less burden on human annotation, as illustrated in the flow diagram in S1 Appendix. The hybrid method utilized in this study also has the potential to generate nuanced training data for machine learning classifier models without the need for annotating thousands of posts.

This study aims to further bridge the gap between computational and qualitative methodologies by demonstrating how NLP and manual annotation can be used synergistically to emphasize the positive aspects of each approach while minimizing their respective drawbacks. In order to assess the utility of this methodology, themes generated by this approach will be compared for consistency with 5G-COVID conspiracy themes previously identified on Twitter [2628], Facebook [29], and Instagram [30]. Further, this study will extend the current literature by providing in-depth characterization of discourse focusing on correcting false information, which can be used to inform counterstrategies for online misinformation propagation.

1.2 Using machine learning to detect misinformation

Even before the proliferation of misinformation spurred by the emergence of COVID-19, there have been several research efforts over the past decade demonstrating that machine learning approaches can be effective methods for detecting misinformation. Previous work has developed models with at least 70% accuracy for correctly classifying misinformation across multiple platforms and topics, including Twitter posts related to the zika virus [31, 32], YouTube videos about prostate cancer [33], and false health-related information on medical forum posts [34] and Chinese posts on the platforms Weibo and WeChat [35]. Social media users can be classified as well, as shown by Ghenai and Mejova [36], who were able to identify with a 90% accuracy Twitter users prone to sharing false information promoting ineffective cancer treatments.

Many investigations using misinformation classifiers also examine the role of feature selection (i.e., what variables to include in the model) on prediction accuracy. While classifiers can effectively detect misinformation from only using textual features of the post, which are typically keywords associated with specific conspiracy theories or false narratives [37, 38], the inclusion of variables accounting for properties of the post beyond the content of the specific message, such as emotional sentiment, has been shown to improve model performance across multiple studies [17, 18, 20]. Deriving sentiment features from posts only requires basic arithmetic, where sentiment scores are calculated by first counting the number of words associated with a sentiment category and then dividing that count by the total number of words within the post [39]. Each sentiment category includes a dictionary, which is the list of words selected to represent the category. Sentiment dictionaries can measure a wide array of topics, such as emotional affect states (e.g., anger, anxiety, joy), biological processes (e.g., health, death), cognitive processes (e.g., certainty, causation), time orientations (e.g., focus on past, present, or future), and linguistic properties (e.g., number of 1st-person pronouns used) [40].

Assigning words to sentiment dictionaries can be accomplished using data-driven approaches, as shown by Stanford’s Empath project that derived over 200 classification categories from analyzing more than 1.8 billion words of modern fiction [41]. However, for online misinformation research, psychometrically validated sentiment dictionaries provided by the software Linguistic Inquiry and Word Count (LIWC) have been widely used for characterizing misinformation discourse [42, 43] and calculating sentiment features across multiple studies that evaluate performance of misinformation classifiers [4446]. A likely reason sentiment features are effective predictor variables is due to the fact that texts containing misinformation and conspiracies often have an emotional signature such as higher frequency of anger and anxiety words [15, 16, 43], although recent work indicates that emotional valence can vary depending on the type of misinformation being studied [47].

Despite the demonstrated utility of machine learning approaches for static misinformation detection, sentiment analysis and misinformation classifiers are both limited in their ability to adapt to the ever-evolving nature of human language. That is, in the same way that Gestalt psychology demonstrates that the perception of a color shifts depending on what other colors surround it [48], the shades of meanings from words also shift based on the other words surrounding it within a sentence or paragraph. Since word meanings can change due to differences in context-use and language shifts throughout time, results from sentiment scores and classifier models may not always reflect the actual semantic meaning of texts. These limitations can be clearly depicted in other applications such as automated detection models for abusive language as demonstrated by Yin and Zubiaga [49], who show that these classifier models are limited in their generalizability for other abusive language datasets, and may over-rely on the appearance of keywords such as slurs and profanity. While slurs and profanity can be strong predictors of online abusive language, abuse can also be expressed using implicit meanings and subtext, which results in models that overlook abuse in posts not including slurs and profanity. It is also possible that a post containing profanity keywords is not abuse at all, such as instances of teasing between friends, yet a model would falsely label as abuse [49]. Due to the continually evolving nature of language and context-dependent meanings, detecting misinformation and conspiracies only using machine learning models faces similar semantic challenges.

Interpreting sentiment scores can be further complicated when accounting for the fact that the discussion topic can also influence the average emotional tone of a discourse. For example, discussions about a pandemic may have higher percentages of negative affect words (e.g., “death”, “tragedy”) compared to lighter conversation topics such as gardening. Hence, it is difficult to determine what is an appropriate threshold for meaningful sentiment scores across topics. While this limits what researchers can infer from these analyses, this study addresses this limitation by using a relativistic interpretation of sentiment scores. More specifically, discourse will be marked as high in a sentiment category based on whether it is in the 90th percentile of scores within the corpus. Using the 90th percentile as a threshold marker allows us to account for the specific context of the 5G discourse when making judgements for determining what comprises a high level of sentiment. This approach is also adaptable across a wide array of topics since percentiles indicate which cases are high or low for a given metric based on the sample distribution. To further illustrate this point, a post containing 5% of death-related words may be in the 95th percentile for discourse about gardening, making it a “high” amount, but be within the 50th percentile for pandemic-related discourse, making it a typical percentage within the context of that corpus. Additionally, this study incorporates a qualitative coding approach to account for contextual information that is typically overlooked in sentiment analysis but recognizable by a human coder.

1.3 Inductive and deductive approaches for content coding

In recent years there have been calls for researchers to recognize the impact of contextualization when interpreting data, such as accounting for changes in semantics within a dataset over time [50]. When identifying emerging narratives within online discourse, where meanings of text vary due to novel co-occurrences of words, the boundaries between noise and relevant signal are constantly shifting. Compared to supervised machine learning approaches, which requires an existing training dataset of annotated posts to be effective, human coders are more adaptable to updating their background knowledge of events, making them effective at recognizing text containing previously undocumented narratives. To account for contextualization factors, our study utilized both inductive and deductive analytic techniques to examine tweets related to conspiracy discourse stating a relationship between COVID-19 and 5G technology.

Based in grounded theory [51], inductive coding is an iterative data analytic process centered on constant examination and comparison, which allows for theory development and explicit coding procedures [52, 53]. The primary benefits of an inductive approach is that it allows researchers to code texts using labels that are both aligned with the data and free from the influence of extant concepts, as well as detect tacit elements or connotations of the data that may not be apparent from a superficial reading of denotative content [54]. For the current study, inductive coding was used to identify narratives and rumors prominent within the 5G discourse.

Deductive coding, on the other hand, refers to a top-down coding process intended on testing whether data coincide with existing assumptions, theories, or hypotheses [55]. For the deductive coding scheme in the current study, we coded for whether posts contained misinformation or misinformation corrections based on whether it made statements claiming that 5G wireless technology causes COVID-19, or explicitly refuted the conspiracy. The criteria for classifying 5G-related misinformation and corrections were adapted from previous frameworks identifying COVID-related misinformation [5658]. The researchers also coded for whether tweets expressed a positive, negative, or neutral stance towards 5G conspiracies. Coding for the user’s stance makes it possible to trace and compare general sentiment across topics without having to account for specific themes.

By using both inductive and deductive coding schemes in our content analysis, we aimed to extract and interpret the underlying patterns of COVID-19 misinformation in a more comprehensive manner that are both consistent with the textual data and aligned with existing conceptual frameworks. This approach also allows us to track both general conceptual categories (e.g., misinformation) and more case-specific incidents (e.g., religious pastor claiming 5G causes COVID-19), or in other words, seeing both “the forest and the trees.”

1.4 Combining unsupervised machine learning and content analysis approaches

Previous studies that only use content analysis when investigating social media activity can require the coding of hundreds, if not thousands of posts, which can be an extensive cost in time and resources. Fortunately, unsupervised machine learning and NLP approaches have been used to aid in content analysis on social media, which includes the detection of illicit drug sales or promotions [5962], self-report symptoms on social media [63], and identifying prominent themes in COVID-related misinformation [56]. These approaches group posts together based on textual similarity, which can help filter out irrelevant topic clusters based on keywords [64], or organize the posts together to facilitate the identification of higher-level themes [56]. Since most social media discourse is driven by a small number of highly active users while the majority of users typically engage in passive behaviors such as browsing [6568], characterizing discourse based on the most highly shared posts can be an effective approach for assessing public response to an issue due to most users not often generating their own content. This is further demonstrated in a recent study that conducted a social network analysis of Twitter users, which showed that the propagation of a COVID-related conspiracy theory was initiated by only a handful of prominent accounts during the early stages of the pandemic [69].

Another advantage of applying topic modeling to social media data is that it allows for other prominent sub-themes to be identified from the topic clusters that might have been overlooked by just assessing a more general sampling of the most shared posts, such as the most retweeted tweets. This is demonstrated in previous work by the authors [56] who were able to identify themes related to the misuse of scientific authority within Twitter misinformation discourse after applying biterm topic modeling (BTM) to the data and then coding the top 10 retweeted tweets from each topic cluster. Using this approach, informally referred to as “BTM + 10,” we were able to build on themes identified in their previous characterization of misinformation discourse based on a more general sampling of the top 100 retweeted tweets [58]. While BTM + 10 does not cover the content in every post assigned to a topic, the top 10 most retweeted tweets were able to account for at least 50% of the total tweet volume of each topic cluster within the context of the hydroxychloroquine twitter discourse [56], indicating that influential posts tend to account for a majority volume of tweets within a discourse. It is also worth noting that uncoded tweets are still assigned to a topic cluster based on textual similarity, which increases the likelihood that it would be similar in message content to the most influential posts that were assigned to the same cluster. Increasing the number of coded posts from the 10 most retweeted tweets to 15 or 20 can also account for greater tweet volume without adding substantial burden on the coders. In the current study, we wish to build on the BTM + 10 methodology as carried out in previous work [56] by incorporating sentiment analysis when characterizing 5G-COVID conspiracy discourse on Twitter. Themes identified from previous studies examining 5G conspiracy discourse will be compared to assess the validity of the methodology proposed in the current study. For readers interested in adapting this methodology, see the flow chart diagram in S1 Appendix that further outlines the current approach.

1.5 5G conspiracy theories during the COVID-19 pandemic

Beginning in early April 2020, there had been reports that telecom engineers were facing verbal and physical threats, and that at least 80 mobile towers had been burned down in the United Kingdom (UK), actions fueled by false conspiracy theories blaming the spread of COVID-19 on 5G wireless signals [70]. In order to increase understanding of these destructive acts, recent work has shown associations between belief in 5G-COVID conspiracies with states of anger and greater justification of real-life and hypothetical violence alongside greater intent to engage in similar behaviors in the future [71].

Public figures and celebrities who, whether with deliberate malintent or not, share rumors and falsehoods to their large groups of followers [7275] are also involved in 5G-COVID conspiracy discourse, as shown in a recent study characterizing discourse on Facebook that identifies celebrities and religious leaders as 5G-COVID conspiracy propagators [29]. Bruns et al. (2020) also identify prominent 5G conspiracy theories such as claims stating that 5G reduces the ability of the human body to absorb oxygen, and that 5G is related to a complex agenda involving bioengineered viruses and deadly 5G-activated vaccines led by elite figures such as Bill Gates, George Soros, the World Health Organization (WHO), and secret-society organizations like the Illuminati [29]. From these findings, Bruns et al. (2020) conclude that 5G conspiracy theorists had retrofitted the new information emerging about the virus and its effects on human health into pre-existing worldviews, beliefs, and ideologies to further propagate conspiracist narratives [29]. This is consistent with findings from another 5G conspiracy study conducted on Twitter, which shows that the conspiracies were built on existing ideas set against wireless technologies [28]. On Twitter more specifically, researchers found that videos played a more crucial role in 5G rumor propagation than posts [28], and other work has examined how spatial data has been misconstrued by conspiracists, as shown with the promotion of maps that assert false correlations between the distribution of COVID-19 cases and installations of 5G towers [27].

Another study that used social network analysis found that influential accounts tweeting 5G-COVID conspiracies tended to form a broadcast network structure resembling the structure most typical for accounts from mainstream news outlets and celebrities that are frequently retweeted [26]. A content analysis from this study also shows that over a third of randomly sampled tweets contained views claiming that COVID and 5G were linked, and that there was a lack of an authority figure who was actively combating said misinformation [26]. In order to examine the authors of influential tweets within 5G discourse, the current study will also content code affiliations from the most retweeted user accounts based on publicly available profile data.

2. Methods

2.1 Data collection and analysis overview

A total of 256,562 tweets were collected from the public streaming Twitter API per the terms available at the time of the study using keywords “5G” and covid-related words such as “coronavirus”, “covid-19” between March 25th and April 3rd 2020. We chose this time frame as it represents a period when the 5G conspiracy theory first became prominent, as shown in the spike in volume for “5G” posts in Fig 1. All personal identifiable information from tweets was removed in the reporting of the results to preserve anonymity. We note that due to the change in ownership, API policies, and name of the platform (Twitter has been renamed “X”), the terms and conditions of the streaming API used for data collection for this study are no longer applicable for current studies. IRB approval was not required as all data collected in this study was available in the public domain and results from the study have been deidentified and anonymized. The dataset and R syntax used to generate the results can be found at the following Open Science Framework (OSF) link: https://bit.ly/5G_Conspiracies.

Fig 1. Number of Tweets with keyword “5G” through time highlighting timeframe of current study.

Fig 1

To analyze the relatively large volume of tweets collected in this study, we used the biterm topic model (BTM), an unsupervised machine learning approach using natural language processing (NLP) further described in the next section, to extract themes from text of tweets as used in prior studies examining COVID-19 topics on social media [10, 56, 63, 64]. The top 10 most retweeted tweets associated with each topic cluster (i.e., the “BTM + 10” approach previously mentioned) were coded using a deductive coding scheme adapted from previous COVID-19 misinformation work [56, 58] to classify posts on whether they contain misinformation, or a misinformation correction (further discussed in Section 2.3). For tweets that may not contain misinformation but still support the notion that 5G signals cause COVID-19 infections, positive and negative stances were coded to assess whether a tweet supported or opposed the conspiracy (Table 1). While the metric of stance is traditionally referred to as “sentiment” in related social media research, this current study will refer to it as “stance” to prevent potential confusion with NLP sentiment analysis metrics further described in the next section. An inductive coding scheme was also used to characterize themes and narratives for both misinformation and correction discourses.

Table 1. Examples of content coded tweets (paraphrased and redacted to retain anonymity).

Misinformation Categories Misinformation:
1. #5G produces the same symptoms as this supposed #coronavirus or #COVID19. Watch this video [LINK]
2. RT! 5G is the real silent killer, not the Corona Virus!!!
Misinformation Correction:
1. Scientists say any suggestion that coronavirus and 5G are linked is “complete rubbish” and biologically impossible
2. I can confirm that 5G is in no way giving people #coronavirus because that would require converting radio waves into organic molecules. This is pretty much God-level ability, and anyone capable of it would be running the planet already
Stance towards 5G Conspiracy Positive Stance:
1. #Ireland. You’re about to be fried. We have pleaded with people to listen to the 23,000 scientists who want #5G banned
2. Unnamed 5G #Whistleblower Claims That People Are Being Infected With #Coronavirus Via #Covid19 Tests
Negative Stance:
1. Imagine being months into a relationship with a person you could see a potential future with, and one morning they send you links and paragraphs on the 5G conspiracy being the source of the coronavirus.
2. If you’re into this 5G conspiracy theory, here are some other just as likely causes of Covid-19 to look into: ferns, Ed Sheeran, clams and clam sauce, those disposable flossers, Golden Retrievers & hemp granola.

2.2 Biterm Topic Model (BTM)

Unsupervised topic modeling strategies, such as BTM, are methods particularly well suited for sorting short text (such as the 280-character limit for tweets) into highly prevalent themes without the need for predetermined coding or a training/labelled dataset to classify specific content. This is particularly useful in characterizing large volumes of unstructured data where predefined themes are unavailable, such as in the case of emerging social movements, novel disease outbreaks, and other emergency events where information changes rapidly [10, 56, 60, 61, 63, 64, 76, 77]. The corpus of tweets containing the 5G keywords was categorized into highly correlated topic clusters using BTM based on splitting all text into a bag of words and then producing a discrete probability distribution for all words for each theme that places a larger weight on words that are most representative of a given theme [78]. While other NLP approaches use unigrams or bigrams for splitting text, BTM uses ‘biterms’, which is a combination of two words from a text (e.g., the text “go to school” has three biterms: ‘go to’, ‘go school’, ‘to school’) and models the generation of biterms in a collection rather than documents [79].

BTM was used for this study because biterms directly model the co-occurrence of words, which increases performance for sparse-text documents such as tweets. Conducting BTM analysis is done initially by setting the BTM topic number (k) and “n” words (for the first round of analysis we set at k = 10, n = 20 to cover several possible misinformation topics that might be present in the corpus). A coherence score is then used to measure how strong the top words from each topic correspond to its respective topic. For this study, the model with k = 20 was chosen because it had the highest coherence score compared to other iterations tested. All data collection and processing were conducted using the programming language Python.

2.3 Deductive and inductive coding schemes

In order to characterize highly prevalent misinformation and conspiratorial narratives in the corpus, the top 10 most retweeted tweets from all 20 BTM topic outputs were extracted and manually coded for relevance first using a deductive coding scheme adapted from existing COVID-19 misinformation themes from the literature [5658], and then coded again using an inductive approach identifying context specific themes related to 5G. While misinformation and conspiracies are distinct concepts, the current study will refer to both as ‘misinformation’ within the analysis for brevity.

In total, 200 unique tweets were reviewed by coders. Tweets were classified as misinformation if they contained declarative statements claiming that 5G causes COVID-19, or statements supporting the conspiracy from sources that convey scientific authority such as from medical experts, scientists, or scientific studies [56]. A tweet was considered a misinformation correction if it explicitly opposes the 5G conspiracy and provided information countering the claims about 5G causing COVID-19. Tweets were also annotated for stance in relation to 5G conspiracy theories, with 1 indicating positive stance, -1 indicating negative stance, and 0 if the tweet only reports information about 5G without stating an opinion, exhibits neutral user sentiment, or is not directly related to the 5G COVID conspiracy discourse.

An inductive coding approach was then used to sub-code for reoccurring themes and narratives associated with each topic cluster that is unique to 5G-related misinformation (see Table 2 below for a description of each identified sub-theme and example tweets). Of the total 200 tweets used for inductive coding, the first and third author divided the sample in half to identify themes. Once theme labels were generated by each coder, both coders met to compare inductive coding labels and combine overlapping themes. All 200 tweets were then reviewed again to classify them based on the finalized coding scheme. Using this approach, tweets can be categorized with the same theme label from inductive coding but classified as misinformation or a correction based on the deductive framework. For example, a tweet labeled as misinformation and the inductive coding theme causative explanation (see Table 2 for description) indicates that the tweet is making a causal claim about a correlation between COVID-19 and 5G. However, a tweet labeled as a misinformation correction but labeled as the same theme indicates that the tweet is making causal claims refuting correlations between COVID-19 and 5G technology. An advantage of this approach is that it showcases overlap in discourse themes between misinformation and correction topic clusters.

Table 2. Inductive coding themes.

Theme Classification Label Description Example Tweet (paraphrased to preserve anonymity)
Causative explanations of 5G & COVID Caus Exp Claims about the causal effects of 5G and COVID-19, sometimes using scientific language (e.g., describing cause and effect mechanisms) “[Redacted] declares that COVID-19 is not a virus at all, but a reaction to the radiation created by 5G technology.”
Medical authority Med Aut Claims made from or endorsed by medical authority figures (such as medical scientist, doctor, or scientific study) “Dr. [Redacted], M.D. hypothesizes that Coronavirus may be history repeating itself & caused by 5G.”
Religious figure Pastor Messages made by religious pastors about the connection between 5G and COVID, or messages in response to the pastor’s remarks “Pastor [Redacted] said corona virus doesn’t kill but 5G does. He said telecom companies are installing 5G networks in lagos and Abuja that’s why federal government enforced a lockdown on both states”
Celebrity Celeb Statements addressing celebrities who state their stance on 5G-COVID conspiracy theories “[Celebrity Singer] out here spreading conspiracy theories that corona virus is a result of 5G networks”
Antivax Anti-Vax Claims that vaccines, including future COVID-19 vaccines are linked to conspiracy “The entire Pandemic crisis has 3 components. . . First Vaccines to increase metal levels in body. Then bombard the body with high energy radiation to destroy the immune system. Then launch the virus to create the Pandamic. Check the date of 5G launch in Wuhan. . .”
Negative health effects Neg Health Statements claiming exposure to 5G causes negative health effects “5G just rolled out in Europe, China, and North America. Microwave radiation cripples immunity and increases inflammation. It’s the perfect fuel for a pandemic. We have to stop it. 5G: Fueling COVID”
Video Video References to video being shared on YouTube about 5G conspiracies “Corona = 5G weapon used as disguise of virus? This video is VERY shadow banned. It won’t show up to my history.…”
Technology correlations with disease outbreaks Tech Corr Statements with timelines showing disease outbreaks alongside the release of new radio wave technologies “2003 - 3G introduced to the world
2003—SARS outbreak
2009 - 4G introduced to the world
2009—Swine flu outbreak
2020 - 5G introduced to the world
2020—Coronavirus outbreak”
Huawei Huawei Story about the UK government pulling out of 5G contract with Chinese communications company Huawei “Britain pulls out of 5G contract with Chinese firm Huawei after test kits were found contaminated with Corona virus…”
Group of elites Elite Statements referring to group of elites, which typically included figures such as Bill Gates, Elon Musk, and the illuminati “If people knew what was unfolding & what satan &his Illuminati agents, his disciples like
@elonmusk & @BillGates & the proponents of the New World Order have in mind &are doing with 5G, coronavirus & the coming Anti-coronavirus vaccines, everyone would join forces & PRAY!”
Towers on fire Tower Fire Discussions about people setting 5G towers on fire “UK mobile phone masts torched and engineers abused over "baseless" theories linking coronavirus to 5G”
Geography comparisons Geo Posts comparing maps of 5G coverage and COVID outbreaks, or mention areas that have 5G coverage and no COVID cases or vice-versa “Let’s take a quick look at Italy. Left: Coronavirus cases Right: Number of 5G deployments with commercial available tech”

2.4 Content coding of Twitter account profiles

Twitter profiles of accounts that produced the top 10 most retweeted tweets for each BTM cluster output were content coded to investigate publicly self-reported occupation among these influential users who were active in the Twitter 5G conspiracy discourse during the study period. Publicly available metadata from user account profiles were retrieved and coded to determine whether descriptions stated that they were a medical doctor or scientist, a religious leader, a government official, or affiliated with the media (e.g., a journalist, TV, or radio personality).

2.5 Calculating LIWC sentiment scores

Sentiment scores were calculated using Linguistic Inquiry and Word Count (LIWC) and reflect the percentage of words within a post that correspond to a given sentiment category. The sentiment scores were calculated to assess language related to: Analytic thinking (metric of logical, formal thinking), Clout (language of leadership, status), Authenticity (perceived honesty, genuineness), and Netspeak (internet slang), cognitive-processes related to information evaluation (i.e., Causation, Discrepancy, Tentative, Certitude), Emotional Affect (i.e., Affect, Positive Tone, Negative Tone, Emotion, Positive emotion, Negative emotion, Anxiousness, Anger, Sadness), social- and health-related topics (i.e., Prosocial behavior, Interpersonal conflict, Moralization, Health, Illness, Death, and Risk), and time-related sentiments (i.e., Time, Past focus, Present focus, Future focus). Since some tweets receive a much greater proportion of engagement within a discourse than others, average sentiment scores for each topic cluster were also weighted based on number of retweets received. This allowed us to characterize emotional sentiment of a topic cluster based on the most influential posts, rather than by posts that received low exposure. Due to the lack of a normal distribution of sentiment scores, we applied a Kruskal-Wallis nonparametric approach and Dunn’s post-hoc test to detect statistically significant differences across BTM topic clusters, with a Bonferroni Correction to account for multiple comparisons.

3 Results

3.1 Content analysis and characterization

From the 256,562 tweets analyzed in this study, Table 3 shows all BTM topics generated and assessed. In this study, the majority of topics have at least 50% of their tweet volume composed of the top 10 tweets with highest volume of retweets, indicating that the top 10 retweets characterize a substantive volume of tweets assigned to each topic. There were 4 topics below 50%, with the lowest percentage being 40.9%. While this still accounts for a substantive percentage of tweet volume, it should be noted that there is more uncertainty associated with the characterizations based on the metrics derived from the top 10 most retweeted tweets for these topic clusters. Topics were further classified as high in misinformation or high in misinformation corrections if the topic had at least 33% of top 10 retweeted tweets associated with the respective categories. Topics were also labeled high in both (i.e., Mixed) if it contained at least 33% of misinformation and corrections, and low for the remaining topics. A threshold of 33% was chosen for this analysis since it indicates a substantive number of tweets for each topic. Out of the 20 topics in this study, 9 were classified as high in misinformation, 5 as high in misinformation corrections, 2 as high in both misinformation and corrections, and 4 as low in both. The topic clusters low in both were not chosen for further analysis.

Table 3. Properties of BTM clusters.

BASED ON ALL TWEETS PER CLUSTER BASED ON TOP 10 RETWEETED TWEETS BY CLUSTER
TOPIC CLUSTER ID Total Tweets Unique tweets Top10% Twitter URL YouTube URL Avg Stance Stance Level Misinfo % MisCorr %
0 5669 1617 58.2% 48.6% 16.6% 0.766 Strongly Positive 94.0% 0.0%
1 14122 3529 54.7% 77.6% 1.2% -0.166 Mixed 15.1% 43.8%
2 66434 15967 71.9% 48.4% 1.7% -0.035 Mixed 11.2% 1.1%
3 14857 4490 48.7% 33.8% 1.2% -0.793 Strongly Negative 0.0% 80.3%
4 9703 1988 57.2% 76.5% 4.8% 0.755 Strongly Positive 87.8% 12.2%
5 5939 2023 63.6% 61.9% 1.4% -0.345 Negative 59.6% 35.1%
6 2217 703 80.5% 46.9% 0.1% 0.118 Mixed 43.6% 0.0%
7 1472 1358 50.3% 49.6% 35.6% 1 Strongly Positive 100.0% 0.0%
8 2429 892 72.7% 69.5% 9.8% 0.81 Strongly Positive 69.9% 3.1%
9 7030 3219 46.6% 71.4% 0.5% -0.296 Negative 31.7% 52.2%
10 2551 727 80.7% 84.7% 3.7% 0.606 Positive 60.6% 0.0%
11 2358 345 89.1% 79.4% 0.6% 0.926 Strongly Positive 92.6% 0.0%
12 20520 3594 47.0% 76.4% 2.4% 0.448 Positive 12.4% 0.0%
13 10478 700 88.1% 73.7% 1.1% 0.97 Strongly Positive 92.7% 0.0%
14 4765 1510 61.2% 78.0% 4.2% 0.46 Positive 27.9% 3.2%
15 5866 770 86.5% 68.7% 10.9% 0.162 Mixed 17.2% 5.7%
16 1523 628 68.0% 56.5% 22.3% 1 Strongly Positive 93.3% 0.0%
17 18512 4788 54.2% 31.9% 2.3% -0.871 Strongly Negative 0.0% 83.8%
18 50014 19149 40.9% 61.5% 1.3% -0.847 Strongly Negative 4.4% 65.2%
19 10103 2625 76.3% 37.3% 2.9% -0.098 Mixed 41.2% 40.4%

Columns “Misinfo %” and “MisCorr %” indicate the percentage of tweets categorized as misinformation or misinformation corrections respectively and weighted by number of retweets. The column “Avg Stance” is based on the average stance scores of the top 10 retweeted tweets for each topic. Additionally, the column “Top10%” indicates the percentage of total tweets for each topic that is accounted for by the top 10 retweeted tweets.

Table 4 shows each topic cluster with the themes identified from inductive coding, as well as a list of LIWC sentiment categories in which the topic scores in the 90th percentile. Most of the topic clusters score in the 90th percentile for at least one sentiment category, indicating that the topic clusters cover a wide range of emotional sentiments. The results in Table 4 show that themes were spread across topics, with sentiment varying across the discourse as well. While some topic clusters are predominately associated with one theme, as seen in topics 7, 8, 11, and 12, other topic clusters have multiple themes. The misinformation topic with the highest number of themes is topic 10, which includes themes of causative explanations, medical authorities, negative health effects, and video. Within the correction discourse, medical authority was paired with themes related to 5G towers on fire and causative explanations in topics 3 and 18. Among the topic clusters that were predominately misinformation, none of the correction tweets corresponded to any of the inductive coding themes. This suggests that misinformation discourses are closed off and less likely to have involvement from those providing corrections. However, this is not consistent with topic clusters high in corrections, as shown in topic 1 where corrections that include causative explanations are in the same discourse as anti-vax misinformation tweets, and topic 9, which has both misinformation and correction tweets that mention religious figures. These results could indicate misinformation discourse is more likely to interfere with correction discourse. The correction discourse also does not mention themes for Tech Corr (i.e., claims that wireless technology advancements are correlated with disease outbreaks), Elite (i.e., claims that figures like Bill Gates and Elon Musk are promoting COVID-19 with 5G technology), Huawei (i.e., the UK government pulling out a contract from a Chinese communications company due to health concerns over 5G), anti-vax (i.e., claims about future COVID-19 vaccines having adverse health effects) and Geo (i.e., maps showing overlapping distribution between COVID cases and 5G cellular towers), indicating that the corrections are not addressing specific misinformation narratives.

Table 4. Inductive coding themes and LIWC sentiment scores across topics.

TOPIC CLUSTER ID MISINFO THEME (%) CORRECT THEME (%) LIWC SENTIMENT
HIGH IN MISINFORMAITON 0 Video
(94.6%)
Neg Health
(5.4%)
None Netspeak
Negative Tone
Emotion
Negative Emotion
4 Anti-Vax
(34.5%)
Elite
(29.5%)
None Clout
Moral
Future Focus
6 Video
(37.1%)
None Sadness
Conflict
Risk
7 Video
(91.8%)
None Time
Past Focused
8 Anti-Vax
(71.2%)
None Health
Illness
10 Caus Exp
(12.0%)
Med Aut
(4.6%)
Neg Health
(7.0%)
Video
(2.6%)
Tech Corr
(3.1%)
None Discrepancy
Health
Death
11 Tech Corr
(100.0%)
None Analytic
Illness
13 Huawei
(98.1%)
None Analytic
Anger
Past Focused
16 Caus Exp
(78.4%)
Med Aut
(21.6%)
None Anxiety
Pro-social
HIGH IN CORRECTIONS 1 Anti-Vax
(26.6%)
Neg Health
(49.7%)
Caus Exp
(86.1%)
Causation
Tentativeness
Present Focused
3 None Med Aut
(67.3%)
Tower Fire
(32.7%)
None
9 Pastor
(30.9%)
Pastor
(76.9%)
Clout
17 None Celeb
(60.8%)
Negative Tone
Emotion
Negative Emotion
18 None Caus Exp
(17.3%)
Med Aut
(54.5%)
Certitude
Anger
MIXED 5 Geo
(50.5%)
None Authentic
Sadness
Time
19 Video
(24.9%)
Med Aut
(75.1%)
Pastor
(77.1%)
Celeb
(22.9%)
Positive Tone
Positive Emotion
Moral
Risk

The column “Misinfo Theme (%)” lists the themes that appear in each topic with the percentage indicating the proportion of coded misinformation tweets that were also assigned the theme label. The column “Correct theme (%)” is similar to the previous column, except that it shows the proportion of coded correction tweets assigned to a theme.

Table 5 groups each topic cluster together based on misinformation classification and shows the percentage of coded tweets corresponding to each inductive coding theme. Among misinformation topics, the themes Huawei (33.5%), video (14.1%), and anti-vax (9.9%) were the three most prominent. The three most prevalent themes for corrections were medical authority (22.1%), causative explanations (10.3%), and celebrities (10.1%). The difference in prevalent themes further illustrates that the two conversations focused on differing conversation topics. Within misinformation discourse, the conspiracy theme Huawei claims that the UK government pulled out of a 5G contract due to concerns about it causing COVID, while the actual circumstance was due to concerns about surveillance from the Chinese government. In this case, a situation with a kernel of truth was misconstrued to perpetuate a false claim. The other themes video and anti-vax indicate conversations occurring off Twitter and plant the seed for future distrust in medical efforts. While the correction discourse made general refutations against the false correlation between 5G and COVID-19, it mostly did not address the specific narratives that were prominent in misinformation discourse, which could mitigate the effectiveness of the corrections. For mixed topic clusters, religious figures (29.8%), medical authority (20.8%), and geographic comparisons (9.9%) were the most common themes. It is possible that themes such as religious figures could be associated with conflicts between people’s faith and public health guidelines, which may contribute to the high volume of both misinformation and corrections. It is also possible that the appearance of medical authorities both supporting and refuting claims that 5G causes COVID-19 could lead to ambiguity and confusion among users.

Table 5. Themes from inductive coding by misinformation classification.

TOPIC CAUS EXP MED AUT PASTOR CELEB ANTI-VAX NEG HEALTH VIDEO TECH CORRE HUAWEI ELITE TOWER FIRE GEO
MISINFO 3.5% 1.0% 0.9% 0.0% 9.9% 1.0% 14.1% 7.7% 33.5% 5.6% 0.0% 0.0%
CORRECT 10.3% 22.1% 3.9% 10.1% 0.7% 1.1% 0.6% 0.0% 2.7% 0.0% 5.7% 0.0%
MIXED 0.0% 20.8% 29.8% 6.2% 0.0% 0.0% 6.9% 0.0% 0.0% 0.0% 0.0% 9.9%

3.2 LIWC sentiment scores

LIWC sentiment scores were also averaged across BTM topics based on misinformation classifications as shown in Table 6 and visualized as bar graphs in Figs 26. Topics with low amounts of misinformation and corrections were kept in this analysis for comparison. Sentiment categories listed in Table 6 that had statistically significant differences across all misinformation classifications are marked with an asterisk (*). Among Text Analytic sentiments (i.e., Analytic, Clout, Authentic, and Netspeak), misinformation topics on average scored higher than corrections across all categories except Authentic, with these differences being statistically significant (p < .001). These differences are most pronounced when comparing Analytic (78.74% misinformation vs 67.66% correction) and Clout sentiments (46.57% vs 41.45%), indicating that misinformation topics were more likely to use words reflecting logic and social status compared to corrections. The mixed BTM topics that contained high levels of both misinformation and corrections scored second highest in Analytic (77.23%) and Clout (46.35%) sentiments but lowest in Authentic (18.24%) against all other classifications. In contrast, topics that contained low amounts of both misinformation and corrections on average scored the lowest in Analytic (59.06%) and Clout (29.54%), but highest in Authentic (27.06%) and Netspeak (6.44%).

Table 6. Average LIWC sentiment scores (i.e., percentage of words corresponding to each sentiment category) across misinformation classifications.

Text Analytics Topic Misinformation Correction Mixed Low
Analytic* 78.74 67.66 77.23 59.06
Clout* 46.57 41.45 46.35 29.54
Authentic* 20.49 21.34 18.24 27.06
Netspeak* 5.10 4.71 4.64 6.44
Cognitive Topic Misinformation Correction Mixed Low
Cause* 0.78 2.06 1.58 1.54
Discrep* 0.61 1.11 0.98 1.98
Tentat* 0.75 1.66 1.30 2.54
Certitude 0.07c,l 0.52 mi,m,l 0.12c,l 2.02 mi,c,m
Affect Topic Misinformation Correction Mixed Low
Affect (general)* 7.03 6.95 7.13 7.17
Positive Tone* 0.67 0.82 1.30 0.81
Negative Tone* 6.17 5.90 5.52 5.84
Emotion (general) 5.26c,m 4.92 mi,m,l 5.14 mi,c,l 4.81c,m
Positive Emotion* 0.06 0.14 0.40 0.23
Negative Emotion* 5.15 4.69 4.66 4.54
Anxiousness* 0.06 0.06 0.04 0.09
Anger 0.08 c,m,l 0.18 mi,m,l 0.01mi,c 0.03mi,c
Sadness 0.01c,m 0.02mi,l 0.02mi,l 0.01c,m
Social/Health-related Topics Topic Misinformation Correction Mixed Low
Prosocial 0.18 c,m,l 0.29 mi,l 0.25mi,l 0.22 mi,c,m
Conflict* 0.26 0.24 0.41 0.64
Moral* 0.24 0.32 0.48 0.16
Health* 4.80 4.62 4.63 4.61
Illness 4.18 c,m,l 4.21mi,l 4.08 mi,l 4.35 mi,c,m
Death* 0.33 0.12 0.29 0.52
Risk* 0.16 0.23 0.52 0.14
Time Topic Misinformation Correction Mixed Low
Time (general)* 2.66 1.78 2.65 1.45
Past Focus 3.11 c,m,l 1.51 mi,l 1.59mi,l 1.46mi,c,m
Present Focus* 3.59 4.98 4.75 7.16
Future Focus* 0.53 0.58 0.47 0.53

Categories with

“*” indicate statistical significance (< .01) across all classifications using Bonferroni corrections to adjust for multiple comparisons. For categories without an “*”, significance differences between classifications are depicted if the row contains the following subscripts: mi = Misinformation, c = Corrections, m = Mixed, and l = Low.

Fig 2. Average percentage of text analytic sentiment across misinformation classifications.

Fig 2

Fig 6. Average percentage of time-related sentiment across misinformation classifications.

Fig 6

Fig 3. Average percentage of cognitive sentiment across misinformation classifications.

Fig 3

Fig 4. Average percentage of affect and emotional sentiment across misinformation classifications.

Fig 4

Fig 5. Average percentage of social and health-related sentiment across misinformation classifications.

Fig 5

For sentiments reflecting cognitive processes (i.e., Causation, Discrepancies, Tentativeness, and Certitude), corrections on average scored significantly higher than misinformation topics across all sentiment categories, and was highest in Causation words (2.06%) compared to all other topic classifications (p < .001). Topics low in both misinformation and corrections scored highest in Discrepancy (1.98%), Tentative (2.54%), and Certitude (2.02%) sentiments.

Among sentiment for general Affect, misinformation topics on average score significantly higher than corrections (7.03% vs 6.95%, p < .001), and are also higher in general Emotion (5.26% vs 4.92%, p < .001), Negative Emotion (5.15% vs 4.96%, p < .001), and Negative Tone (6.17% vs 5.90%, p < .001). However, misinformation scores lower than corrections in Positive Tone (0.67% vs 0.82%, p < .001), Positive Emotion (.06% vs .14%, p < .001), and Anger (0.08% vs 0.18%, p < .001). There are also statistical differences between misinformation and correction topics for Anxiousness (0.063% vs .058%, p = .023) and Sadness (0.01% vs 0.02%, p < .001), although the magnitude of the differences is fairly small.

For sentiment categories measuring social and health-related topics, correction topics on average scored higher than misinformation topics for Prosocial (0.29% vs 0.18%, p < .001), Moral (0.32% vs 0.24%, p < .001), Illness (4.21% vs 4.18%, p < .001), and Risk (0.23% vs 0.16%, p < .001) sentiments. Misinformation topics scored higher on Conflict (.26% vs .24%, p = .013), Health (4.80% vs 4.62%, p < .001), and Death (0.33% vs 0.12%, p < .001) related words compared to corrections. When compared against the averages of all other misinformation classifications, the mixed topic scored highest in Moral (0.48%) and Risk (0.52%) sentiments.

Lastly, among time-related sentiment categories misinformation topics on average scored higher than all other classifications for general Time (2.66%) and Past-focused (3.11%), but scored lowest for Present-focused (3.59%) and lower than corrections for Future-focused (0.53% vs 0.58%, p < .001) sentiments. Topics low in both misinformation and corrections scored the highest (7.16%) for Present-focused words compared against all other classifications (p < .001).

3.3 Analysis of user account affiliations

Accounts that composed the top ten most retweeted tweets associated with each topic cluster were grouped together by misinformation classification type, as seen in Figs 7 and 8. Accounts were classified as a spreader of misinformation or corrections based on whether they posted tweets from the respective categories as labeled from deductive coding. See S1 Table for account information broken out by each topic cluster. Accounts that were engaged in misinformation discourse were the most likely to be suspended (36.6%) or deleted (18.3%) after two years (the time of analysis) from posting their tweets compared to correction discourse where only 2.8% of accounts were suspended. Since a suspended account refers to when the platform kicks off a user’s account, these results reflect actions taken by Twitter towards addressing misinformation discourse.

Fig 7. Most retweeted accounts: Suspended of deleted by misinformation classification.

Fig 7

Fig 8. Most retweeted accounts: Affiliations by misinformation classification.

Fig 8

Of the occupational affiliations for most retweeted users, there were 17 medical affiliates / scientists, 27 employed by the media, 7 government officials, and 2 religious leaders. Among accounts engaged in misinformation discourse, there were some occupational affiliations detected with the media (5.6%), medical affiliates / scientists (4.2%), and religious leaders (1.4%). There was no involvement from government officials detected in the misinformation discourse. However, there were relatively more professional affiliations identified in correction discourse where 33.3% of accounts were affiliated with the media, 22.2% were medical affiliates, and 8.3% were government officials. No religious leaders were associated with correction discourse. These findings show that while there were some media and medical affiliates spreading misinformation, the vast majority focused on correcting false information.

4. Discussion

This study collected tweets related to 5G conspiracy theories and applied both NLP and qualitative content coding to characterize a large-scale online discourse. Sentiment analysis results show that misinformation and conspiracy discourse was more likely to use language that is analytic and references social status, death, conflict, and health. This discourse also scored higher in general emotion, negative emotion, and past-orientation. In contrast, discourse that challenged false information was more likely to use language related to cognitive processes, positive emotions, anger, prosocial tendencies, morality, illness, risk, and be future-oriented. These results reveal that correction topics, when compared to misinformation and conspiracy discourse, are more likely to use explanatory words related to causative arguments (e.g., “because”, “how”) and are more concerned with social relations and health-related consequences. The differences in temporal focus between discourses suggest that misinformation narratives evoke past events more often than corrections, whereas the future-oriented language of correction discourse could reflect concerns regarding health-related consequences to 5G-COVID conspiracy theories. These results also reveal that conspiracy-related discourse is more likely to use combative and negative emotional language, and mention extreme consequences more often than corrections (e.g., using death-related words vs illness-related words). The higher degree of negative and emotional language is consistent with previous studies that used sentiment analysis to examine 5G-COVID conspiracies [43, 80], and provides additional insight into rhetorical strategies used in conspiratorial discourse. Further, these findings are consistent with work examining conspiracy spreaders on Twitter more generally, which found significant differences in negative emotion and death-related words compared to those who engage with scientific content [42].

Results from the inductive coding themes were also detected in related work. More specifically, the themes Group of Elites [29, 30, 81], Anti-vax [30, 81], Celebrity [29], Religious Figures [29], Geography Comparisons [27], Videos [82], Technology correlations with disease outbreaks [28], and Negative health effects [29] were all identified in other studies across Facebook, Twitter, and Instagram. The consistency in findings shows the utility of the approach used in the current study, and reveals a pattern of uniformity in conspiracist narratives across social media platforms.

4.1 Using topic modeling to identify discourse for further investigation

This hybrid approach uses topic modeling to identify influential posts within the corpus across multiple topic areas, making it possible to only need to code a small subset of the total volume of tweets in order to detect prominent themes and narratives within both conspiracy and correction discourses. Additionally, segmenting the discourse into topic clusters and then running sentiment analysis makes it possible to detect smaller subsets of tweets within the larger twitter conversation that are high in emotional affect. This can be useful for quickly identifying emotionally charged conversations that could be prioritized for content assessment from public officials or platforms. For instance, since discourse within misinformation topics 0, 6, and 13 are highest in negative emotion, anger and conflict sentiments, as seen in Table 4, these might be of relevance to those interested in tracking or preventing mobilization responses that lead to consequences in the physical world or lead to “offline harm” (such as deciding to burn down a cell tower). When assessing the inductive coding themes related to these topics, results show that themes concerning viral videos, negative health effects, and the Chinese technology company Huawei are associated with these sentiments. This could lead to follow-up analyses examining the users who lead these conversations, and can help prioritize which narratives within a conspiracy discourse might be most harmful to the public, especially in cases such as Huawei that could be linked to xenophobic attitudes towards Chinese and Asian Americans.

Another topic of interest is cluster 8, which is associated with the anti-vax theme and high in health and illness sentiments. Despite the timeframe of the study being late March to early April, where for those in the US and most English-speaking countries COVID-19 was still an emerging pandemic, it is surprising to encounter anti-vax sentiment towards the COVID-19 vaccine almost nine months before the announcement of the first approved vaccine. Topic 16, which has inductive coding themes related to causative explanations and medical authorities, and high in sentiment scores for Anxiousness and Pro-social language, could also be of interest for those looking to identify persuasive strategies used by online conspiracy spreaders. In this discourse, the combination of credibility signifiers reflected in scientific terminology and pro-social language could make messages more persuasive to those seeking certainty and safety during an unknown health crisis or searching for more credible sources of information to fill an existing information gap.

It is also possible to track reactions to misinformation discourse by examining themes prominent in correction topics, where strategies using causative explanations and calls to medical authorities appear to be a common strategy used in corrections. However, the discourse from the corrections did not directly address many of the narratives prevalent in misinformation conversations (e.g., Huawei, Anti-vax) with the most specific talking-points focusing on criticizing celebrities or religious leaders for spreading conspiracies as also detected by Honcharov et al. in a separate study examining anti-vaccination hashtags of public figures on Twitter [82]. Not addressing specific conspiracist narratives could tamper the effectiveness of correction strategies by not clarifying the information gaps that are capitalized on by misinformation spreaders, especially during times of uncertainty. This is particularly important in light of the findings from the affiliations analysis of the most retweeted users, which showed that media figures, medical scientists, and government officials were more likely to be involved in corrections discourse. Since these accounts are more likely to have higher perceived credibility and a greater reach in audience compared to the average user, it is important that emerging conspiracist narratives are identified early in order to design more effective counter-messaging strategies.

4.2 Drawbacks of solely relying on machine learning approaches

Machine learning approaches for detecting misinformation (or classifying categories within online discourse more generally) have continued to evolve over the years. Recent efforts to improve accuracy of misinformation classifiers have found success using theoretical frameworks based on human information processing to guide feature selection [83], combining features from multiple modalities (i.e., text and visual) [84], and including features related to user response to post and message sources [85]. Multiple algorithms have also been evaluated, ranging from traditional logistic regression models and ensemble approaches such as voting and bagging classifiers [86] to advanced models using deep convolutional neural networks and knowledge graphs that use word embeddings instead of requiring feature selection [87, 88].

In response to the high accuracy of machine learning classifiers at detecting misinformation, there have been calls for action in recent years for researchers to make publicly available textual data with labeled fake news articles to build comprehensive training datasets for misinformation classification [8991]. These findings, in addition to the prominence of advanced LLMs such as chatGPT [92, 93], make it tempting to infer that AI technologies can fully address the issue of large-scale content coding without the use of humans. However, as previously discussed, an important drawback of machine learning approaches is that they are generally limited to detecting fake news within specific contexts. Since misinformation topics and conspiracy theories are always evolving, NLP approaches are unable to generalize to novel situations that are not already documented. For example, classifiers developed in response to the misinformation proliferation surrounding the 2016 US presidential election would not be able to adequately detect COVID-related misinformation since the relevant keywords do not correspond across topics. Maintaining up-to-date classifier models would also be difficult to achieve during the COVID-19 pandemic, where scientific consensus and understanding of the virus have shifted multiple times from its early stages, and continues to evolve even over three years after its initial outbreak. It is also difficult for machine learning approaches to detect instances where outdated information or findings that were considered accurate at the time of publication are then used misleadingly in contemporary scenarios, such as in the case surrounding misinformation promoting the use of hydroxychloroquine for treating COVID-19 [56, 58].

While building effective classifier models for 5G COVID conspiracies is possible during 2023 after being well documented, only human coders had the capability to accurately recognize text that contains these narratives when they first emerged online in 2020. We intended to address these limitations by demonstrating the general utility of a hybrid approach that incorporates human coders to take advantage of the efficiencies machine learning techniques provide researchers when working with big datasets while still accounting for contextual nuance from using qualitative approaches. Even though the definition of misinformation may vary in other discussions (e.g., politics, climate change, other social issues), the general principles of the methodology described could be leveraged to provide more up-to-date and richer contextual insights into how these conspiracy-related discourses evolve over time.

4.3 Future directions in infodemiology

In order to properly make sense of findings generated from large scale communication networks, it is important that we also advance our conceptual understandings of information transmission dynamics. Other fields such as cognitive science and human-computer interactions (HCI) have developed frameworks for scientifically examining information as a phenomenon, which can be incorporated into future infodemiology research. One such theory is information foraging, which is based on ecological models of food scavenging behaviors where online users are considered “foragers” who balance the value gained from finding new information with the time cost needed to obtain it [9498]. The types of analyses conducted using an information foraging framework include the following: information patch models that assess engagement activities in environments where information is encountered in clusters (e.g., webpages), information scent models that assess how information value is evaluated from proximal cues (e.g., titles and images on a website), and information diet models that examine decisions concerning the selection and pursuit of information items [94]. Future infodemiological work can design behavioral experiments using information scent models to test how cues on sites or social media posts influence safety perceptions of potentially dangerous transactions such as illicit drug purchasing, and adapt information diet models to examine how users evaluate the truthfulness of health-claims across online environments.

Another relevant theoretical framework is distributed cognition, which extends human intelligence beyond the boundaries of individual actors to encompass interactions between people and resources within the environment [99101]. According to this framework, social organization determines the way information flows through groups [100], and more recent work emphasizes the ways cognition is distributed over a vast array of social, institutional, political, and technological systems that are shaped by, and shape, the individuals who develop and operate within them [99, 102]. Within the context of social media environments, where topics of discourse are constantly shifting and information is transmitted from a wide array of sources including other users, celebrities, institutions, and political figures, this framework can make sense of how narratives evolve by examining the interplay between the broadcaster of a message and their audiences’ reaction, engagement, and retransmission of that message. The theory of distributed cognition can also guide research questions that distinguish between message transmission due to an individual’s personal beliefs and propagation driven by conformity to other users. Overall, both information foraging and distributed cognition frameworks can be adapted in future work to generate deeper insights concerning health-related information seeking behaviors, enrich interpretations from social network analysis, and guide the design of interventions targeting misinformation propagation.

4.4 Limitations

Content coding took place several months after the initial timeframe of the study. While having the time gap allowed us to assess which accounts were deleted or suspended since the initial 5G discourse, we were unable to code for affiliations of those deleted users using publicly available profile data. As stated in the methods section, the top 10 retweeted tweets do not account for every tweet associated within a topic cluster. While the most retweeted tweets account for a substantive proportion of the topic cluster’s tweet volume, and in most cases a majority, there is still some level of uncertainty when characterizing the discourse even if the uncoded tweets share textual similarity to the coded posts. Additional measures such as the use of sentiment analysis, which in this study was applied to the full topic clusters, can also mitigate these concerns since they account for information provided in the uncoded tweets. Finally, though false information can be categorized as “misinformation,” “disinformation,” “mal-information,” and “conspiracy” based on intent and content, this study did not differentiate between these categories, opting to call all false information, regardless of intent, “misinformation.” This lack of differentiation limits the study’s ability to identify potential differences in rhetoric associated with the nuances of false information dissemination.

5. Conclusion

The advancement of communication technologies and the continued emergence of new social media platforms present difficult challenges for researchers looking to investigate the highly dynamic information ecosystem of the 21st century. Fortunately, these same rapid advancements in technology can also be harnessed by researchers as powerful tools to navigate these complex environments. However, as demonstrated in the current paper, the human perspective is equally crucial in this line of work to compensate for the shortcomings that artificial intelligences have toward understanding human endeavors. Due to the rapid pace of modern discourse, words that can be key identifiers for a dangerous conspiracy in one context can be completely irrelevant in a different grouping of text. Within these online conversations, where the boundary between signal and noise is constantly shifting due to emerging and continually evolving narratives, it is crucial to recruit the signal detection capabilities of both machine learning models and human beings to adequately address current and future misinformation challenges now endemic in our global information society.

Supporting information

S1 Checklist. STROBE statement—Checklist of items that should be included in reports of observational studies.

(DOCX)

S1 Appendix. Hybrid approach for characterizing social media discourse using machine learning and qualitative methodologies.

(PDF)

S1 Table. Most retweeted accounts associated with each BTM cluster.

(PDF)

Acknowledgments

The authors would like to thank our collaborators in the Cognitive Media Lab at UC San Diego (cognitivemedialab.ucsd.edu) for the productive discussions about online conspiracy discourse that helped inform the current manuscript.

Data Availability

The full dataset collected from Twitter and the top retweeted tweets selected for content coding can be found at the following link on the Open Science Framework (OSF): https://bit.ly/5G_Conspiracies (DOI 10.17605/OSF.IO/YRNMX). Usernames, links to the tweet, and all personally identifiable information were removed to preserve anonymity.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Allcott H, Gentzkow M. Social Media and Fake News in the 2016 Election. J Econ Perspect. 2017;31: 211–36. doi: 10.1257/jep.31.2.211 [DOI] [Google Scholar]
  • 2.Bovet A, Makse HA. Influence of fake news in Twitter during the 2016 US presidential election. Nat Commun. 2019;10: 7. doi: 10.1038/s41467-018-07761-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Grinberg N, Joseph K, Friedland L, Swire-Thompson B, Lazer D. Fake news on Twitter during the 2016 U.S. presidential election. Science. 2019;363: 374–378. doi: 10.1126/science.aau2706 [DOI] [PubMed] [Google Scholar]
  • 4.Zarocostas J. How to fight an infodemic. The lancet. 2020;395: 676. doi: 10.1016/S0140-6736(20)30461-X [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Eysenbach G. Infodemiology: The epidemiology of (mis) information. Am J Med. 2002;113: 763–765. doi: 10.1016/s0002-9343(02)01473-0 [DOI] [PubMed] [Google Scholar]
  • 6.Eysenbach G. Infodemiology: tracking flu-related searches on the web for syndromic surveillance. American Medical Informatics Association; 2006. p. 244. [PMC free article] [PubMed] [Google Scholar]
  • 7.Eysenbach G. Infodemiology and Infoveillance: Framework for an Emerging Set of Public Health Informatics Methods to Analyze Search, Communication and Publication Behavior on the Internet. J Med Internet Res. 2009;11: e11. doi: 10.2196/jmir.1157 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Cuomo RE, Purushothaman V, Li J, Cai M, Mackey TK. A longitudinal and geospatial analysis of COVID-19 tweets during the early outbreak period in the United States. BMC Public Health. 2021;21: 793. doi: 10.1186/s12889-021-10827-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Li J, Xu Q, Cuomo R, Purushothaman V, Mackey T. Data Mining and Content Analysis of the Chinese Social Media Platform Weibo During the Early COVID-19 Outbreak: Retrospective Observational Infoveillance Study. JMIR Public Health Surveill. 2020;6: e18700. doi: 10.2196/18700 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Mackey TK, Li J, Purushothaman V, Nali M, Shah N, Bardier C, et al. Big Data, Natural Language Processing, and Deep Learning to Detect and Characterize Illicit COVID-19 Product Sales: Infoveillance Study on Twitter and Instagram. JMIR Public Health Surveill. 2020;6: e20794. doi: 10.2196/20794 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Chiou H, Voegeli C, Wilhelm E, Kolis J, Brookmeyer K, Prybylski D. The Future of Infodemic Surveillance as Public Health Surveillance. Emerg Infect Dis J. 2022;28: 121. doi: 10.3201/eid2813.220696 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Mackey T, Baur C, Eysenbach G. Advancing Infodemiology in a Digital Intensive Era. JMIR Infodemiology. 2022;2: e37115. doi: 10.2196/37115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Haupt MR, Xu Q, Yang J, Cai M, Mackey TK. Characterizing Vaping Industry Political Influence and Mobilization on Facebook: Social Network Analysis. J Med Internet Res. 2021;23: e28069. doi: 10.2196/28069 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Xu Q, Yang J, Haupt MR, Cai M, Nali MC, Mackey TK. Digital Surveillance to Identify California Alternative and Emerging Tobacco Industry Policy Influence and Mobilization on Facebook. Int J Environ Res Public Health. 2021;18. doi: 10.3390/ijerph182111150 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Freckelton QC I. COVID-19: Fear, quackery, false representations and the law. Int J Law Psychiatry. 2020;72: 101611. doi: 10.1016/j.ijlp.2020.101611 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.van Prooijen J-W, Douglas KM. Conspiracy theories as part of history: The role of societal crisis situations. Mem Stud. 2017;10: 323–333. doi: 10.1177/1750698017701615 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Ajao O., Bhowmik D., Zargari S. Sentiment Aware Fake News Detection on Online Social Networks. ICASSP 2019–2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 2019. pp. 2507–2511. doi: 10.1109/ICASSP.2019.8683170 [DOI] [Google Scholar]
  • 18.Bhutani B., Rastogi N., Sehgal P., Purwar A. Fake News Detection Using Sentiment Analysis. 2019 Twelfth International Conference on Contemporary Computing (IC3). 2019. pp. 1–5. doi: 10.1109/IC3.2019.8844880 [DOI] [Google Scholar]
  • 19.Kolluri N, Liu Y, Murthy D. COVID-19 Misinformation Detection: Machine-Learned Solutions to the Infodemic. JMIR Infodemiology. 2022;2: e38756. doi: 10.2196/38756 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Lee CJ, Chua HN. Using Linguistics and Psycholinguistics Features in Machine Learning for Fake News Classification Through Twitter. Springer; 2022. pp. 717–730. [Google Scholar]
  • 21.Caramancion K. M. Harnessing the Power of ChatGPT to Decimate Mis/Disinformation: Using ChatGPT for Fake News Detection. 2023 IEEE World AI IoT Congress (AIIoT). 2023. pp. 0042–0046. doi: 10.1109/AIIoT58121.2023.10174450 [DOI] [Google Scholar]
  • 22.Johnson SB, King AJ, Warner EL, Aneja S, Kann BH, Bylund CL. Using ChatGPT to evaluate cancer myths and misconceptions: artificial intelligence and cancer information. JNCI Cancer Spectr. 2023;7: pkad015. doi: 10.1093/jncics/pkad015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Rashkin H, Choi E, Jang JY, Volkova S, Choi Y. Truth of varying shades: Analyzing language in fake news and political fact-checking. 2017. pp. 2931–2937. [Google Scholar]
  • 24.Massey PM, Kearney MD, Hauer MK, Selvan P, Koku E, Leader AE. Dimensions of Misinformation About the HPV Vaccine on Instagram: Content and Network Analysis of Social Media Characteristics. J Med Internet Res. 2020;22: e21451. doi: 10.2196/21451 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Zhou C, Li K, Lu Y. Linguistic characteristics and the dissemination of misinformation in social media: The moderating effect of information richness. Inf Process Manag. 2021;58: 102679. doi: 10.1016/j.ipm.2021.102679 [DOI] [Google Scholar]
  • 26.Ahmed W, Vidal-Alaball J, Downing J, López Seguí F. COVID-19 and the 5G Conspiracy Theory: Social Network Analysis of Twitter Data. J Med Internet Res. 2020;22: e19458. doi: 10.2196/19458 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Flaherty E, Sturm T, Farries E. The conspiracy of Covid-19 and 5G: Spatial analysis fallacies in the age of data democratization. Soc Sci Med. 2022;293: 114546. doi: 10.1016/j.socscimed.2021.114546 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Langguth J, Filkuková P, Brenner S, Schroeder DT, Pogorelov K. COVID-19 and 5G conspiracy theories: long term observation of a digital wildfire. Int J Data Sci Anal. 2022;15: 329–346. doi: 10.1007/s41060-022-00322-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Bruns A, Harrington S, Hurcombe E. ‘Corona? 5G? or both?’: the dynamics of COVID-19/5G conspiracy theories on Facebook. Media Int Aust. 2020;177: 12–29. doi: 10.1177/1329878X20946113 [DOI] [Google Scholar]
  • 30.Quinn EK, Fazel SS, Peters CE. The Instagram infodemic: cobranding of conspiracy theories, coronavirus disease 2019 and authority-questioning beliefs. Cyberpsychology Behav Soc Netw. 2021;24: 573–577. doi: 10.1089/cyber.2020.0663 [DOI] [PubMed] [Google Scholar]
  • 31.Sicilia R., Giudice S. L., Pei Y., Pechenizkiy M., Soda P. Health-related rumour detection on Twitter. 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). 2017. pp. 1599–1606. doi: 10.1109/BIBM.2017.8217899 [DOI] [Google Scholar]
  • 32.Sicilia R, Lo Giudice S, Pei Y, Pechenizkiy M, Soda P. Twitter rumour detection in the health domain. Expert Syst Appl. 2018;110: 33–40. doi: 10.1016/j.eswa.2018.05.019 [DOI] [Google Scholar]
  • 33.Hou R, Pérez-Rosas V, Loeb S, Mihalcea R. Towards automatic detection of misinformation in online medical videos. 2019. pp. 235–243. [Google Scholar]
  • 34.Kinsora A., Barron K., Mei Q., Vydiswaran V. G. V. Creating a Labeled Dataset for Medical Misinformation in Health Forums. 2017 IEEE International Conference on Healthcare Informatics (ICHI). 2017. pp. 456–461. doi: 10.1109/ICHI.2017.93 [DOI] [Google Scholar]
  • 35.Liu Y., Yu K., Wu X., Qing L., Peng Y. Analysis and Detection of Health-Related Misinformation on Chinese Social Media. IEEE Access. 2019;7: 154480–154489. doi: 10.1109/ACCESS.2019.2946624 [DOI] [Google Scholar]
  • 36.Ghenai A, Mejova Y. Fake cures: user-centric modeling of health misinformation in social media. Proc ACM Hum-Comput Interact. 2018;2: 1–20. [Google Scholar]
  • 37.Ajao O, Bhowmik D, Zargari S. Fake news identification on twitter with hybrid cnn and rnn models. 2018. pp. 226–230. [Google Scholar]
  • 38.Granik M., Mesyura V. Fake news detection using naive Bayes classifier. 2017 IEEE First Ukraine Conference on Electrical and Computer Engineering (UKRCON). 2017. pp. 900–903. doi: 10.1109/UKRCON.2017.8100379 [DOI] [Google Scholar]
  • 39.Tausczik YR, Pennebaker JW. The Psychological Meaning of Words: LIWC and Computerized Text Analysis Methods. J Lang Soc Psychol. 2010;29: 24–54. doi: 10.1177/0261927X09351676 [DOI] [Google Scholar]
  • 40.Boyd RL, Ashokkumar A, Seraj S, Pennebaker JW. The development and psychometric properties of LIWC-22. Austin TX Univ Tex Austin. 2022; 1–47. [Google Scholar]
  • 41.Fast E, Chen B, Bernstein MS. Empath: Understanding topic signals in large-scale text. 2016. pp. 4647–4657. [Google Scholar]
  • 42.Fong A, Roozenbeek J, Goldwert D, Rathje S, van der Linden S. The language of conspiracy: A psychological analysis of speech used by conspiracy theorists and their followers on Twitter. Group Process Intergroup Relat. 2021;24: 606–623. doi: 10.1177/1368430220987596 [DOI] [Google Scholar]
  • 43.Rains SA, Leroy G, Warner EL, Harber P. Psycholinguistic Markers of COVID-19 Conspiracy Tweets and Predictors of Tweet Dissemination. Health Commun. 2021;38: 21–30. doi: 10.1080/10410236.2021.1929691 [DOI] [PubMed] [Google Scholar]
  • 44.Castelo S, Almeida T, Elghafari A, Santos A, Pham K, Nakamura E, et al. A topic-agnostic approach for identifying fake news pages. 2019. pp. 975–980. [Google Scholar]
  • 45.Che X, Metaxa-Kakavouli D, Hancock JT. Fake News in the News: An Analysis of Partisan Coverage of the Fake News Phenomenon. 2018. pp. 289–292. [Google Scholar]
  • 46.Giachanou A, Ghanem B, Ríssola EA, Rosso P, Crestani F, Oberski D. The impact of psycholinguistic patterns in discriminating between fake news spreaders and fact checkers. Data Knowl Eng. 2022;138: 101960. doi: 10.1016/j.datak.2021.101960 [DOI] [Google Scholar]
  • 47.Charquero-Ballester M, Walter JG, Nissen IA, Bechmann A. Different types of COVID-19 misinformation have different emotional valence on Twitter. Big Data Soc. 2021;8: 20539517211041280. doi: 10.1177/20539517211041279 [DOI] [Google Scholar]
  • 48.Beau Lotto R, Purves D. The empirical basis of color perception. Conscious Cogn. 2002;11: 609–629. doi: 10.1016/s1053-8100(02)00014-4 [DOI] [PubMed] [Google Scholar]
  • 49.Yin W, Zubiaga A. Hidden behind the obvious: Misleading keywords and implicitly abusive language on social media. Online Soc Netw Media. 2022;30: 100210. doi: 10.1016/j.osnem.2022.100210 [DOI] [Google Scholar]
  • 50.Poirier L. Reading datasets: Strategies for interpreting the politics of data signification. Big Data Soc. 2021;8: 20539517211029320. doi: 10.1177/20539517211029322 [DOI] [Google Scholar]
  • 51.Glaser B, Strauss A. Discovery of grounded theory: Strategies for qualitative research. Routledge; 2017. [Google Scholar]
  • 52.Corbin JM, Strauss A. Grounded theory research: Procedures, canons, and evaluative criteria. Qual Sociol. 1990;13: 3–21. doi: 10.1007/BF00988593 [DOI] [Google Scholar]
  • 53.Boyatzis RE. Transforming qualitative information: Thematic analysis and code development. sage; 1998. [Google Scholar]
  • 54.Suddaby R. From the Editors: What Grounded Theory is Not. Acad Manage J. 2006;49: 633–642. doi: 10.5465/amj.2006.22083020 [DOI] [Google Scholar]
  • 55.Fereday J, Muir-Cochrane E. Demonstrating Rigor Using Thematic Analysis: A Hybrid Approach of Inductive and Deductive Coding and Theme Development. Int J Qual Methods. 2006;5: 80–92. doi: 10.1177/160940690600500107 [DOI] [Google Scholar]
  • 56.Haupt MR, Li J, Mackey TK. Identifying and characterizing scientific authority-related misinformation discourse about hydroxychloroquine on twitter using unsupervised machine learning. Big Data Soc. 2021;8: 20539517211013844. [Google Scholar]
  • 57.Islam MS, Sarkar T, Khan SH, Mostofa Kamal A-H, Hasan SMM, Kabir A, et al. COVID-19–Related Infodemic and Its Impact on Public Health: A Global Social Media Analysis. Am J Trop Med Hyg. 2020;103: 1621–1629. doi: 10.4269/ajtmh.20-0812 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Mackey TK, Purushothaman V, Haupt M, Nali MC, Li J. Application of unsupervised machine learning to identify and characterise hydroxychloroquine misinformation on Twitter. Lancet Digit Health. 2021;3: e72–e75. doi: 10.1016/S2589-7500(20)30318-6 [DOI] [PubMed] [Google Scholar]
  • 59.Haupt MR, Cuomo R, Li J, Nali M, Mackey TK. The influence of social media affordances on drug dealer posting behavior across multiple social networking sites (SNS). Comput Hum Behav Rep. 2022;8: 100235. doi: 10.1016/j.chbr.2022.100235 [DOI] [Google Scholar]
  • 60.Mackey TK, Kalyanam J, Katsuki T, Lanckriet G. Twitter-Based Detection of Illegal Online Sale of Prescription Opioid. Am J Public Health. 2017;107: 1910–1915. doi: 10.2105/AJPH.2017.303994 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Mackey T, Kalyanam J, Klugman J, Kuzmenko E, Gupta R. Solution to Detect, Classify, and Report Illicit Online Marketing and Sales of Controlled Substances via Twitter: Using Machine Learning and Web Forensics to Combat Digital Opioid Access. J Med Internet Res. 2018;20: e10029. doi: 10.2196/10029 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Shah Neal, Nali Matthew, Bardier Cortni, Li Jiawei, Maroulis James, Cuomo Raphael, et al. Applying topic modelling and qualitative content analysis to identify and characterise ENDS product promotion and sales on Instagram. Tob Control. 2021;32: e153. doi: 10.1136/tobaccocontrol-2021-056937 [DOI] [PubMed] [Google Scholar]
  • 63.Mackey T, Purushothaman V, Li J, Shah N, Nali M, Bardier C, et al. Machine Learning to Detect Self-Reporting of Symptoms, Testing Access, and Recovery Associated With COVID-19 on Twitter: Retrospective Big Data Infoveillance Study. JMIR Public Health Surveill. 2020;6: e19509. doi: 10.2196/19509 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Haupt MR, Jinich-Diamant A, Li J, Nali M, Mackey TK. Characterizing twitter user topics and communication network dynamics of the “Liberate” movement during COVID-19 using unsupervised machine learning and social network analysis. Online Soc Netw Media. 2021;21: 100114. doi: 10.1016/j.osnem.2020.100114 [DOI] [Google Scholar]
  • 65.Benevenuto F, Rodrigues T, Cha M, Almeida V. Characterizing user behavior in online social networks. 2009. pp. 49–62. [Google Scholar]
  • 66.Lerman K, Ghosh R. Information Contagion: An Empirical Study of the Spread of News on Digg and Twitter Social Networks. Proceedings of the International AAAI Conference on Web and Social Media. 2010. May 16;4(1):90–7. doi: 10.1609/icwsm.v4i1.14021 [DOI] [Google Scholar]
  • 67.Papakyriakopoulos O, Serrano JCM, Hegelich S. Political communication on social media: A tale of hyperactive users and bias in recommender systems. Online Soc Netw Media. 2020;15: 100058. doi: 10.1016/j.osnem.2019.100058 [DOI] [Google Scholar]
  • 68.Van Mieghem P, Blenn N, Doerr C. Lognormal distribution in the digg online social network. Eur Phys J B. 2011;83: 251. doi: 10.1140/epjb/e2011-20124-0 [DOI] [Google Scholar]
  • 69.Gruzd A, Mai P. Going viral: How a single tweet spawned a COVID-19 conspiracy theory on Twitter. Big Data Soc. 2020;7: 2053951720938405. doi: 10.1177/2053951720938405 [DOI] [Google Scholar]
  • 70.Hamilton IA. 77 cell phone towers have been set on fire so far due to a weird coronavirus 5G conspiracy theory. In: Business Insider [Internet]. [cited 30 Mar 2023]. Available: https://www.businessinsider.com/77-phone-masts-fire-coronavirus-5g-conspiracy-theory-2020-5 [Google Scholar]
  • 71.Jolley D, Paterson JL. Pylons ablaze: Examining the role of 5G COVID-19 conspiracy beliefs and support for violence. Br J Soc Psychol. 2020;59: 628–640. doi: 10.1111/bjso.12394 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Arora VM, Madison S, Simpson L. Addressing Medical Misinformation in the Patient-Clinician Relationship. JAMA. 2020;324: 2367–2368. doi: 10.1001/jama.2020.4263 [DOI] [PubMed] [Google Scholar]
  • 73.Bruns A, Harrington S, Hurcombe E. Coronavirus Conspiracy Theories: Tracing Misinformation Trajectories from the Fringes to the Mainstream. In: Lewis M, Govender E, Holland K, editors. Communicating COVID-19: Interdisciplinary Perspectives. Cham: Springer International Publishing; 2021. pp. 229–249. doi: 10.1007/978-3-030-79735-5_12 [DOI] [Google Scholar]
  • 74.Calac AJ, Haupt MR, Li Z, Mackey T. Spread of COVID-19 Vaccine Misinformation in the Ninth Inning: Retrospective Observational Infodemic Study. JMIR Infodemiology. 2022;2: e33587. doi: 10.2196/33587 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75.Shin I, Wang L, Lu Y-T. Twitter and endorsed (fake) news: The influence of endorsement by strong ties, celebrities, and a user majority on credibility of fake news during the COVID-19 pandemic. Int J Commun. 2022;16: 23. [Google Scholar]
  • 76.Blekanov I, Tarasov N, Maksimov A. Topic modeling of conflict ad hoc discussions in social networks. 2018. pp. 122–126. [Google Scholar]
  • 77.Schück S, Foulquié P, Mebarki A, Faviez C, Khadhar M, Texier N, et al. Concerns Discussed on Chinese and French Social Media During the COVID-19 Lockdown: Comparative Infodemiology Study Based on Topic Modeling. JMIR Form Res. 2021;5: e23593. doi: 10.2196/23593 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Kalyanam J, Katsuki T, R.G. Lanckriet G, Mackey TK. Exploring trends of nonmedical use of prescription drugs and polydrug abuse in the Twittersphere using unsupervised machine learning. Addict Behav. 2017;65: 289–295. doi: 10.1016/j.addbeh.2016.08.019 [DOI] [PubMed] [Google Scholar]
  • 79.Yan X, Guo J, Lan Y, Cheng X. A biterm topic model for short texts. A biterm topic model for short texts. 2013. pp. 1445–1456. doi: 10.1145/2488388.2488514 [DOI] [Google Scholar]
  • 80.Gerts D, Shelley CD, Parikh N, Pitts T, Watson Ross C, Fairchild G, et al. “Thought I’d Share First” and Other Conspiracy Theory Tweets from the COVID-19 Infodemic: Exploratory Study. JMIR Public Health Surveill. 2021;7: e26527. doi: 10.2196/26527 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81.Erokhin D, Yosipof A, Komendantova N. COVID-19 Conspiracy Theories Discussion on Twitter. Soc Media Soc. 2022;8: 20563051221126052. doi: 10.1177/20563051221126051 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82.Honcharov V, Li J, Sierra M, Rivadeneira NA, Olazo K, Nguyen TT, et al. Public Figure Vaccination Rhetoric and Vaccine Hesitancy: Retrospective Twitter Analysis. JMIR Infodemiology. 2023;3: e40575. doi: 10.2196/40575 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83.Zhao Y, Da J, Yan J. Detecting health misinformation in online health communities: Incorporating behavioral features into machine learning based approaches. Inf Process Manag. 2021;58: 102390. doi: 10.1016/j.ipm.2020.102390 [DOI] [Google Scholar]
  • 84.Singh VK, Ghosh I, Sonagara D. Detecting fake news stories via multimodal analysis. J Assoc Inf Sci Technol. 2021;72: 3–17. doi: 10.1002/asi.24359 [DOI] [Google Scholar]
  • 85.Ruchansky N, Seo S, Liu Y. Csi: A hybrid deep model for fake news detection. 2017. pp. 797–806. [Google Scholar]
  • 86.Ahmad I, Yousaf M, Yousaf S, Ahmad MO. Fake News Detection Using Machine Learning Ensemble Methods. Uddin MI, editor. Complexity. 2020;2020: 8885861. doi: 10.1155/2020/8885861 [DOI] [Google Scholar]
  • 87.Kaliyar RK, Goswami A, Narang P, Sinha S. FNDNet–A deep convolutional neural network for fake news detection. Cogn Syst Res. 2020;61: 32–44. doi: 10.1016/j.cogsys.2019.12.005 [DOI] [Google Scholar]
  • 88.Pan JZ, Pavlova S, Li C, Li N, Li Y, Liu J. Content based fake news detection using knowledge graphs. Springer; 2018. pp. 669–683. [Google Scholar]
  • 89.Nielsen DS, McConville R. Mumin: A large-scale multilingual multimodal fact-checked misinformation social network dataset. 2022. pp. 3141–3153. [Google Scholar]
  • 90.Torabi Asr F, Taboada M. Big Data and quality data for fake news and misinformation detection. Big Data Soc. 2019;6: 2053951719843310. doi: 10.1177/2053951719843310 [DOI] [Google Scholar]
  • 91.Zhang X, Ghorbani AA. An overview of online fake news: Characterization, detection, and discussion. Inf Process Manag. 2020;57: 102025. doi: 10.1016/j.ipm.2019.03.004 [DOI] [Google Scholar]
  • 92.De Angelis L, Baglivo F, Arzilli G, Privitera GP, Ferragina P, Tozzi AE, et al. ChatGPT and the rise of large language models: the new AI-driven infodemic threat in public health. Front Public Health. 2023;11. Available: https://www.frontiersin.org/articles/10.3389/fpubh.2023.1166120 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 93.King MR, chatGPT. A Conversation on Artificial Intelligence, Chatbots, and Plagiarism in Higher Education. Cell Mol Bioeng. 2023;16: 1–2. doi: 10.1007/s12195-022-00754-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 94.Pirolli P, Card S. Information foraging. Psychol Rev. 1999;106: 643. [Google Scholar]
  • 95.Pirolli P, Card S. Information foraging in information access environments. 1995. pp. 51–58. [Google Scholar]
  • 96.Pirolli P. Rational Analyses of Information Foraging on the Web. Cogn Sci. 2005;29: 343–373. doi: 10.1207/s15516709cog0000_20 [DOI] [PubMed] [Google Scholar]
  • 97.Pirolli P. An elementary social information foraging model. 2009. pp. 605–614. [Google Scholar]
  • 98.Pirolli P, Fu W-T. SNIF-ACT: A Model of Information Foraging on the World Wide Web. In: Brusilovsky P, Corbett A, de Rosis F, editors. User Modeling 2003. Berlin, Heidelberg: Springer Berlin Heidelberg; 2003. pp. 45–54. [Google Scholar]
  • 99.Cash M. Cognition without borders: “Third wave” socially distributed cognition and relational autonomy. Socially Ext Cogn. 2013;25–26: 61–71. doi: 10.1016/j.cogsys.2013.03.007 [DOI] [Google Scholar]
  • 100.Hollan J, Hutchins E, Kirsh D. Distributed cognition: toward a new foundation for human-computer interaction research. ACM Trans Comput-Hum Interact TOCHI. 2000;7: 174–196. [Google Scholar]
  • 101.Rogers Y. A brief introduction to Distributed Cognition. In 1997. [cited 2023 Mar 31]. Available from: https://www.semanticscholar.org/paper/A-brief-introduction-to-Distributed-Cognition-Rogers/5633038212a61652781835c02d681922c1620b48. [Google Scholar]
  • 102.Gallagher S. The socially extended mind. Socially Ext Cogn. 2013;25–26: 4–12. doi: 10.1016/j.cogsys.2013.03.008 [DOI] [Google Scholar]

Decision Letter 0

Jasna Karacic Zanetti

31 Jul 2023

PONE-D-23-14668Detecting Nuance in Conspiracy Discourse: Advancing Methods in Infodemiology and Communication Science with Machine Learning and Qualitative Content CodingPLOS ONE

Dear Dr. Haupt,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. The reviewers have offered valuable insights and constructive criticism aimed at strengthening your manuscript. Their expertise and perspectives will undoubtedly contribute to refining your research and ensuring its significance within the field. It is crucial that you carefully consider their comments and respond to each point thoughtfully.

Information regarding ethical challenges in digitalization could be found in:

https://www.mdpi.com/2073-431X/12/6/114.

Please review and evaluate the requested work to determine whether they are relevant and should be cited. It is not a requirement to cite these works. We appreciate your attention to this request.

Please submit your revised manuscript by Sep 14 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Regards,

Jasna Karacic Zanetti, Ph.D

Academic Editor 

Plos One

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. In your Methods section, please include additional information about your dataset and ensure that you have included a statement specifying whether the collection and analysis method complied with the terms and conditions for the source of the data.

3. Thank you for stating the following in the Competing Interests section:

“Authors TKM and ZL are employees of the startup company S-3 Research LLC. S-3 Research is a startup funded and currently supported by the National Institutes of Health – National Institute of Drug Abuse through a Small Business Innovation and Research contract for opioid-related social media research and technology commercialization. TKM is also the CEO and a member of S-3 Research LLC with ownership.  Author reports no other conflict of interest associated with this manuscript.”

Please confirm that this does not alter your adherence to all PLOS ONE policies on sharing data and materials, by including the following statement: "This does not alter our adherence to  PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests).  If there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please include your updated Competing Interests statement in your cover letter; we will change the online submission form on your behalf.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Partly

********** 

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Yes

********** 

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Yes

********** 

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Yes

********** 

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The manuscript ensures that the message conveyed is clear, concise, and easily comprehensible to the intended audience. Its well-written manuscript uses appropriate grammar, spelling, punctuation, and vocabulary to convey the author's ideas effectively. It also follows a logical structure that enables readers to follow the flow of ideas easily.

Reviewer #2: 1. Summary of research and overall impression

The manuscript is a well written study of misinformation recognized in online platforms such as Twitter. It takes a thorough view of various approaches to misinformation detection and classification and also examines the societal need for this activity, a field termed as Infodemiology in the health context.

2. Major Issues

The work has some excessively wide-ranging phrases and general statements, for example, ``Since science itself is an iterative process..." which result in a slightly long manuscript and could be pruned. Moreover, such statements are not strictly necessary, apparently.

Minor Issues

Details such as the specific script/algorithm used for analysis in each Table could be presented in Supplementary Data or in an Appendix to make the work more replicable by the readership.

As an additional point, as the manuscript is long, adding a `table of contents' type element right at the start might be useful for the reader to get a bird's-eye view of all the sections and subsections; but journal format guidelines have the last word on this.

3. Overall Feedback

The work is well referenced and the presentation conveys the impression of thoroughness.

Reviewer #3: This is an excellent study and manuscript describing the use of a hybrid method for detecting conspiracy tweets linking 5G technology to COVID-19. I especially appreciated the authors’ suggested applications of their findings in the Discussion. A few points are provided below for the authors to consider.

• The Introduction provides an excellent and erudite background on infodemiology and content coding. It is longer than is standard, but perhaps worth keeping given the broad audience of PLOS One.

• Given the large amount of information scientists encounter daily, many will only see the abstract of this study. It is therefore important to include some findings or results in the abstract. I urge the authors to add at least two sentences describing the outcome of their analysis of tweets linking 5G to COVID-19; several of the background sentences can be omitted, if needed, to accommodate the word limit. Even though the authors position the study as a showcase for their methods, it seems important to demonstrate the novel insights that the method provides. Similarly, the authors state that “this study showcases . . . .” in the abstract, without describing the study. These changes will help rectify the current feel of over-promoting the method instead of letting the data/outcomes speak for itself/themselves.

• COVID-19 or coronavirus should be listed as a keyword, as it does not appear in the abstract or title. Consider added 5G as a keyword also, although that does appear in the abstract.

• A description of the units shown should be added to each graph or in the figure captions. This could be accomplished by labeling the Y axis of each graph.

• This study—including both the methods and findings—are of broad societal interest and may attract a wide readership. For this reason, I suggest that the authors add a simplified (flow?) diagram of the general methods used, highlighting the novel features and/or outcomes it permits.

• I suggest deleting the last sentence of the Conclusion, as it seems a bit “preachy” for a scientific paper and may work better in an editorial.

Reviewer #4: This study investigated conspiracy discourse related to 5G wireless technology by using a combination of NLP (i.e., topic modeling and sentiment analysis) and qualitative content analysis.

1. The abstract should reflect the main findings of the current study.

2. The first paragraph concludes with a call to advance methodologies that address the ongoing proliferation of misinformation that is overwhelming the social media ecosystem. I got a little lost here. This conclusion seems disconnected from the main reasoning of this paragraph. Why can't the aforementioned infodemiology applications address the current problems? What areas in particular need to be advanced?

3. The authors should clearly distinguish between misinformation and conspiracies. They are two distinct concepts, but the authors used them interchangeably throughout the manuscript.

4. I have some concerns over the generalizability of the current study. The NLP of this study generally evolves around the 5G-COVID conspiracy. How could the findings be generalized to misinformation detection in other disciplines, say politics, climate change, etc.?

5. Please elaborate on the relativistic approach: define it and justify its use in the current context.

6. The first paragraph on page 9 appears disjointed as it initially discusses three modes of data interpretation, then transitions to discussing inductive and deductive analysis. It would be beneficial to establish a coherent connection between these two sets of approaches to ensure the flow of ideas and the logical progression of the argument.

7. I guess most of the audience of this paper should get some foundational knowledge, like what inductive and deductive coding are, as well as their pros and cons. So the authors could cut this part.

8. Deductive coding is a coding process intended to test whether data coincides with existing assumptions, theories, or hypotheses. I agree with that. But I don’t see any assumptions or hypotheses that the authors sought to test.

9. My understanding is that this study summarized the topic and linguistic characteristics of 5G-related conspiracies on Twitter using topic modeling, LIWC, and manual content analysis. First, it is not about detection. The paper did not present any algorithms or computations that could be used to detect falsehoods in the future. Second, these methods are not quite new (see below)

Rashkin, H., Choi, E., Jang, J. Y., Volkova, S., & Choi, Y. (2017, September). Truth of varying shades: Analyzing language in fake news and political fact-checking. In Proceedings of the 2017 conference on empirical methods in natural language processing (pp. 2931-2937).

Zhou, C., Li, K., & Lu, Y. (2021). Linguistic characteristics and the dissemination of misinformation in social media: The moderating effect of information richness. Information Processing & Management, 58(6), 102679.

Massey, P. M., Kearney, M. D., Hauer, M. K., Selvan, P., Koku, E., & Leader, A. E. (2020). Dimensions of misinformation about the HPV vaccine on Instagram: Content and network analysis of social media characteristics. Journal of medical Internet research, 22(12), e21451.

********** 

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

Reviewer #4: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2023 Dec 20;18(12):e0295414. doi: 10.1371/journal.pone.0295414.r002

Author response to Decision Letter 0


7 Sep 2023

To: Dr. Jasna Karacic Zanetti, Academic Editor, PLOS ONE

Re: Revised Manuscript, “Detecting Nuance in Conspiracy Discourse: Advancing Methods in Infodemiology and Communication Science with Machine Learning and Qualitative Content Coding”

Date: September 7th, 2023

Dear Prof. Zanetti,

Thank you for your July 31st, 2023 editorial and reviewer comments for the manuscript titled “Detecting Nuance in Conspiracy Discourse: Advancing Methods in Infodemiology and Communication Science with Machine Learning and Qualitative Content Coding” (Manuscript ID PONE-D-23-14668).

We thank you for your continued interest in our submission along with helpful reviewer comments. As directed by your correspondence, we are submitting a revised manuscript with changes highlighted in the text and this cover letter that details how we have addressed all reviewer comments and suggestions. Our responses are highlighted in yellow, with quoted text from the revised manuscript in 11-point font. The convention of pX, ¶A (i.e., page X, paragraph A) is adopted to indicate from where in the revised manuscript the text we reproduce arises.

Below, please find the editor and reviewer’s comments (in bold) and how each is addressed in a point-by-point response in the revised manuscript. Changes to the manuscript have also been highlighted in yellow in the revision submitted.

We thank you for the opportunity to submit this revised manuscript and look forward to your comments.

Journal Requirements

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

Author Response: Thank you for your feedback about the formatting of our manuscript. We have updated our manuscript so that it is now aligned with PLOS ONE’s style requirements.

2. In your Methods section, please include additional information about your dataset and ensure that you have included a statement specifying whether the collection and analysis method complied with the terms and conditions for the source of the data.

Author Response: Thank you for your comment for more information concerning our data collection and analysis method. We have added language to the Methods section detailing that our collection was compliant with the terms and conditions associated with publicly available data and terms per the Twitter API at the time of the study.

2.1 Data Collection and Analysis Overview

A total of 256,562 tweets were collected from the public streaming Twitter API per the terms available at the time of the study using keywords “5G” and covid-related words such as “coronavirus” and “covid-19” between March 25th to April 3rd 2020. We chose this time frame as it represents a period when the 5G conspiracy theory first became prominent, as shown in the spike in volume for “5G” posts in Figure 1. All personal identifiable information from tweets was removed in the reporting of the results to preserve anonymity. We note that due to the change in ownership, API policies, and name of the platform (Twitter has been renamed “X”), the terms and conditions of the streaming API used for data collection for this study are now longer applicable for current studies. IRB approval was not required as all data collected in this study was available in the public domain and results from the study have been deidentified and anonymized. The dataset and R syntax used to generate the results can be found at the following Open Science Framework (OSF) link: https://bit.ly/5G_Conspiracies. p13, ¶3

3. Thank you for stating the following in the Competing Interests section:

“Authors TKM and ZL are employees of the startup company S-3 Research LLC. S-3 Research is a startup funded and currently supported by the National Institutes of Health – National Institute of Drug Abuse through a Small Business Innovation and Research contract for opioid-related social media research and technology commercialization. TKM is also the CEO and a member of S-3 Research LLC with ownership. Author reports no other conflict of interest associated with this manuscript.”

Please confirm that this does not alter your adherence to all PLOS ONE policies on sharing data and materials, by including the following statement: "This does not alter our adherence to PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests). If there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Author Response: We have included the previously mentioned phrase into our disclosure statement to clarify that author affiliation with S-3 Research does not alter our adherence to PLOS ONE policies on sharing data and materials.

Disclosure Statement

TKM and ZL are employees of the startup company S-3 Research LLC. S-3 Research is a startup funded and currently supported by the National Institutes of Health – National Institute of Drug Abuse through a Small Business Innovation and Research contract for opioid-related social media research and technology commercialization. TKM is also the CEO and a member of S-3 Research LLC with ownership. This does not alter our adherence to PLOS ONE policies on sharing data and materials. Author reports no other conflict of interest associated with this manuscript.

REVIEWER #1

The manuscript ensures that the message conveyed is clear, concise, and easily comprehensible to the intended audience. Its well-written manuscript uses appropriate grammar, spelling, punctuation, and vocabulary to convey the author's ideas effectively. It also follows a logical structure that enables readers to follow the flow of ideas easily.

Author Response: We thank the reviewer for their positive comments regarding our manuscript.

REVIEWER #2

The manuscript is a well written study of misinformation recognized in online platforms such as Twitter. It takes a thorough view of various approaches to misinformation detection and classification and also examines the societal need for this activity, a field termed as Infodemiology in the health context.

2. Major Issues

The work has some excessively wide-ranging phrases and general statements, for example, ``Since science itself is an iterative process..." which result in a slightly long manuscript and could be pruned. Moreover, such statements are not strictly necessary, apparently.

Author Response: Thank you for your feedback concerning the phrasing used in our manuscript and overall length. In response we have restructured our introduction and discussion in order to remove repetitive language and overly general statements. We also rephrased sentences throughout the manuscript to make the language more concise and clearer when possible. Changes are numerous, please see highlighted changes in the revised manuscript submitted.

Minor Issues

Details such as the specific script/algorithm used for analysis in each Table could be presented in Supplementary Data or in an Appendix to make the work more replicable by the readership.

Author Response: Thank you for your feedback about making our analysis scripts accessible for replicability. We have added the R-script used for analysis into our open science framework link, as well as referencing the availability of the syntax in the revised methods section. See revised language below:

2.1 Data Collection and Analysis Overview

A total of 256,562 tweets were collected from the public streaming Twitter API per the terms available at the time of the study using keywords “5G” and covid-related words such as “coronavirus” and “covid-19” between March 25th to April 3rd 2020. We chose this time frame as it represents a period when the 5G conspiracy theory first became prominent, as shown in the spike in volume for “5G” posts in Figure 1. All personal identifiable information from tweets was removed in the reporting of the results to preserve anonymity. We note that due to the change in ownership, API policies, and name of the platform (Twitter has been renamed “X”), the terms and conditions of the streaming API used for data collection for this study are now longer applicable for current studies. IRB approval was not required as all data collected in this study was available in the public domain and results from the study have been deidentified and anonymized. The dataset and R syntax used to generate the results can be found at the following Open Science Framework (OSF) link: https://bit.ly/5G_Conspiracies. p13, ¶3

As an additional point, as the manuscript is long, adding a `table of contents' type element right at the start might be useful for the reader to get a bird's-eye view of all the sections and subsections; but journal format guidelines have the last word on this.

Author Response: Thank you for your comment about adding a table of contents element to our manuscript. We will reach out to the editor in order to see the feasibility of adding this feature.

REVIEWER #3

This is an excellent study and manuscript describing the use of a hybrid method for detecting conspiracy tweets linking 5G technology to COVID-19. I especially appreciated the authors’ suggested applications of their findings in the Discussion. A few points are provided below for the authors to consider.

• The Introduction provides an excellent and erudite background on infodemiology and content coding. It is longer than is standard, but perhaps worth keeping given the broad audience of PLOS One.

Author Response: Thank you for your feedback concerning the introduction section. In response to reviewer’s suggestions and those of other reviewers, we have edited the Introduction to maintain the main points in the previous version while trimming excess statements and rephrasing the language in order to make it clearer and more concise when possible. Changes are numerous, please see highlighted changes in the revised manuscript submitted.

• Given the large amount of information scientists encounter daily, many will only see the abstract of this study. It is therefore important to include some findings or results in the abstract. I urge the authors to add at least two sentences describing the outcome of their analysis of tweets linking 5G to COVID-19; several of the background sentences can be omitted, if needed, to accommodate the word limit. Even though the authors position the study as a showcase for their methods, it seems important to demonstrate the novel insights that the method provides. Similarly, the authors state that “this study showcases . . . .” in the abstract, without describing the study. These changes will help rectify the current feel of over-promoting the method instead of letting the data/outcomes speak for itself/themselves.

Author Response: Thank you for your important comment about featuring our results within the abstract. In response we have included in the abstract the major results from the study, including sentiment analysis and inductive coding in order to make the insights generated from our approach more explicit. We have also edited the abstract to remove any language that might be interpreted as self-promoting, such as the language suggested by reviewer. See revised abstract below:

Abstract

The spread of misinformation and conspiracies has been an ongoing issue since the early stages of the internet era, resulting in the emergence of the field of infodemiology (i.e., information epidemiology) which investigates the transmission of health-related information. Due to the high volume of online misinformation in recent years, there is a need to continue advancing methodologies in order to effectively identify narratives and themes. While machine learning models can be used to detect misinformation and conspiracies, these models are limited in their generalizability to other datasets and misinformation phenomenon, and are often unable to detect implicit meanings in text that require contextual knowledge. In order to rapidly detect evolving conspiracist narratives within high volume online discourse while identifying nuanced themes requiring the comprehension of subtext, this study describes a hybrid methodology that combines natural language processing (i.e., topic modeling and sentiment analysis) with qualitative content coding approaches to examine conspiracy discourse related to 5G wireless technology and COVID-19 on Twitter (currently known as ‘X’). Discourse that focused on correcting 5G conspiracies was also analyzed for comparison. Sentiment analysis shows that conspiracy-related discourse was more likely to use language that was analytic, reference social status, combative, expressed negative emotions, and be past-oriented. Corrections discourse was more likely to use words reflecting cognitive processes, prosocial relations, health-related consequences, and future-oriented language. Inductive coding identified conspiracist narratives related to global elites, anti-vax sentiment, medical authorities, religious figures, and false correlations between technology advancements and disease outbreaks. Further, the corrections discourse did not address many of the narratives prevalent in conspiracy conversations. This paper aims to further bridge the gap between computational and qualitative methodologies by demonstrating how both approaches can be used in tandem to emphasize the positive aspects of each methodology while minimizing their respective drawbacks.

• COVID-19 or coronavirus should be listed as a keyword, as it does not appear in the abstract or title. Consider added 5G as a keyword also, although that does appear in the abstract.

Author Response: Thank you for your suggestion concerning our keywords. In response we have now added the terms ‘COVID-19’ and ‘5G’.

• A description of the units shown should be added to each graph or in the figure captions. This could be accomplished by labeling the Y axis of each graph.

Author Response: Thank you for your comment about specifying the units used in Y axis for our figures. In response we have added to the figure caption the phrase “Average percentage of…” to specify what the Y axis is depicting.

• This study—including both the methods and findings—are of broad societal interest and may attract a wide readership. For this reason, I suggest that the authors add a simplified (flow?) diagram of the general methods used, highlighting the novel features and/or outcomes it permits.

Author Response: Thank you for your suggestion about including a flow chart that further illustrates the methods used in our study. We have now included a chart in the appendix and referenced it in the Introduction section.

• I suggest deleting the last sentence of the Conclusion, as it seems a bit “preachy” for a scientific paper and may work better in an editorial.

Author Response: Thank you for your feedback concerning the language used in our Conclusion. In response we have edited the language so that it focuses back on the advantages of using the approach featured in the manuscript and deleted the last sentence as suggested. See revised language below:

5. Conclusion

The advancement of communication technologies and the continued emergence of new social media platforms present difficult challenges for researchers looking to investigate the highly dynamic information ecosystem of the 21st century. Fortunately, these same rapid advancements in technology can also be harnessed by researchers as powerful tools to navigate these complex environments. However, as demonstrated in the current paper, the human perspective is equally crucial in this line of work to compensate for the shortcomings that artificial intelligences have towards understanding human endeavors. Due to the rapid pace of modern discourse, words that can be key identifiers for a dangerous conspiracy in one context can be completely irrelevant in a different grouping of text. Within these online conversations, where the boundary between signal and noise is constantly shifting due to emerging and continually evolving narratives, it is crucial to recruit the signal detection capabilities of both machine learning models and human beings to adequately address current and future misinformation challenges now endemic in our global information society. p41, ¶2

REVIEWER #4

This study investigated conspiracy discourse related to 5G wireless technology by using a combination of NLP (i.e., topic modeling and sentiment analysis) and qualitative content analysis.

1. The abstract should reflect the main findings of the current study.

Author Response: Thank you for your important comment about featuring our results within the abstract. In response we have included in the abstract the major results from the study, including sentiment analysis and inductive coding in order to make the insights generated from our approach more explicit. See revised abstract below.

Abstract

The spread of misinformation and conspiracies has been an ongoing issue since the early stages of the internet era, resulting in the emergence of the field of infodemiology (i.e., information epidemiology) which investigates the transmission of health-related information. Due to the high volume of online misinformation in recent years, there is a need to continue advancing methodologies in order to effectively identify narratives and themes. While machine learning models can be used to detect misinformation and conspiracies, these models are limited in their generalizability to other datasets and misinformation phenomenon, and are often unable to detect implicit meanings in text that require contextual knowledge. In order to rapidly detect evolving conspiracist narratives within high volume online discourse while identifying nuanced themes requiring the comprehension of subtext, this study describes a hybrid methodology that combines natural language processing (i.e., topic modeling and sentiment analysis) with qualitative content coding approaches to examine conspiracy discourse related to 5G wireless technology and COVID-19 on Twitter (currently known as ‘X’). Discourse that focused on correcting 5G conspiracies was also analyzed for comparison. Sentiment analysis shows that conspiracy-related discourse was more likely to use language that was analytic, reference social status, combative, expressed negative emotions, and be past-oriented. Corrections discourse was more likely to use words reflecting cognitive processes, prosocial relations, health-related consequences, and future-oriented language. Inductive coding identified conspiracist narratives related to global elites, anti-vax sentiment, medical authorities, religious figures, and false correlations between technology advancements and disease outbreaks. Further, the corrections discourse did not address many of the narratives prevalent in conspiracy conversations. This paper aims to further bridge the gap between computational and qualitative methodologies by demonstrating how both approaches can be used in tandem to emphasize the positive aspects of each methodology while minimizing their respective drawbacks.

2. The first paragraph concludes with a call to advance methodologies that address the ongoing proliferation of misinformation that is overwhelming the social media ecosystem. I got a little lost here. This conclusion seems disconnected from the main reasoning of this paragraph. Why can't the aforementioned infodemiology applications address the current problems? What areas in particular need to be advanced?

Author Response: Thank you for your feedback concerning the language used in the introduction. Our intent was to preview further discussion of the potential need to examine new methodologies to address misinformation within the field of infodemiology and not to specifically critique prior work. In response we deleted the last sentence of the first paragraph, edited the first paragraph for greater clarity, and restructured the introduction more generally to better contextualize prior infodemiology and machine learning approaches to the misinformation challenge, while also providing more rationale of why new methodologies also warrant further study. Changes are numerous, please see our revised Introduction section in the revised manuscript.

3. The authors should clearly distinguish between misinformation and conspiracies. They are two distinct concepts, but the authors used them interchangeably throughout the manuscript.

Author Response: Thank you for your comment about distinguishing between misinformation and conspiracies. We have edited the manuscript in the introduction and discussion to include the phrase “misinformation and conspiracies” when referring to discourse containing false information. For our analysis we have continued to just refer to this discourse as misinformation, however, we included language in the methods to explicitly state that we’re using the term misinformation to refer to both concepts for brevity and have also included additional language in the revised Limitations section clearly stating the limitations of this coding approach. See additional explanation below:

2.3 Deductive and Inductive Coding Schemes

In order to characterize highly prevalent misinformation and conspiratorial narratives in the corpus, the top 10 most retweeted tweets from all 20 BTM topic outputs were extracted and manually coded for relevance first using a deductive coding scheme adapted from existing COVID-19 misinformation themes from the literature (Haupt et al., 2021a; Islam et al., 2020), and then coded again using an inductive approach identifying context specific themes related to 5G. While misinformation and conspiracies are distinct concepts, the current study will refer to both as ‘misinformation’ within the analysis for brevity. p17, ¶1

4.4. Limitations

Content coding took place several months after the initial timeframe of the study. While having the time gap allowed us to assess which accounts were deleted or suspended since the initial 5G discourse, we were unable to code for affiliations of those deleted users using publicly available profile data. As stated in the methods section, the top 10 retweeted tweets do not account for every tweet associated within a topic cluster. While the most retweeted tweets account for a substantive proportion of the topic cluster’s tweet volume, and in most cases a majority, there is still some level of uncertainty when characterizing the discourse even if the uncoded tweets share textual similarity to the coded posts. Additional measures such as the use of sentiment analysis, which in this study was applied to the full topic clusters, can also mitigate these concerns since they account for information provided in the uncoded tweets. Finally, though false information can be categorized as “misinformation,” “disinformation,” “mal-information”, and some information can also be classified as “conspiracy” based on intent and content, this study did not differentiate between these categories, opting to call all false information, regardless of intent, “misinformation.” This lack of differentiation limits internal reliability and the study’s ability to identify potential differences in rhetoric associated with the nuances of false information dissemination. p40, ¶2

4. I have some concerns over the generalizability of the current study. The NLP of this study generally evolves around the 5G-COVID conspiracy. How could the findings be generalized to misinformation detection in other disciplines, say politics, climate change, etc.?

Author Response: Thank you for your feedback concerning the generalizability of the NLP methods showed in the current study. In response we have added additional language discussing how our approach can be adapted for other topics. We have also included a flow chart diagram in the appendix (Figure A1) to further illustrate how our approach can be adapted for analyzing other discourse topics besides 5G conspiracies. We have also clarified that the Discussion section of the study that results are not necessarily generalizable to misinformation discourse on other pressing social and health topics, but that the general methodological approach could be leveraged in a similar fashion to provide more contextual classification to misinformation specific to these topics. See revised language below:

Interpreting sentiment scores can be further complicated when accounting for the fact that the discussion topic can also influence the average emotional tone of a discourse. For example, discussions about a pandemic may have higher percentages of negative affect words (e.g., “death”, “tragedy”) compared to lighter conversation topics such as gardening. For this reason it is difficult to determine what is an appropriate threshold for meaningful sentiment scores across topics. While this can limit what researchers can infer from these analyses, this study addresses this limitation by using a relativistic interpretation of sentiment scores. More specifically, discourse will be marked as high in a sentiment category based on whether it is in the 90th percentile of scores within the corpus. Using the 90th percentile as a threshold marker allows us to account for the specific context of the 5G discourse when making judgements for determining what is a high level of sentiment. This approach is also adaptable across a wide array of topics since percentiles indicate which cases are high and low for a given metric based on the sample distribution. To further illustrate this point, a post containing 5% of death-related words may be in the 95th percentile for discourse about gardening, making it a “high” amount, but be within the 50th percentile for pandemic-related discourse, making it a typical percentage within the context of that corpus. Additionally, this study incorporates a qualitative coding approach to account for contextual information that is typically overlooked in sentiment analysis but can be recognized by a human coder. p8, ¶2

While building effective classifier models for 5G COVID conspiracies is possible during 2023 after being well documented, only human coders had the capability to accurately recognize text that contains these narratives when they first emerged online in 2020. We intended to address these limitations by demonstrating the general utility of a hybrid approach that incorporates human coders to take advantage of the efficiencies machine learning techniques provide researchers when working with big datasets while still accounting for contextual nuance from using qualitative approaches. Even though the definition of misinformation may vary in other discussions (e.g., politics, climate change, other social issues), the general principles of the methodology described could be leveraged to provide more up-to-date and richer contextual insights into how these conspiracy-related discourses evolve over time p38, ¶2

5. Please elaborate on the relativistic approach: define it and justify its use in the current context.

Author Response: Thank you for your comment about expanding on the relativistic approach mentioned in our study. We have included additional language to the manuscript to further define this approach as well as including examples for how this approach can be adapted for other discourses.

Interpreting sentiment scores can be further complicated when accounting for the fact that the discussion topic can also influence the average emotional tone of a discourse. For example, discussions about a pandemic may have higher percentages of negative affect words (e.g., “death”, “tragedy”) compared to lighter conversation topics such as gardening. For this reason it is difficult to determine what is an appropriate threshold for meaningful sentiment scores across topics. While this can limit what researchers can infer from these analyses, this study addresses this limitation by using a relativistic interpretation of sentiment scores. More specifically, discourse will be marked as high in a sentiment category based on whether it is in the 90th percentile of scores within the corpus. Using the 90th percentile as a threshold marker allows us to account for the specific context of the 5G discourse when making judgements for determining what is a high level of sentiment. This approach is also adaptable across a wide array of topics since percentiles indicate which cases are high and low for a given metric based on the sample distribution. To further illustrate this point, a post containing 5% of death-related words may be in the 95th percentile for discourse about gardening, making it a “high” amount, but be within the 50th percentile for pandemic-related discourse, making it a typical percentage within the context of that corpus. Additionally, this study incorporates a qualitative coding approach to account for contextual information that is typically overlooked in sentiment analysis but can be recognized by a human coder. p8, ¶2

6. The first paragraph on page 9 appears disjointed as it initially discusses three modes of data interpretation, then transitions to discussing inductive and deductive analysis. It would be beneficial to establish a coherent connection between these two sets of approaches to ensure the flow of ideas and the logical progression of the argument.

Author Response: Thank you for your comment about the modes of data interpretation discussed in the manuscript. Our initial intent when discussing these modes were to further illustrate the advantages of human coders when interpreting data, however, we realize that this could be interpreted as introducing a formal coding approach for analysis. In order to make this clearer, we have edited the language in this paragraph to more explicitly emphasize the utility of human coders when identifying emerging narratives compared to machine learning approaches.

1.3 Inductive and Deductive approaches for content coding

In recent years there have been calls for researchers to recognize the impact of contextualization when interpreting data, such as accounting for changes in semantics within a dataset over time (Poirier, 2021). When identifying emerging narratives within online discourse, where meanings of text vary due to novel co-occurrences of words, the boundaries between noise and relevant signal are constantly shifting. Compared to machine learning approaches, which requires an existing training dataset of annotated posts to be effective, human coders are more adaptable at updating their background knowledge of events, making them effective at recognizing text containing previously undocumented narratives. To account for contextualization factors, our study utilized both inductive and deductive analytic techniques to examine tweets related to conspiracy discourse that states a relationship between COVID-19 and 5G technology. p9, ¶1

7. I guess most of the audience of this paper should get some foundational knowledge, like what inductive and deductive coding are, as well as their pros and cons. So the authors could cut this part.

Author Response: Thank you for your comment about the inductive and deductive coding sections of our manuscript. We have edited the language in these sections to make these sections more concise while still providing a description of these approaches.

Based in grounded theory (Glaser and Strauss, 1967), inductive coding is an iterative data analytic process centered on constant examination and comparison, which allows for theory development and explicit coding procedures (Corbin and Strauss, 1990; Boyatzis 1998). The primary benefits of an inductive approach is that it allows researchers to code texts using labels that are both aligned with the data and free from the influence of extant concepts, and detect tacit elements or connotations of the data that may not be apparent from a superficial reading of denotative content (Suddaby, 2006). For the current study, inductive coding was used to identify narratives and rumors prominent within the 5G discourse.

Deductive coding, on the other hand, refers to a top-down coding process intended on testing whether data coincide with existing assumptions, theories, or hypotheses (Fereday & Muir-Cochrane, 2006). For the deductive coding scheme in the current study, we coded for whether posts contained misinformation or misinformation corrections based on whether it made statements claiming that 5G wireless technology causes COVID-19, or actively refutes the conspiracy. The criteria for classifying 5G-related misinformation and corrections was adapted from previous frameworks identifying COVID-related misinformation (Haupt et al., 2021a; Islam et al., 2020; Mackey et al., 2021). The researchers also coded for whether tweets expressed a positive, negative, or neutral stance towards 5G conspiracies. Coding for the user’s stance makes it possible to trace and compare general sentiment across topics without having to account for specific themes. p9, ¶2-3

8. Deductive coding is a coding process intended to test whether data coincides with existing assumptions, theories, or hypotheses. I agree with that. But I don’t see any assumptions or hypotheses that the authors sought to test.

Author Response: Thank you for your comment about clarifying the existing assumptions and theories associated with our deductive coding approach. We have edited this section to make our coding criteria clearer as well as made it more explicit that we adapted existing frameworks for defining misinformation from previous COVID-related studies.

Deductive coding, on the other hand, refers to a top-down coding process intended on testing whether data coincide with existing assumptions, theories, or hypotheses (Fereday & Muir-Cochrane, 2006). For the deductive coding scheme in the current study, we coded for whether posts contained misinformation or misinformation corrections based on whether it made statements claiming that 5G wireless technology causes COVID-19, or actively refutes the conspiracy. The criteria for classifying 5G-related misinformation and corrections was adapted from previous frameworks identifying COVID-related misinformation (Haupt et al., 2021a; Islam et al., 2020; Mackey et al., 2021). The researchers also coded for whether tweets expressed a positive, negative, or neutral stance towards 5G conspiracies. Coding for the user’s stance makes it possible to trace and compare general sentiment across topics without having to account for specific themes. p9, ¶3

9. My understanding is that this study summarized the topic and linguistic characteristics of 5G-related conspiracies on Twitter using topic modeling, LIWC, and manual content analysis. First, it is not about detection. The paper did not present any algorithms or computations that could be used to detect falsehoods in the future. Second, these methods are not quite new (see below)

Rashkin, H., Choi, E., Jang, J. Y., Volkova, S., & Choi, Y. (2017, September). Truth of varying shades: Analyzing language in fake news and political fact-checking. In Proceedings of the 2017 conference on empirical methods in natural language processing (pp. 2931-2937).

Zhou, C., Li, K., & Lu, Y. (2021). Linguistic characteristics and the dissemination of misinformation in social media: The moderating effect of information richness. Information Processing & Management, 58(6), 102679.

Massey, P. M., Kearney, M. D., Hauer, M. K., Selvan, P., Koku, E., & Leader, A. E. (2020). Dimensions of misinformation about the HPV vaccine on Instagram: Content and network analysis of social media characteristics. Journal of medical Internet research, 22(12), e21451.

Author Response: Thank you for your comment about clarifying the impact and potential novelty of our study. In response we have added additional language to our introduction to clarify that we are incorporating existing methodologies into a streamlined approach in order to conduct more in-depth characterization (versus “identification”) of emerging narratives in online discourse and have made this distinction consistent throughout the manuscript. We have also included the above mentioned references as examples of work that use these methods. We also clarify that by “detection”, this word can refer to a recognition process that is accessible to both humans and machine learning classifier models. This is further illustrated by the fact that the word “detect” has existed long before machine learning techniques were prevalent. Further, we emphasize that the approach described in this study can be used to streamline annotation of posts to develop relevant training datasets, which is a crucial step that needs to be taken before it’s possible to develop effective classifier models for future conspiracy discourse. Changes are numerous, please find some of the main changes addressing this comment below:

To gain a more thorough understanding of the nuanced patterns and dynamics underlying the spread of misinformation, our study introduces a hybrid analytic approach using metadata from 5G-COVID conspiracy discourse on Twitter (currently known as ‘X’ but referred to as Twitter for this paper) aimed at leveraging both the efficiency of NLP techniques and the qualitative schema afforded by human coders. More specifically, this study used topic modeling and sentiment analysis to identify influential posts and characterize the discourse, and then used both inductive and deductive coding to detect context-specific narratives that would not be measured by standard sentiment dictionaries. While NLP and qualitative coding methods have been widely used throughout the literature for identifying and characterizing misinformation (see the following for examples: Haupt et al., 2021a; Rashkin et al., 2017; Massey et al., 2020; Zhou et al., 2021), this paper combines these methods into a streamlined approach that can be utilized for rapidly characterizing nuanced themes and emerging narratives within large scale online discourses that requires significantly less burden on human annotation, as illustrated in the flow diagram in S1 Appendix. The hybrid method utilized in this study also has the potential to generate nuanced training data for machine learning classifier models without the need for annotating thousands of posts.

This study aims to further bridge the gap between computational and qualitative methodologies by demonstrating how NLP and manual annotation can be used synergistically to emphasize the positive aspects of each approach while minimizing their respective drawbacks. In order to assess the utility of this methodology, themes generated by this approach will be compared for consistency with 5G-COVID conspiracy themes previously identified on Twitter (Ahmed et al., 2020; Flaherty et al., 2022; Langguth et al., 2022), Facebook (Bruns et al., 2020), and Instagram (Quinn et al., 2021). Further, this study will extend the current literature by providing in-depth characterization of discourse focusing on correcting false information, which can be used to inform counter strategies for online misinformation propagation p4-5

Attachment

Submitted filename: R1 Response Letter 5G submit.docx

Decision Letter 1

Stefano Cresci

22 Nov 2023

Detecting Nuance in Conspiracy Discourse: Advancing Methods in Infodemiology and Communication Science with Machine Learning and Qualitative Content Coding

PONE-D-23-14668R1

Dear Dr. Haupt,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Stefano Cresci

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Thank you for the thorough work carried out as part of this revision. The reviewers are now fully satisfied with your work and the state of the manuscript. All concerns appear to be adequately addressed, including those from Reviewer #4 who was unavailable for this second round of review. In light of this revision, I can recommend publication.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: (No Response)

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: (No Response)

Reviewer #3: I Don't Know

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: (No Response)

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: (No Response)

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The manuscript presented in an intelligible fashion and written in standard English is a testament to the author's meticulousness and commitment to professionalism. From the very first line, clarity and coherence reign supreme, allowing readers to effortlessly follow the author's train of thought. The use of standard English further enhances the manuscript's accessibility, ensuring that it is easily readable to a wide range of audiences. The choice of appropriate language, precise grammar, and coherent structure all contribute to this professional tone. This manuscript showcases the author's dedication to delivering a high-quality piece of work, reflecting their expertise and attention to detail. It is a shining example of how effective communication and adherence to language standards can elevate the impact and credibility of any written document.

Reviewer #2: (No Response)

Reviewer #3: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Abdulhamid Abuaniza

Reviewer #2: No

Reviewer #3: No

**********

Acceptance letter

Stefano Cresci

11 Dec 2023

PONE-D-23-14668R1

Detecting Nuance in Conspiracy Discourse: Advancing Methods in Infodemiology and Communication Science with Machine Learning and Qualitative Content Coding

Dear Dr. Haupt:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Stefano Cresci

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Checklist. STROBE statement—Checklist of items that should be included in reports of observational studies.

    (DOCX)

    S1 Appendix. Hybrid approach for characterizing social media discourse using machine learning and qualitative methodologies.

    (PDF)

    S1 Table. Most retweeted accounts associated with each BTM cluster.

    (PDF)

    Attachment

    Submitted filename: R1 Response Letter 5G submit.docx

    Data Availability Statement

    The full dataset collected from Twitter and the top retweeted tweets selected for content coding can be found at the following link on the Open Science Framework (OSF): https://bit.ly/5G_Conspiracies (DOI 10.17605/OSF.IO/YRNMX). Usernames, links to the tweet, and all personally identifiable information were removed to preserve anonymity.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES