Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2021 Aug 9;16(8):e0253300. doi: 10.1371/journal.pone.0253300

Detecting sentiment dynamics and clusters of Twitter users for trending topics in COVID-19 pandemic

Md Shoaib Ahmed 1,*,#, Tanjim Taharat Aurpa 1,#, Md Musfique Anwar 1,#
Editor: Hocine Cherifi2
PMCID: PMC8351922  PMID: 34370730

Abstract

COVID-19 caused a significant public health crisis worldwide and triggered some other issues such as economic crisis, job cuts, mental anxiety, etc. This pandemic plies across the world and involves many people not only through the infection but also agitation, stress, fret, fear, repugnance, and poignancy. During this time, social media involvement and interaction increase dynamically and share one’s viewpoint and aspects under those mentioned health crises. From user-generated content on social media, we can analyze the public’s thoughts and sentiments on health status, concerns, panic, and awareness related to COVID-19, which can ultimately assist in developing health intervention strategies and design effective campaigns based on public perceptions. In this work, we scrutinize the users’ sentiment in different time intervals to assist in trending topics in Twitter on the COVID-19 tweets dataset. We also find out the sentimental clusters from the sentiment categories. With the help of comprehensive sentiment dynamics, we investigate different experimental results that exhibit different multifariousness in social media engagement and communication in the pandemic period.

1 Introduction

People’s involvement in the online social network (OSN) has increased during the COVID-19 pandemic, as regular activities move online. Numerous uses of OSN (e.g., people use OSN for expressing their opinion, communicating with family members, online meetings, etc.) are showed up at this time. Like other OSN, the use of popular microblogging service Twitter has also been impacted. It becomes a popular media for the leaders to communicate with general people and make them aware of public health during this health crisis [1]. So, people usually spend more time on Twitter, and users are more active than at any other time. Their involvements increase during the lockdown period to get the latest news on COVID-19. At the same time, they share their opinions and feelings with their friends through it. As a result, analysis of Twitter data draws vast attention from researchers in this pandemic.

Sentiment analysis is a technical study about people’s emotions, opinions, and attitudes [2]. It is an effective way to measure people’s thoughts on particular topics. Moreover, sentiment analysis can convey various impacts on society in several ways. Different types of mental anxieties arise in this pandemic situation, and all those mental conditions can be summarized through sentiment analysis. We can quickly determine the extensive state of depression and panic disorder of persons in a society or community from the sentiment analysis result. We need to apply different virtual depression optimizers in those depressed persons to bring some positive ramifications to society. Again, the success of many applications like recommendation systems depends on the sentiments of social users. Sentiment analysis for active users is a more efficient way to track public opinion. In the coronavirus pandemic, these types of research have significant contributions to help government and policymakers. Authors in [3] analyze Indian people’s sentiment during corona lockdown. They used some popular hashtags for measuring positivity and negativity in people.

People concentrate on many different topics during this whole pandemic period. Some people posted tweets about the COVID-19 tests and deaths. Again, some people focused on job cuts, online education, or politics. Besides the new topics arrival among people, many different thoughts regarding those topics are shown in this pandemic situation. In [4], authors determine top trending topics using hashtags for detecting COVID-19 conspiracy theories. Another work [5] detected trending topics and clustered them using the k-mean clustering algorithm. So, the determination of trending sub-topics at different time windows is essential to understand the public’s changing interests properly.

Our work includes the concept of analyzing active users’ different sentiments, such as positive, negative, and neutral sentiments at a particular time interval for trending topics related to COVID-19. This work concentrates on people’s positive, negative and neutral sentiments on top-k trending sub-topics in Twitter related to COVID-19. We also track the changes that occurred in top trending topics in Twitter and user’s sentiment. The main contributions of our research are summarised below:

  • Propose a model that lists top-k trending topics in Twitter due to COVID-19 pandemic at a different time interval.

  • We are modeling and evaluating users’ sentiments towards different topics of a given query.

  • Modeling the sentiment dynamics of different topics.

  • Detection of sentiment clusters and tracking their changes for top-k trending topics over time.

We have accomplished this work as an extensive version of our extended abstract that appeared at [6]. The significant key points of our additional contributions in this journal version are listed below:

  1. We cluster the Twitter users based on their sentiments on different topics related to COVID-19.

  2. We model the degree of topical activeness of the users according to the rank of the topics of a given query.

  3. We revise the existing algorithm to list top-r users according to their overall activities related to top-k trending topics.

  4. We conduct our experiment on a new dataset that contains COVID-19 related tweets. We collect those tweets with real-time Twitter lookup API and prepare them according to our requirements.

  5. The COVID-19 outbreak results in an overwhelming amount of information on different topics, and also users’ sentiments vary quickly. As a result, we consider a non-overlapping time window with shorter time intervals to monitor social users’ sentiments.

  6. In most cases, tweets are very informal, extremely noisy, and also contain grammatically incorrect phrases. To improve the quality of data, we apply a set of pre-processing steps such as Tokenization, Lemmatization, Stemming, Sentence Segmentation, etc. for performance enhancement.

  7. We consider a self-regulating topic modeling approach known as T-LDA (Twitter-Latent Dirichlet Allocation) [7] to detect the topic from a tweet.

2 Related work

Rajesh et al. [8] scrutinized Tweets related to the coronavirus to get out the appropriate and most accurate with minor misinformation spread. Here, applied only the LDA (Latent Dirichlet Allocation) analysis to find out the negative sentiments dominated the tweet as expected as the virus highly contagious that was clear from the sentiment analysis significantly depends on some words. This work only shows the negative sentiment just from some particular topics without analyzing any model in time intervals and devoid of any sentiment model and analysis. Jim Samuel et al. [9] presented an issue surrounding public sentiment leading to the testimony of growth in fear sentiment and negative sentiment. This approach does not examine the change of sentiment aloft time. An evolving method [10] illustrates the sentiment analysis country-wise related to COVID-19. The author evokes sentiments from tweets only with the judgment of some growing keyword about coronavirus of examining the top trending topics over time. They also discuss just positive and negative sentiments. This approach does not consider any extensive topical model (ex. T-LDA) and neutral sentiment. Yin et al. [11] introduced a structure to study the topic and sentiment dynamics due to COVID-19 from extensive Twitter posts. A recent proposal [12] to analyze social media (micro-blogging like as Twitter called Weibo) data in the early stage of COVID-19 in China and proposed a topic extraction and classification model. The opinion’s appearance showed that the topic’s approach is stable and viable for understanding public opinions. Moreover, they showed the statistical results of the percentage of first-level topics of COVID-19. A machine learning-based sentiment analysis [13] introduced a hybrid approached to find out the sentiments on regular tweets with polarity calculations. The polarity score is measured from a score range of -1 to 1 based on words used and then used three sentiment analyzer W-WSD, TextBlob, and SentiWordNet. Those analyzers are then validated with the Waikato Environment for Knowledge Analysis (Weka) to measure the best result. Pandey et al. [14] proposed a metaheuristic method depend on K-means and cuckoo search. This method is applied to the different tweeter datasets to determine the optimum cluster-heads in terms of sentiment. It is also compared with differential evolution, particle swarm optimization, cuckoo search, improved cuckoo search, two n-grams, and gauss-based cuckoo search.

A clustering-based approach on sentiment analysis is proposed by Gang [15] where they accosted a weighting method called term frequency-inverse document frequency (TF-IDF) on document-based content. Over the two existing forms of propositions, they listed a competitive advantage, one is allegorical methods, and another is supervised learning methods. They used the simple k-mean clustering algorithm to find the positive and negative categories of clusters. An SVM classifier combined with a cluster organization provided better classification accuracies than a stand-alone SVM to control the impressions, feelings, and biases presented in the source material to assess tweet sentiment analysis [16]. They used an algorithm called C3E-SL in their analysis, capable of combining classifier and cluster assemblies. This algorithm will improve tweet classifications from clusters’ additional details, assuming the same classmark is more likely to be shared by similar instances from the same clusters. Shreya et al. [17] suggested a study that came from various clusters that belong to polarity wise and subjective wise with sentiment ratings. The sentiment scores are assessed here using Afinn and TextBlob. Therefore, they used extensive data, calculating the Euclidean distance in less time and using the K-means clustering algorithm technique. An extensive approach [18] to find out the appearance of clustering techniques on document sentiment analysis. In their first approach, they showed two types of notices. The first one is a good performance, and the second one is the poor performance when applying the K-means-type clustering algorithm on balanced and unbalanced datasets, respectively. To avoid this problem, they designed a weighting model that worked well on both unbalanced and balanced datasets that were better than the conventional weighting model. Feng et al. [19] researched clustering methods on standard blog posts and got natural emotions from web blogs by topics or keywords, which is a typical approach. A novel approach based on Probabilistic Latent Semantic Analysis (PLSA) is performed. An emotion-oriented clustering technique is proposed to find common emotions affirming the connection of fine-grained sentiment between blogs and blog posts. Farhadloo et al. [20] proposed a score representation with aspect level sentiment identification. This identification is based on positiveness, neutralness, and negativeness. This process is designed with a 3-class SVM classifier to determine feature sets according to a 3-dimensional representation (positive, negative, and neutral). To improve clustering results, authors utilized a bag of nouns (BON) rather than a bag of words (BOW).

3 Preliminary and proposed framework

We introduce some relevant concepts before defining the problem statement. Then we give an overview of our proposed framework.

Social Graph: We model the Twitter network as a social graph G = (U, E, T), where U is the set of nodes (users), E is the set of connections or virtual social relationships among the Twitter users (such as the following relationships in Twitter), and T={T1,T2...,Tm} is the set of topics discussed by the social users U [21].

Topic: A topic is a collection of the most representative words for that topic. For example, politics topic has words like election, government, democratic, parliament, etc. about politics [22, 23].

Social Stream: A social stream S is a continuous and temporal sequence of the tweets posted by the social users U.

Query: An input query Q={Tq} consisting top-k trending Topics Tq={Ti,Ti+1...,Tk} at a particular time interval.

Overlapping Time Window: A window of a predefined length len is moved over the social stream S and specifies the intervals to analyze. Let Γ = <t1, t2, …, tn> be a sequence of points in time, Im an interval [tilen, ti] of len, where 0 < leni. We partition Γ into set of equal-length intervals denoted as I={I1,...,Im}. We consider an overlapping window partially overlaps with the prior window. The degree of overlap is controlled by the parameter Δt [24].

Topical Involvement Score: For each user uiU, we compute her involvement score towards the query Q in a time interval Im using Eqs 1 and 2 which measures ui’s relative participation compared with the most active users at that time interval Im.

σ(ui,Q,Im)=ψ(ui,Q,Im)maxuzU(Q,Im){ψ(uz,Q,Im)} (1)
ψ(uz,Q,Im)=l=1k(k+1-l)·κ(uz,Q,Im) (2)

where κ(ui, Q, Im) indicates the total number of tweets posted by ui related to lth number topic in Q at Im.

Our proposed approach has three stages as presented in Fig 1. Firstly, the pre-processing is performed to remove irrelevant data from the social stream S. Secondly, we apply the topic modeling method on the cleaned data to infer the latent topics and then select top-k trending topics. Then we apply our proposed algorithm to the processed social streams to find top-k trending topics and users’ involvement scores. Finally, we detect involved uses’ sentiment dynamics and clusters of top-involved users at different time intervals.

Fig 1. The workflow of the proposed framework (a methodical diagram representing the entire process from data collection to topic modeling and find out the sentiment dynamics and clusters).

Fig 1

3.1 Data pre-processing for topic detection

In general, tweets are informally written and often contain grammatically wrong sentence formations with misspellings and non-standard words. Tweets also contain numerous non-standard forms (e.g., comeee for Come, goooood for good), informal abbreviations (e.g., tmrw for tomorrow, lemme for let me, wknd for weekend), phonetic substitutions (e.g., gdn8 for good night, 4eva for forever, 2day for today), etc. For removing those, we follow some steps to lead the standardization for our next stages. In the first stage, we remove the noise entities such as HTML tag, Stop words, Punctuations, White Spaces, URLs, etc. The next stage is text normalization like as Tokenization, Lemmatization, Stemming, Sentence Segmentation, etc. Finally, word standardization gives us cleaned texts. To improve the quality of our tweet corpus and the fulfillment of the consequent steps, mentioned normalization of the tweets through linear substitution of lexical variants with their conventional forms proposed by Han et al. [25]

3.2 Topic detection from social stream

The use of hashtags (for example #coronavirus, #StayHome) to point out a tweet’s topic is common on Twitter. However, neither every tweet contains hashtags, nor hashtags have been written by following any rule. Thus, tracking hashtags rarely leads to the exact topic. Another topic modeling approach T-LDA (Twitter-Latent Dirichlet Allocation) [7], is a popular way of inferring topics on Twitter. It is a textual analysis tool that deals with microblogs like tweets. Tweets are limited to 140 characters, and within this limitation, a single tweet can refer to a single topic. This restricted characteristic of tweets intercepts traditional text mining tools in their successive execution. T-LDA resolves this issue and potentially works with tweets.

Twitter LDA has been implemented based on the following assumptions.

  • Assuming there are T topics on Twitter and each topic t is generated from background word distribution θB and topic word distribution θt. Latent value y dominated by Bernoulli distribution π. identifies a word w to be a background word (y = 0) or a topic word (y = 1).

  • Φu represents a user u’s topic of interest. It also determines the assignment of topic t for each word in tweets posted by u.

  • αd, βd, γd, and λd are the parameters of the Dirichlet prior on Φu, θt, π and θB respectively.

  • z is the determined topic for a tweet [26].

Fig 2 shows the graphical representation of T-LDA. Table 1 shows the word distribution for top-k topics (k = 3) in different time intervals from 23rd March, 2020 to 31st March, 2020.

Fig 2. Graphical representation of Twitter-LDA model.

Fig 2

Table 1. Topic word distribution for top-k trending topics in Twitter.

News Health COVID-19 Test Lockdown
Press Hospital Tested Home
Live Died Positive Locked
Conference Ventilator Confirmed Safe
CNN Drug Report Distancing
Fox News Patient Negative Staying
Briefing Medical Update Friend
World Nurse Total Hope
Staff Equipment Number Save
Time Supply Covid Family
Medium Doctors Rate Work

3.3 Top-k trending topics from social stream

In our proposed model, we set the value of the query Q at each time interval Im as the top-k trending topics related to COVID-19 at that Im. We define trending score (Λ(Ti, Im)) for each topic Ti according to Eq 3:

Λ(Ti,Im)=α×NTi+(1-α)×UTi,Im (3)

Where NTi indicates the total number of tweets related to topic Ti and UTi,Im represents the number of unique Twitter users who posted tweets on Ti at time interval Im. The parameter α ∈ [0, 1] balances the above two factors.

Table 2 shows how the changing value of alpha can effect the top trending topics at a particular time window. In this table top-k topics for three different values of α at two different time intervals (Im) Ia = 25/03 − 28/03 and Ib = 26/03 − 29/03. We consider the length of each Im for four days and shift this Ims for Δt = 1 day.

Table 2. Top-k trending topics for different values of α.

α Ia(25/03-28/03) Ib = (26/03-29/03)
0.3 News Health Economy News Health Politics
0.4 News Health Lockdown Health News COVID-19 Test
0.5 News Health Lockdown News Health COVID-19 Test

We use Eq 3 (α = 0.5) on different time intervals and detect top-k (k = 3) trending topics on Twitter. We consider seven-time intervals starting from 23rd March 2020 to 31st March 2020. Each of these time interval’s length is 3 days and we shift them by Δt = 1 days. We also measure how much of these trending topics are discussed by users at those time intervals. Table 3 shows the percentages of top-k = 3 trending topics that indicate it’s popularity among users.

Table 3. Top-k = 3 trending topics in different Im (we consider seven time intervals where len = 3 and Δt = 1).

I1
(23/3–25/3)
I2
(24/3–26/3)
I3
(25/3–27/3)
I4
(26/3–28/3)
I5
(27/3–29/3)
I6
(28/3–30/3)
I7
(29/3–31/3)
News
(19%)
News
(20%)
News
(20%)
News
(20%)
News
(21%)
Health
(22%)
Health
(23%)
Health
(16%)
Health
(18%)
Health
(20%)
Health
(18%)
Health
(21%)
News
(21%)
News
(22%)
Lock down
(16%)
Covid Test
(17%)
Covid Test
(19%)
Covid Test
(17%)
Covid Test
(20%)
Politics
(17%)
Politics
(16%)

Fig 3 represents the percentage of topics discussed by users at different time windows. It is a heatmap where different blue shades are used to indicate the trendiness of topics at various Im. We consider α = 0.5 to determine the trendiness of topics. We depict this heatmap using seven different time windows from 23/03/2020 to 31/03/2020. Each of them has a length of 3 days and is shifted by one day. We pick the top eight trending topics (Health; News; COVID-19 test; Lockdown; Economy, Politics, Relief, and Infection) on Twitter to visualize the distinction of the percentage of those topics discussion rate among users.

Fig 3. Heatmap representing the percentage of each topic discussed among users at various time interval Im (displays the top eight discussed topics at seven different time-intervals).

Fig 3

3.4 Users’ involvement Detection Algorithm

We develop an algorithm that can detect top-k trending topics before determining the top involved users at a particular time.

Algorithm overview. The algorithm, called Query Algorithm, identifies top-k topics from social stream S at each time interval Im through procedure TOP_K_TOPICS (line 9-17) at first. It enumerates the trending score η(Tj, Im) for each topic Tj and adds that score to a priority queue of size k (line 11-16). Then it returns the top-k topics based on their trending scores. Next, the algorithm finds the set of users UImQ from U for a given Q at each time interval Im and then computes users’ involvement score σ(ui, Q, Im) (line 3-6). Finally, the users are sorted by their involvement scores, and the proposed algorithm returns the top 20 users as output (line 7-8).

Algorithm 1 Query Algorithm

Require: G=(U,E,T),I,Q,S,k,α

Ensure: top-r active users UImr

1: for each ImI do

2:  Q ← TOP_K_TOPICS(S, Im, α)

3:  select UImQ from U ⊳ each uiU has to post certain number of tweets related to Q

4:  for each uiUImQ do

5:   compute σ(ui, Q, Im)

6:  end for

7:  Sort the list UImQ according to σ(ui, Q, Im)

8:  top-r active users UImr at each time interval Im

9:  Procedure TOP_K_TOPICS(S, Im, α)

10:  PPriorityQueue(k)

11:  for each TjT do

12:   compute the total number of tweets |ψ(ui, Q, Im)|

13:   generate user frequency matrix UTj,Im

14:   compute η(Tj, Im)

15:   P.add(η(Tj, Im))

16:  end for

17:  return top-k results from P

18: end for

3.5 Sentiment identification from social stream

For sentiment identification, we use VADER (Valence Aware Dictionary and Sentiment Reasoner) [27] is a lexicon and rule-based sentiment analysis appliance that precisely harmonizes to sentiments expressed in social media. VADER is open-source and licensed under the MIT available in GitHub. It is the rule-based sentiment analysis engine that carries out the grammatical and syntactical rules. In addition, it recognizes the intensity of sentiment in sentence-level text.

Our processed social streams pass through this engine for the analysis of sentiments and give a score. The scoring formulation is given below:

  • The compound score (ϱ) is calculated by summing the valence scores of each word in the lexicon, adjusted according to the rules, and then normalized to be between Υmax and Υmin. It is suitable for a single uni-dimensional measure of sentiment for a given sentence.

    Where, Υmin = −1 = most extreme negative and Υmax = 1 = most extreme positive. Here we take the graded thresholds for classifying sentences as either positive, neutral, or negative. Typical threshold values are:
    • positive sentiment: ϱ ≥ 0.05
    • neutral sentiment: ϱ > -0.05 and ϱ < 0.05
    • negative sentiment: ϱ ≤ -0.05
  • The pos, neg, and neu scores are the proportion of each category and the multidimensional measures of sentiment for a given sentence.

From Table 4, there we look at three columns. The first column is the social streams (tweets), the second column is polarity, where we observe the value of different sentiments between Υmax and Υmin after applying VADER. Then eventually, we classify the social streams as either positive, negative, or neutral.

Table 4. Sample tweets with sentiment polarity by VADER.

Social stream Polarity Result
Emerging markets have limited power to tackle recession #economy #businessnews #coronavirus #emerging #emergingmarkets #healthcareindustry Neg: 0.299,
Neu: 0.701,
Pos: 0.0,
Com.: -0.5719
Negative
My Mom’s a nurse and just tested positive for COVID-19. The caregiver is now the patient. Stay home for all the brave Neg: 0.0,
Neu: 0.562,
Pos: 0.438,
Com.: 0.7906
Positive
Under Armour Manufacturing Face Masks For Hospital Workers Amid Coronavirus Pandemic #economy #armour #coronavir… https://t.co/J7Rkz8WMjw Neg: 0.367,
Neu: 0.412,
Pos: 0.221,
Com.: -0.2019
Neutral

4 Experimental evaluation

In this section, we estimate the performance of our algorithm on a real Twitter dataset. We perform all experiments on an AMD Ryzen 7 3700U with Radeon Vega 10 Gfx (8 CPUs), 2.3 GHz Windows 10 PC with 32 GB RAM and 512GB NVME M.2 SSD.

4.1 Data set

We collect COVID-19 related tweets through Twitter lookup API’s endpoint that contains 100 million tweets with 10,000 users from 23 March 2020 to 31 March 2020.

4.2 Performance evaluation measure

We consider two performance evaluation measures, one is entropy, and another one is semantic cohesion.

Entropy measures with the Equation of 4 and 5 that betokens the randomness of topics discussed in clusters.

entropy({Cj}j=1r)=jr|U(Cj)||U|entropy(Cj) (4)
entropy(Cj)=-i=1npijlog2pij (5)

Here, |U(Cj)||U| is the weighted probability of a user in cluster Cj for discussing a trending topic. pij is the percentage of active users for that topic in the cluster. entropy({Cj}j=1r) measures the weighted entropy considering all topics over all the (r) clusters. Usually, a good topical cluster should have a low entropy value.

Semantic cohesion is measured with the following Equation from 6 to 8. For this purpose, we find out the main topic of activity of each user ui according to Eq 6.

λ(ui,Im)=freqmaxQACTS(ui,ψui) (6)

Then, the most recurrent topic in a cluster Cj at time interval Im defines with the Eq 7.

λ(Cj,Im)=freqmaxQλ(ui,Im) (7)

Finally, we find the semantic cohesion (expertness of cluster) denoted as ρ(Cj,Im) for a particular topic Tj at time interval Im with Eq 8.

ρ(Cj,Im)=#{uiCj,λ(ui,Im)=λ(Cj,Im)}|Cj| (8)

4.3 Experimental results

In this section, we have mentioned the findings of our experiments. Firstly we detect top-k = 3 trending topics and identify the involved users for those topics using our query algorithm. Then we determine their sentiments. Based on these experiments, we make different types of observations regarding users’ sentiments.

We consider our Table 3’s topics set in each time interval for further experiment. Our query is a set of topics and we fixed k = 3 and α = 0.5 for determining these topic sets. We consider all negative, positive, and neutral tweets for different time windows. We demonstrate the result not only for our query Q = {set of top-k = 3 trending topics} but also for individual topic in the query set. We sketch bar diagrams in Figs 46 to represent percentages of positive, negative and neutral tweets posted by our users on a particular time interval for News, Health and COVID-19 test. These three topics appear again and again as top trending topics. Sea green, Coral red and Royal blue bars are indicating the positive, negative, and neutral tweets, respectively.

Fig 4. Overall sentiment dynamics on Twitter at different time intervals Im for topic news.

Fig 4

Fig 6. Overall sentiment dynamics on Twitter at different time intervals Im for topic COVID-19 test.

Fig 6

Fig 5. Overall sentiment dynamics on Twitter at different time intervals Im for topic health.

Fig 5

After that, we concentrate on the most involved users’ sentiments towards COVID-19 related subtopics. We determine users’ involvement scores for our query topics (top trending topics from Table 3) at different time intervals. We identify the top 20 involved users at each time interval.

Our research has been accomplished with recent year’s tweets related to COVID-19. To preserve the privacy of users, we replace some alphabets with ‘*’ in usernames. Table 5 shows the sentiment dynamics for the top 20 involved users at each Im. To measure overall sentiment dynamics, we sum up users’ sentiment scores towards the query topics. In that table, we can see that the top 20 involved users’ list is changing over time.

Table 5. Top 20 active users with their overall sentiment.

I 1 I 2 I 3 I 4 I 5 I 6 I 7
js**an
(0.262)
Ve**ar
(-0.985)
PK**17
(-1.5)
ww**rr
(-2.20)
wo**er
(-1.843)
ze**ee
(-0.533)
we**de
(0.906)
ge**jo
(-0.287)
ma**te
(1.431)
tr**28
(0.222)
PK**17
(-2.7)
tr**28
(0.673)
wh**ra
(-1.377)
Ve**ar
(-0.16)
PK**17
(0.7989)
Pk**17
(0.441)
ra**99
(5.332)
mj**n7
(-2.050)
ki**59
(-0.886)
pu**v8
(-0.235)
HE**rk
(0.37)
ev**67
(-4.746)
da**sa
(0.221)
nu**me
(0.650)
mi**k0
(0.475)
jg**00
(0.002)
me**at
(0.183)
PK**17
(-5.4)
ca**oo
(0.588)
br**om
(0.642)
lg**87
(-2.051)
m2**30
(-3.171)
jb**28
(0.922)
ju**o2
(0.809)
da**y3
(-0.249)
bb**92
(-1.641)
Wa**ay
(0.078)
jg**00
(0.534)
ki**59
(-0.844)
is**gs
(-1.049)
ev**67
(0.602)
br**46
(-0.786)
ma**85
(-1.869)
RC**ic
(1.116)
br**20
(-1.038)
js**vr
(0.501)
go**bo
(3.167)
ck**rz
(-1.522)
Vi**rW
(-0.398)
le**84
(1.239)
Ma**85
(-0.642)
Th**te
(0.844)
jg**00
(1.068)
He**rk
(0.011)
Ma**85
(-1.156)
US**aK
(-1.194)
st**st
(-0.6)
Jb**hn
(-1.136)
Nu**60
(-2.599)
ex**99
(-0.407)
PK**17
(-4.2)
Ma**64
(-0.221)
TB**ne
(-0.661)
gr**r6
(-2.728)
HE**rk
(-0.442)
ma**te
(-0.398)
br**20
(-2.449)
Ka**49
(0.108)
PK**17
(-5.1)
Sh**3R
(-1.381)
ei**an
(-0.618)
De**y_
(-0.393)
Ki**59
(-0.118)
SC**22
(1.639)
Jb**hn
(-1.203)
Je**no
(-0.429)
Nu**60
(1.057)
co**ly
(0.921)
Ar**82
(-2.563)
Dj**a1
(0.840)
Ro**16
(-2.122)
Fi**ee
(-0.553)
Fi**ee
(-0.958)
Ma**85
(-2.478)
ch**ger
(0.348)
19**in
(-0.685)
De**y_
(-0.329)
ka**49
(-0.753)
Cr**ef
(0.365)
Eb**Jr
(1.351)
Je**rs
(-1.526)
be**ly
(0.029)
wr**ub
(-1.016)
Co**dy
(-0.522)
Ha**99
(-0.797)
Ci**t1
(-0.739)
Dj**a1
(0.098)
Fl**x1
(-0.745)
cc**ly
(0.458)
th**ia
(-0.090)
1f**at
(0.859)
Cr**ef
(0.365)
Ch**ger
(-0.093)
De**ns
(0.161)
De**ns
(-1.20)
wr**ub
(0.988)
js**vr
(-0.759)
wr**ub
(-1.016)
Ad**lG
(-0.368)
Ap**ne
(-3.092)
Ci**t1
(-1.840)
Bi**06
(-1.24)
ma**te
(-1.839)
ev**67
(-3.682)
st**st
(0.349)
Aa**ee
(-0.263)
24**ia
(-0.661)
st**st
(-2.2)
BF**in
(1.548)
lo**79
(-1.471)
br**at
(-1.344)
li**ze
(-1.064)
zi**is
(-1.043)
ra**an
(0.043)
wo**er
(-1.843)
19**in
(-2.913)
is**s
(2.898)
br**20
(0.926)
ev**on
(-1.303)
tr**28
(0.923)
ma**ra
(-1.524)
ge**jo
(1.2)
zi**is
(-2.281)
ex**99
(-0.153)
Un**MN
(2.203)
da**sh
(-0.562
ro**73
(-0.458)
lo**79
(1.194)
wa**27
(0.971)
ge**jo
(-0.042)

The reason behind this is that user’s interests and their involvements in trending topics vary over time. Another remarkable fact in this table is the change of users’ sentiments over time. Users who remain in the top 20 on the next Im have different sentiment scores.

Here we analyze some users’ sentiment dynamics with their involvement below:

  • PK**17 is highly involved in each Im. He has positive sentiments in I1,I2 and I3 and diverts to negative sentiments from I4 to I7. More highly involved users like PK**17 have different sentiment dynamics at each Im.

  • Other types of users like ma**te remain in the top 20 at some time windows. But also drop from the involvement list at the next or previous time windows. ma**te is a top involved user in I1, I2, I3, but vanishes from the list after I4. These users have various sentiment dynamics at a particular Im.

  • Some users suddenly appear in the top involved list, who have no existence in the list previously (e.g., jg**00). User jg**00 is not one of the top involved users at I1 and I2, while he is scoring top at the next three Im. This user has non-identical sentiment scores over time.

We analyze the top 20 users and bring out ten users whose average involvement in seven Im is greater than other users. We track the changes in their sentiment dynamics. In Fig 7 we sketch these 10 users’ sentiments. Here we can observe that users have different types of sentiment scores at different time windows. Even for some users, their sentiment dynamics changes from positive to negative or neutral. After another shift, it is changing into positive again. This heatmap provides clear visualization of sentiments’ change over time for a particular user.

Fig 7. Heatmap of selected ten users sentiment dynamics at each time-interval Im (shows the sentiments of selected ten users at different seven time-interval).

Fig 7

We also determine sentimental clusters based on the users’ sentiment scores. We find cluster CPos, CNeg and CNeu. We identify these clusters in two different ways. In Table 6, we identify clusters for each trending topic in the query set. For a user’s cluster membership identification, we sum up the sentiment scores of all tweets posted by that user on a particular topic. If she/he achieves a positive sentiment score, then he/she is the member of cluster CPos. For negative and neutral sentiment scores, a user is the member of CNeg and CNeu respectively. In all time intervals, the CNeu clusters have the highest number of members.

Table 6. Topic-wise sentimental clusters size at different Im.

Time interval → I 1 I 2 I 3 I 4 I 5 I 6 I 7
Clusters Trending topics
C Pos Topic 1 592
(News)
325
(News)
324
(News)
341
(News)
332
(News)
592
(Health)
606
(Health)
Topic 2 351
(Health)
301
(Health)
393
(Health)
317
(Health)
480
(Health)
351
(News)
357
(News)
Topic 3 197
(Lock Down)
253
(Covid Test)
309
(Covid Test)
281
(Covid Test)
307
(Covid Test)
197
(Politics)
219
(Politics)
C Neg Topic 1 644
(News)
631
(News)
642
(News)
595
(News)
673
(News)
644
(Health)
710
(Health)
Topic 2 337
(Health)
360
(Health)
617
(Health)
335
(Health)
634
(Health)
337
(News)
367
(News)
Topic 3 192
(Lock Down)
308
(Covid Test)
412
(Covid Test)
293
(Covid Test)
418
(Covid Test)
192
(Politics)
199
(Politics)
C Neu Topic 1 973
(News)
812
(News)
995
(News)
1001
(News)
778
(News)
912
(Health)
927
(Health)
Topic 2 931
(Health)
775
(Health)
948
(Health)
977
(Health)
981
(Health)
903
(News)
897
(News)
Topic 3 907
(Lock Down)
707
(Covid Test)
893
(Covid Test)
900
(Covid Test)
705
(Covid Test)
816
(Politics)
878
(Politics)

Next, we sum up each users’ sentiment scores for top-k = 3 trending topics and consider these scores for clustering them. We determine CPos, CNeg, and CNeu clusters following the same procedure of identifying topic-wise clusters. Table 7 represents the sizes of overall clusters at different time windows. Here, from the first time interval to the fifth time-interval, the positive, negative, and neutral clusters’ size change typically. But, in the sixth and seventh time-interval, neutral cluster size increases than usual. From the analysis of this change, top-k = 3 trending topics change that arises with Health, News and Polities topics. The UK’s prime minister and health secretary test positive on 28 March 2020 that refers to the sixth time-interval. We sketch Fig 8 and visualize the changes in the clusters’ size precisely for a graphical representation. The size of these clusters is changing with the shift of time windows. We also notice that the neutral clusters (CNeu) always have the largest sizes among all.

Table 7. Overall sentimental clusters size for top-k = 3 trending topics.

Time Interval → I 1 I 2 I 3 I 4 I 5 I 6 I 7
Clusters ↓
C Pos 987 809 1007 813 918 1073 1103
C Neg 1023 1109 1592 1211 1610 1192 1209
C Neu 1975 1821 1873 1993 1509 1912 1708

Fig 8. Overall sentimental clusters at different time intervals Im.

Fig 8

As a first evaluation measure, we find out the entropy of our mentioned positive, negative, and neutral clusters. These clusters are shown in Table 8. Hence, a good sentimental cluster should have a low entropy value, and here CNeg is 1.401 that depicts the lowest entropy value in the first time interval. The highest value of entropy is 1.584 as CPos in the second time interval that refers to a bad sentimental cluster comparatively. We also see the diversity of entropy values where some values explicate good sentimental clusters, and some define bad sentimental clusters.

Table 8. Entropy of sentimental clusters at different Im.

Clusters → C Pos C Neg C Neu
Time Interval ↓
I 1 1.437 1.401 1.53
I 2 1.584 1.503 1.479
I 3 1.579 1.563 1.58
I 4 1.585 1.512 1.521
I 5 1.548 1.561 1.497
I 6 1.45 1.419 1.545
I 7 1.465 1.402 1.46

Schematic cohesion, which is our second evaluation measure, is represented in Table 9 that leads to clusters’ expertness. Here, we see the most outstanding value of schematic cohesion of Cpos and Cneg are 0.60 (News) and 0.63 (News) respectively in the first time interval. Furthermore, the most economical value of Cpos is 0.39 (Health) in the third time interval. From the observation of this table, we see the heterogeneity among the clusters as values. Here, considering two topics, one is News, and another one is Health.

Table 9. Schematic cohesion of sentimental clusters at different Im.

Time window C pos C neg C neu
I 1
(23/3–25/3)
0.60 (News) 0.63 (News) 0.493 (News)
I 2
(24/3–26/3)
0.402 (News) 0.569 (News) 0.534 (News)
I 3
(25/3–27/3)
0.39 (Health) 0.403 (News) 0.434 (News)
I 4
(26/3–28/3)
0.419 (News) 0.491 (News) 0.502 (News)
I 5
(27/3–29/3)
0.523 (Health) 0.418 (News) 0.518 (Health)
I 6
(28/3–30/3)
0.552 (Health) 0.54 (Health) 0.472 (Health)
I 7
(29/3–31/3)
0.549 (Health) 0.587 (Health) 0.525 (Health)

5 Discussion

Tracking top involved users’ sentiments and sentimental clusters over time is the main objective of this work. Therefore, we conduct these experiments focusing on the topics that have the most trendiness on Twitter at a particular time.

Depending on the unique users’ number and their activities on Twitter about a specific sub-topic related to COVID-19, we identify top-k trending sub-topics. Table 1 holds information regarding trending sub-topics. Table 2 shows how the value of α controls given two parameters for a sub-topics trendiness detection. When we change the value of α, the list of top trending subtopics is changing. It changes by either the topic title or by the serial of topics in the list. Notably, very few users can sustain the top involved list at all time intervals for related trendy topics. In Table 3 topic ’Lockdown’ appears in the top sub-topics list at I1 and then vanishes from the list after that. Other topics may remain on the list at more than a time window, but the percentages of their popularity change over time. This observation becomes clearer when we notice the heatmap in Fig 3. To find out the sentiment from the social stream, we use VADER. It depicts in Fig 9 as an architecture view. Table 4 shows some examples of social streams with sentiment results.

Fig 9. Architecture of VADER.

Fig 9

By using ‘Users’ involvement Detection Algorithm, we bring out top r involved users’ sentiments and analysis over time. Table 5 holds the top 20 involved user’s sentiment scores. Notably, very few users can sustain the top involved list at all time intervals for related trendy topics.

COVID-19 has a particular impact on users’ sentiments. So we intend to focus on the most involved users’ sentiments. With the flow of time, users’ overall sentiments on top COVID-19 topics are changing. In Table 5, we can observe that the change of time window brings changes in the top 20 involved users’ lists and their sentiments. This list in each Im is mixed with negative, positive, and neutral sentiments. Fig 7 has ten specific users’ overall sentiment scores on various time windows represented by a heat map that indicates these changes more specifically and visually.

We also illustrate the sentimental clusters topic-wise and overall. Table 6 shows the topic wise sentimental clusters and Table 7 displays the overall sentimental clusters in each time window. The 3D visualization can help to regulate the behavior of overall sentimental clusters. It is sketched in Fig 8.

Finally, Table 8 exhibits the entropy of clusters at each time window that serves the randomness of a cluster as the reference of entropy value. Table 9 depicts the schematic cohesion at each time window that mirrors the clusters’ expertness.

6 Conclusion

Users’ sentiment for diverse purposes has brought attention to research on social networks. It contains great importance in the COVID-19 pandemic situation. This paper proposed a model to identify users’ sentiment dynamics for top-k trending sub-topics related to COVID-19. It has also detected the top active users based on their involvement score on those trending topics.

This work successfully derives a function to calculates user’s involvement scores towards Query topics and determines the top 20 involved users to analyze their sentiment at the different periods. We accomplish this research with the latest Twitter data and bring out that both users’ involvement and their sentiments vary after a particular time. In the future, besides the determination of active users, we want to develop a methodology to track top negative and positive users by analyzing their sentiments.

Data Availability

The sample data used in this study can be found on figshare at https://figshare.com/articles/dataset/Covid19RelatedTweets/14735972.

Funding Statement

The authors received no specific funding for this work.

References

  • 1.Rufai SR, Bunce C. World leaders’ usage of Twitter in response to the COVID-19 pandemic: a content analysis. Journal of Public Health. 2020Aug18;42(3):510–6. doi: 10.1093/pubmed/fdaa049 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Ravi K, Ravi V. A survey on opinion mining and sentiment analysis: tasks, approaches and applications. Knowledge-based systems. 2015Nov1;89:14–46. doi: 10.1016/j.knosys.2015.06.015 [DOI] [Google Scholar]
  • 3.Barkur G, Vibha GB. Sentiment analysis of nationwide lockdown due to COVID 19 outbreak: Evidence from India. Asian journal of psychiatry. 2020Jun;51:102089. doi: 10.1016/j.ajp.2020.102089 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Ahmed W, Vidal-Alaball J, Downing J, Seguí FL. COVID-19 and the 5G conspiracy theory: social network analysis of Twitter data. Journal of medical internet research. 2020May6;22(5):e19458. doi: 10.2196/19458 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Asgari-Chenaghlu M, Nikzad-Khasmakhi N, Minaee S. Covid-transformer: Detecting trending topics on Twitter using universal sentence encoder. arXiv preprint arXiv:2009.03947. 2020 Sep 8.
  • 6.Aurpa TT, Ahmed MS, and Anwar MM. “Clustering Active Users in Twitter Based on Top-k Trending Topics”. Extended abstract of Complex Network 2020. doi: 10.13140/RG.2.2.19936.30726 [DOI] [Google Scholar]
  • 7.Zhao WX, Jiang J, Weng J, He J, Lim EP, Yan H, et al. Comparing Twitter and traditional media using topic models. InEuropean conference on information retrieval 2011 Apr 18 (pp. 338-349). Springer, Berlin, Heidelberg.
  • 8.Prabhakar Kaila D, Prasad DA. Informational flow on Twitter–Corona virus outbreak–topic modelling approach. International Journal of Advanced Research in Engineering and Technology (IJARET). 2020Mar31;11(3). [Google Scholar]
  • 9.Samuel J, Ali GG, Rahman M, Esawi E, Samuel Y. COVID-19 public sentiment insights and machine learning for tweets classification. Information. 2020Jun;11(6):314. doi: 10.3390/info11060314 [DOI] [Google Scholar]
  • 10.Sharma K, Seo S, Meng C, Rambhatla S, Liu Y. COVID-19 on social media: Analyzing misinformation in Twitter conversations. arXiv e-prints. 2020 Mar:arXiv-2003.12309.
  • 11.Yin H, Yang S, Li J. Detecting topic and sentiment dynamics due to Covid-19 pandemic using social media. InInternational Conference on Advanced Data Mining and Applications 2020 Nov 12 (pp. 610-623). Springer, Cham.
  • 12.Han X, Wang J, Zhang M, Wang X. Using social media to mine and analyze public opinion related to COVID-19 in China. International Journal of Environmental Research and Public Health. 2020Jan;17(8):2788. doi: 10.3390/ijerph17082788 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Hasan A, Moin S, Karim A, Shamshirband S. Machine learning-based sentiment analysis for Twitter accounts. Mathematical and Computational Applications. 2018Mar;23(1):11. doi: 10.3390/mca23010011 [DOI] [Google Scholar]
  • 14.Pandey AC, Rajpoot DS, Saraswat M. Twitter sentiment analysis using hybrid cuckoo search method. Information Processing & Management. 2017Jul1;53(4):764–79. doi: 10.1016/j.ipm.2017.02.004 [DOI] [Google Scholar]
  • 15.Li G, Liu F. A clustering-based approach on sentiment analysis. In2010 IEEE international conference on intelligent systems and knowledge engineering 2010 Nov 15 (pp. 331-337). IEEE.
  • 16.Coletta LF, da Silva NF, Hruschka ER, Hruschka ER. Combining classification and clustering for tweet sentiment analysis. In2014 Brazilian Conference on Intelligent Systems 2014 Oct 18 (pp. 210-215). IEEE.
  • 17.Ahuja S, Dubey G. Clustering and sentiment analysis on Twitter data. In2017 2nd International Conference on Telecommunication and Networks (TEL-NET) 2017 Aug 10 (pp. 1-5). IEEE.
  • 18.Ma B, Yuan H, Wu Y. Exploring performance of clustering methods on document sentiment analysis. Journal of Information Science. 2017Feb;43(1):54–74. [Google Scholar]
  • 19.Feng S, Wang D, Yu G, Gao W, Wong KF. Extracting common emotions from blogs based on fine-grained sentiment clustering. Knowledge and information systems. 2011May;27(2):281–302. doi: 10.1007/s10115-010-0325-9 [DOI] [Google Scholar]
  • 20.Farhadloo M, Rolland E. Multi-class sentiment analysis with clustering and score representation. In2013 IEEE 13th international conference on data mining workshops 2013 Dec 7 (pp. 904-912). IEEE.
  • 21.Anwar MM, Liu C, Li J, Anwar T. Discovering and Tracking Active Online Social Groups. Proceedings of the 18th International Conference on Web Information System Engineering (WISE 2017), pp. 59–74, October 7-11, 2017, Puschino, Russia.
  • 22.Aurpa TT, Ahmed MS, Anwar MM. Online Topical Clusters Detection for Top-k Trending Topics in Twitter. Accepted to IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM) 2020, pp. 573–577, 7-10 December, 2020, Hague, Netherlands.
  • 23.Das BC, Ahmed MS, Anwar MM. Query Oriented Active Community Search. International Joint Conference on Computational Intelligence (IJCCI), pp. 495–505, 14-15 December, 2018, Dhaka, Bangladesh
  • 24.Anwar MM, Liu C, Li J. Discovering and tracking query oriented active online social groups in dynamic information network. World Wide Web, Vol. 22(4) pp. 1–36, August 2018. [Google Scholar]
  • 25.Han B, Cook P, Baldwin T. Lexical normalization for social media text. ACM Transactions on Intelligent Systems and Technology (TIST). 2013Feb1;4(1):1–27. doi: 10.1145/2414425.2414430 [DOI] [Google Scholar]
  • 26.Sasaki K, Yoshikawa T, Furuhashi T. Online topic model for Twitter considering dynamics of user interests and topic trends. InProceedings of the 2014 conference on empirical methods in natural language processing (EMNLP) 2014 Oct (pp. 1977-1985).
  • 27.Hutto, C.J. & Gilbert, E.E. VADER: A Parsimonious Rule-based Model for Sentiment Analysis of Social Media Text. Eighth International Conference on Weblogs and Social Media (ICWSM-14). Ann Arbor, MI, June 2014.

Decision Letter 0

Hocine Cherifi

2 Jun 2021

Detecting Sentiment Dynamics and Clusters of Twitter Users for Trending Topics in COVID-19 Pandemic

PONE-D-21-06706

Dear Dr. Ahmed,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Hocine Cherifi

Academic Editor

PLOS ONE

Journal requirements:

1. We suggest you thoroughly copyedit your manuscript for language usage, spelling, and grammar. If you do not know anyone who can help you do this, you may wish to consider employing a professional scientific editing service.  

Whilst you may use any professional scientific editing service of your choice, PLOS has partnered with both American Journal Experts (AJE) and Editage to provide discounted services to PLOS authors. Both organizations have experience helping authors meet PLOS guidelines and can provide language editing, translation, manuscript formatting, and figure formatting to ensure your manuscript meets our submission guidelines. To take advantage of our partnership with AJE, visit the AJE website (http://learn.aje.com/plos/) for a 15% discount off AJE services. To take advantage of our partnership with Editage, visit the Editage website (www.editage.com) and enter referral code PLOSEDIT for a 15% discount off Editage services.  If the PLOS editorial team finds any language issues in text that either AJE or Editage has edited, the service provider will re-edit the text for free.

Upon resubmission, please provide the following:

  • The name of the colleague or the details of the professional service that edited your manuscript

  • A copy of your manuscript showing your changes by either highlighting them or using track changes (uploaded as a *supporting information* file)

  • A clean copy of the edited manuscript (uploaded as the new *manuscript* file)

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This is a very interesting research that studied the sentiment analysis with the emergent demand of covid related text. This research can help to identify the trending topics and find the related groups. It will help to fight the covid pandemic using data science techniques.

This work clearly presented the motivation, algorithms and validated the solution using the real datasets. The methodology and the experimental design are reasonable to show the result of such research.

There is only one minor comment: It missed the relevant citation - Detecting topic and sentiment dynamics due to Covid-19 pandemic using social media. H Yin, S Yang, J Li. International Conference on Advanced Data Mining and Applications, 610-623.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Acceptance letter

Hocine Cherifi

30 Jul 2021

PONE-D-21-06706

Detecting Sentiment Dynamics and Clusters of Twitter Users for Trending Topics in COVID-19 Pandemic

Dear Dr. Ahmed:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Professor Hocine Cherifi

Academic Editor

PLOS ONE


Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES