Abstract
Text Classification problem has been thoroughly studied in information retrieval problems and data mining tasks. It is beneficial in multiple tasks including medical diagnose health and care department, targeted marketing, entertainment industry, and group filtering processes. A recent innovation in both data mining and natural language processing gained the attention of researchers from all over the world to develop automated systems for text classification. NLP allows categorizing documents containing different texts. A huge amount of data is generated on social media sites through social media users. Three datasets have been used for experimental purposes including the COVID-19 fake news dataset, COVID-19 English tweet dataset, and extremist-non-extremist dataset which contain news blogs, posts, and tweets related to coronavirus and hate speech. Transfer learning approaches do not experiment on COVID-19 fake news and extremist-non-extremist datasets. Therefore, the proposed work applied transfer learning classification models on both these datasets to check the performance of transfer learning models. Models are trained and evaluated on the accuracy, precision, recall, and F1-score. Heat maps are also generated for every model. In the end, future directions are proposed.
1. Introduction
Natural language processing is a scientific process to train a computer to understand and process human language. NLP gained a lot of importance in recent years because of the researchers and processing powers of machines. Researchers are doing their best to generate interesting facts and figures from human language and implement those results in every field of life from educations to hospitals, industry to shopping malls, etc. In past, NLP problems were solved using rule-based systems. However, due to the different nature of text in the world, machine learning is applied to NLP and it has gained a strong ground using SVM and Naïve Bayes. Natural language processing and text mining refer to the process of human-generated text that came from multiple social media networks using different algorithms, programs, and techniques. It is an important field of AI. With continued research on text mining and NLP using data mining algorithms, machine learning, and deep learning, data mining techniques have gained the best results in the fields of automatic question answering machines, anaphora resolution, automatic abstraction, bioinformatics, and web relation network analysis [1]. Researches show that NLP, data mining, and text classification can be very helpful in every prospect of life. There are also many other researchers who have used NLP in hate speech, sentiment analysis [2], detection of controversial Urdu speeches [3], movie reviews [4], stock market [5], online reviews [6], and restaurant reviews [7].
In recent decades, social media has gained huge importance because of its usage for different purposes. If people use social media often, then it is obvious they will generate a huge amount of data. Because of this huge data generated by social media users, hate speech is also increased. For example, if a movie is released, the audience will have good or bad or neutral reviews or comments about it. Researchers had also done plenty of work in the area of hate speech as well and it is increasing day by day. The paper [8] had explained how NLP is involved in hate speech tasks and how it is able to automate the process to capture and detect hatred social media content. These researches involve NLP as they are using human-generated natural content. Social media content generated by social media platform users is an important source of data for hospitals, industry, scientists, policy-making, and much more. UGC (User Generated Content) on different review platforms or sites holds diverse information in the form of text that is extracted after applying opinion extraction algorithms and sentiment analysis techniques [9] These algorithms provide better performance in the feature extraction phase of text classification as well [10–24].
A group of researchers had worked and highlighted the limitations and gaps in the field of hate speech [25]. A solution to reduce these limitations was also proposed. They had elaborated that a large amount of sufficient data to train an automated approach. Insufficient labeled data related to hate speech is a big problem in the detection of hate speeches on social media. Their proposed approach was pretrained on BERT. One of the important tasks in hate speech detection is to categorize portions of text based on their context and make developers capable of text classification tasks in NLP [26]. Their trained model on the Italian hate speech dataset is named as ALBERTo. This model is highly sensitive regarding the temporal distances of datasets. But its main advantage is as follows: after a time, its performance increases, and it required less training data than previous classifiers. Hate speech and offensive languages are two different things. Separating hate speech from offensive language is a difficult task [27]. Their research uses a crowdsourced lexicon of hate speech to collect tweets and then label them as offensive, hateful, and neutral metaheuristic algorithms which are used for text and data classification tasks [80, 82, 90].
Transfer learning is a phenomenon or task in which the information gained from unlabeled data can be used in relative tasks with a small labeled dataset. And that small labeled dataset achieves high accuracy with the help of previous information. NLP transformers have gained promising accuracy in every practice as compared to ML and DL techniques. They have written in their research that the key idea behind TL is to grab information from related areas to help systems based on machine learning to obtain higher accuracy in the area of interest. Thus, we can also say that transfer learning can also be used to achieve high performance with less human supervision as compared to active learning and supervised learning. There are many examples in our real life which we can relate to transfer learning. For example, if a system is already trained to recognize apples, then it may be also used with little fine-tuning to recognize pear as well. This will need less data and less training time. The key idea behind transformers is attention.
Social media has a great empowerment impact. Every user can post based on their thoughts. They trained already existing models to predict the posts or news related to coronavirus as real or fake. Among all trained models, transformers showed the best results. News is a great source and holds great importance as they keep everybody updated. Fake news took birth in the 18th century [28]. The Internet makes it easier to spread fake news through the excessive use of social media. It is also very tough to distinguish between real and fake news. Already existing approaches have deficiencies which they tried to overcome by developing a hybrid approach.
The rest of the paper is organized as follows. Section 2 discusses the literature review. Section 3 discusses the types of classification algorithms. Section 4 overviews methodology. Experimental results are discussed in Section 5. The conclusion and future prospects of our work are discussed in Section 6.
2. Literature Review
Transfer learning nowadays holds spectacular importance in the research area. Researchers are trying hard to achieve higher accuracies in every research by applying different versions of transformers. A team of researchers [29] performed a comprehensive survey of sentiment analysis in finance in which they have evaluated recent researches and advancements regarding finance. They have evaluated techniques including lexicon-based approaches for text classification, statistical methods, sentence encoders, word encoders, and transformers. Their evaluation of the finance dataset clearly shows that transformers outperformed among all existing methods and techniques with the highest accuracy. They have applied different models of transformers including BERT, FinBERT, XLNet, XLM, ALBERT, RoBERTa, DistlBERT, XLM-RoBERTa, and BART. All these transformers gained a high F1 score among all. Between these NLP transformers, BART launched by Facebook achieved the highest f1 score of 0.85.
Researchers performed an evaluation of sentiment analysis approaches based on transfer learning for the Japanese dataset [30]. They have performed binary sentiment classification and multisentiment classification on product reviews and movie reviews. After their research, they have stated that transfer learning approaches perform way better than models that are generated for task-specific purposes on 3 times greater data. They have stated that better systems exist for the English language but there is much deficiency for the Japanese data. So, they tried 3 transfer learning models including BERT, ELMo, and ULMFit. All these models have achieved less error percentages compared to other models using datasets including the Rakuten dataset and Yahoo movie review dataset. BERT-base gained the lowest rate of 8.42 on the Yahoo movie review dataset and 4.68 on the Rakuten dataset.
These researchers performed a study on the most recent advancements of transfer learning in the field of natural language processing [31]. Firstly, they have checked recent machine learning and deep learning approaches and then also checked recent TL approaches. They have noticed that transfer learning approaches have brought new dimensions for different NLP tasks. Transfer learning can be happily and effectively used in the areas where we have less data to train. We can use a pretrain model and then fine-tune it. They have experienced that transfer learning models can perform better than other state-of-the-art methods in NLP. BERT is trained on BookCorpus, text corpus, and Wikipedia which can give overwhelming results in some areas of natural language processing but it still needs to be improved [32]. It somewhere lacks domain-related and task-related knowledge. It is where improvement is required. They have presented a new version of BERT called BERT4TC BERT for text classification. Their model is rich in sense of domain- and task-related knowledge. They have evaluated their proposed model on publicly available datasets. Results showed that the model they have proposed with compatible auxiliary sentences outperforms compared to both feature-based typical methods and some fine-tuned methods and achieved new state-of-the-art results in multiclass classification.
This research proposes a political sentence-level text classifier using human experts' annotated corpus for political manifestos [33] and then applied to press briefings of COVID-19. They have manually annotated the manifestos as training data on a classifier and then applied that to press briefings to automatically classify existing sentences in press briefings. They have combined CNN with BERT transformer, and it showed that CNN combined with BERT gained the highest accuracy among other models compared with CNN. They have done four experiments named M1, M2, M3, and M4. M4 performed better among all as it is CNN + BERT. It contained high accuracy and an F1 score. Fine-tuning of desired pretrained models is an efficient transfer mechanism. However, fine-tuning may be inefficient in some tasks and need to build entire new techniques for solving multinature problems [34]. As an alternative to it, they have proposed an adapter module with the transfer. These modules generate an extensible model. We only need to add a few parameters which are trainable on every task, and we can add a new task without revision of the previous one. Parameters that are from the original model remain fixed with high parameter sharing. They have evaluated BERT on 26 different classification tasks. And they have used GLUE as a benchmark. GLUE achieved high performance with full fine-tuning of parameters by adding only 3.6% parameters per task. Fine-tuning trains 100% of the parameters.
Evaluation of deep learning approaches and transfer learning approaches for fake news detection using COVID-19 fake news detection dataset (consisting of 10,700 social posts and articles) was performed by [35]. They used classification algorithms bi-LSTM + Attention, HAN (hierarchal attention network) BERT-base, and DistilBERT. Their aim is to classify the news as fake or real. The fake news detection task is formulated as a text classification problem. They rely on the content of the news and ignore other important features like user characteristics, social circle, etc. which might not always be available. The BERT and DistilBERT models pretrained on the COVID-19 tweets corpus perform better than the ones which are only fine-tuned on the dataset. The BERT-cased model which was trained manually on the COVID-19 tweets corpus gives the best results followed by the COVID-Twitter-BERT model. Reference [36] elaborated the impact of social media in our daily lives. They also highlighted the misleading information on social media and its effect on our lives. They proposed an approach to detect the fake and real news about COVID-19. The model achieves high F1 score and occupied the second position on the leaderboard. They used the dataset containing posts and tweets collected from Facebook, Twitter, and Instagram. They have split the dataset into train test and validation parts. They tried different baseline models on this dataset and also used different transformer models. And results clearly show that their RoBERTa model achieves a 0.9864 F1-score and their Electra model achieves a 0.9827 F1-score on the official test set.
This research highlights the impact of fake news related to coronavirus [37]. They stated that most social media posts are not trustworthy as they lead the readers toward wrong information that can cause panic situations among people. They presented their results on COVID-19 Fake News Detection in English and achieved the first position in the leaderboard among 166 submitted results. Their proposed model uses CT-BERT (COVID-Twitter-BERT) and achieves a 98.69 F1-score. Their research developed a method to check the reliability of social media posts that belong to COVID-19 [38]. They ensemble three transfer learning models (BERT, ALBERT, and XLNET) for classifying COVID-19 news into real and fake. They have used the COVID-19 Fake News Detection in English dataset. Their proposed methodology achieves a 0.9855 F1-score on the test set and among 160 teams getting the 5th rank. They split the dataset into training, validation, and testing parts for the experimental setup.
A multimodal approach for fake news detection was developed by [39]. Because in past years, posting wrong, hateful, abusive, offensive, and hateful content on social media tools has increased in exponential format, people spread their inner negativity related to any situation on social media. This may lead other people toward a wrong and hateful path. That is why it is the need of the hour to detect those profiles and people who do this sinful act. Researchers used different strategies to accomplish this purpose. The authors of this research work propose a multimodal approach based on multi-image. In specific, their system uses textual, semantic, and visual data or information. They had used BERT for textual data to extract the semantic and contextual meaning of the text. They further used the VGG-16 model for visual representation and tag extraction. And the rest of the semantic information is calculated using cosine similarity. They had used GossipCop, a part of the FakeNewNet dataset. This multimodal multi-image approach achieved a 0.7955 F1 score on testing. This approach had also increased the performance of baseline models.
With the rapid growth of social media in past years, it has become more convenient for people to access news fast than ever. They said that it is also happening that people are spreading fake news over social platforms for their own purposes. Many researches using supervised learning had been proposed to detect fake news. These approaches focus on different features to make the classification more accurate like news content, social context, user profile, and messages context. These approaches showed accuracy but face limitations as they need a reliable accurate dataset [40]. Their proposed work was an unsupervised framework called the Unsupervised Fake News Detection Technique (UFD) to minimize this problem. They mainly focused on two aspects: user's reliability and truth of news to filter the fake news among real news. They had tested their framework on datasets which are LIAR containing 12,800 short news statements, and BuzzFeed has 1,627 news articles related to U.S. elections. But they used 332 and 144 after filtering datasets. UFD achieved the highest accuracy, precision, recall, and F1-score on both datasets. The model achieved 0.759 and 0.679 of accuracy on LIAR and BuzzFeed, respectively.
Nowadays, society is more and more connected and attracted to the Internet. People around the globe make it a necessary part of their lives. Information we retrieve and gather from Internet has become an essential part of our lives [41]. They had described that this extra dependence on Internet has led us to its wrong impacts as well, as it is leading us toward hatred, abusive, offensive, and toxic language. Machine learning is doing great in the field of NLP. They had developed the DeepHate model for text analysis which is trained on several small datasets to make it more accurate. Their model can learn a single hate speech pattern from unrelated and diverse data sources. The model works on transfer learning and can generate both word representation and sentence representation. They used an English tweets dataset containing 37,520 tweets. Another dataset is also used containing 22,304 tweets including offensive, hateful and harmless tweets.
3. Types of Classification Algorithm
This research work uses machine learning and transfer learning classification algorithms. These models are applicable to many natural processing tasks and work efficiently on these tasks. The following mentioned models are used in this research procedure.
3.1. BERT-Base
Bidirectional Encoder Representation from Transformers (BERT) was proposed by [42]. The main purpose of BERT is to train bidirectional representations from an unlabeled dataset. It works on collaborative left and right context phenomena in all layers. BERT is simple yet powerful. It generates promising results in several machine learning tasks. A fine-tune model of BERT only needs to add one more layer for each new model to perform a variety of tasks. It uses a masked language model. MLM works on the phenomena of masking random words from input and then it predicts the ID of that word by utilizing its context. MLM uses both left and right contexts which enables training of the bidirectional model. They joint MLM with next sentence prediction (NSP) as well. BERT-base is comparatively smaller in its size, it takes less time for computation and processing, and also it is affordable. It is not applicable to ambiguous data mining or text mining tasks. Reference [43] used it in the detection of fake news. The paper [44] used it for content enhancement and it proves its promising results in content enhancement field. It was also used by [45] for distilling its knowledge. Reference [46] performed sentiment analysis using BERT and it has done a great job there as well.
3.2. BERT-Large
BERT-large is a type of BERT model. It works similarly as BERT-base does but it has a larger size than BERT-base. It is more expensive than BERT-base as it takes more time for computation and is applicable to large datasets. The article [47] used BERT-large in his research work to process the COVID-19 related content on Twitter. BERT-large performed well on his dataset but his proposed approach performed better. The paper [48] used BERT-large in offensive tweet classification, and among all evaluated approaches, BERT-large stands the second on a scale with a 0.781 F1 score. The authors of [49] performed multiple experiments on deep learning and transfer learning approaches to access syntactic abilities and they have seen that between all approaches BERT-based transformers performed extremely well.
3.3. RoBERTa-Base
The authors of [50] proposed RoBERTa model with slight advancements in BERT which are as follows: training their model with more data and larger batch size, eliminating the next sentence prediction factor, having larger sequences, and making changes in masking pattern. Their proposed model performs well in many experimental setups. They have also noticed that the linguistic bias of RoBERTa-base is stronger. Roberta uses BookCorpus, OpenWebText, English Wikipedia, STORIES, and CC-News. The authors of [51] did research on learning features that are also important. They explained that RoBERTa obtains linguistic generalization as preferences. Reference [52] compared three methods including LSTM, BERT, and RoBERTa for detecting and classifying mental illness on multiple social media platforms. And RoBERTa outperforms among these three approaches. Reference [53] used RoBERTa to classify informative tweets related to COVID-19 and their approach showed the best results.
3.4. RoBERTa-Large
Reference [54] applied RoBERTa-large with dialog history attention to select the responses based on a randomly wired network. Research has shown that the RoBERTa-large model needs more computer resources than RoBERTa-base. That is why it is not widely used by researchers. The article [55] used RoBERTa to highlight and detect medications on Twitter. They used an unbalanced dataset and their proposed model achieved a 0.8 F1 score. The paper [56] used RoBERTa for a Dutch language model. Their experimentation showed that training a BERT model on the Dutch language shows a lot of variety in multiple tasks for the Dutch language. The authors of [57] used RoBERTa-large for eye-tracking prediction. And their technique showed promising results with a 3.929 MAE score and stands in 3rd position among 13 teams.
3.5. DistilBERT
DistilBERT was introduced in 2019 by [58]. It was a lighter, fast, smaller, and cheap version of BERT with a size reduction of 40% with 60% more speed and 97% understanding of language capabilities. This lighter and useful version was used by many researchers. The authors of [59] used this lighter version of BERT for sociopolitical news classification. DistilBERT showed promising results in their experiments. The authors of [60] combined linguistic knowledge with different transfer learning models to enhance their performance. And their methodology worked really well in this perspective. Ensemble models boosted the performance of used models by many points. The authors of [61] used this version of BERT for detecting health information along with named entity recognition tasks. And the detection was improved by half which was promising. The authors of [62] worked with DistilBERT and proposed a mechanism for answer selection and picking up important words. The performance was improved by 0.6% which is not bad at all. The authors of [63] retrained DistilBERT on universal dependencies for the purpose of a voice shopping assistant. The performance of these downstream tasks is raised by 1.31%.
3.6. ALBERT-Base-v2
With the collaboration of Toyota Technologies and Google Research, they jointly released the scalable and smaller successor of BERT in 2019 which they named ALBERT [64]. It mainly involved reduction in two parameters: increase in training speed of BERT and lower memory consumption. ALBERT performs better in multiple classification tasks. It also uses a very low number of parameters while doing sentiment analysis. The authors of [65] used the ALBERT transformer approach for contextualized sarcasm-based detection on Twitter. They have applied other transformer approaches as well. The authors of [66] also evaluated this model for fake news detection and additionally checked the facts for these fake news which worked really well and lead among all models. The authors of [67] checked this approach for question answers on COVID-19. ALBERT gained the highest exact match score of 13.04. The authors of [68] used ALBERT for medication prescriptions used on social media.
3.7. XLM- RoBERTa-Base
The authors of [50] proposed this model and trained it on hundred languages with two TB of data which was filtered. Their model which was combined with XLM-R outperforms with 23% accuracy compared to many transformers. The authors of [69] identified offensive language using this ensemble technique. They joint XLM-RoBERTa with DPCNN, and this model showed amazing results. These two also used this approach for hope speech detection attention, and this shows promising results in this task. They achieved 0.59, 0.84, and 0.92 F1 scores for Tamil, Malayalam, and English languages. The authors of [70] used this and experiment with it for multilanguage sentiment analysis. And the model achieved a good F1 score. The authors of [71] performed another research work of classification using neurons for the task at EVALITA 2020. They had used the hate speech dataset and performed the experiment. The model achieved a 0.798 F1 score. The authors of [72] used XLM-RoBERTa for context disambiguation in words. The model outperforms all experimented methodologies.
3.8. Electra-Small
BERT uses Masked Language Modeling and replaces some tokens with masks and then reconstructs the model using these masks. But this requires a large amount to compute, so the authors of [73] proposed Electra and overcome this issue. Their proposed approach replaces tokens with alternative samples. And after that, they did not train the model; they made sure that each token in input is swapped with a sample generator or not. The authors of [74] used Electra for profiling fake news. They have created an ensemble model considering 15 models. Then, they are fined tuned according to the tasks and dataset. Electra achieved 0.70 and 0.69 F1 scores for English and Spanish datasets, respectively. Electra was also trained with multiword selection by [75].
3.9. BART-Large
Facebook researchers in October 2019 proposed BART [76]. The formation of BART is similar to BERT and GPT2. Tasks like question-answer and summarization of any text are done accurately. This model showed promising results in these kinds of tasks. This one takes advantage that the encoder and decoder form BERT and GPT AR, respectively. It considers the autoregressive techniques to check dependencies which makes it better than BERT. Its encoders and decoders are connected. They used BART for an automated speech recognition system. They had done the experimentation for 1000 hours on the speech recognition dataset and they have reduced the error rate to 21.7% which is a huge success and way better than the baseline model. The authors of [77] used it for supervised topic label generation. Their model performs better than the baseline model. The authors of [78] also evaluated this model for query suggestions. Their proposed approach has a better understanding of noise and can handle and understand complex queries. The authors of [79] performed visual common sense generation and called it Knowledge Enhanced Multimodal BART. The authors of [80] evaluated BART for knowledge grounded conversation tasks and achieved good results.
4. Methodology
Due to the complex nature of social media data on COVID-19 fake news and hate speech, it is quite obvious that the proposed model must have different aspects to precisely and accurately predict the fake and real news and similarly hateful or nonhateful content. Figure 1 elaborates the steps and architecture of the fine-tuned model.
4.1. Data Sets
Two datasets are used in this research work. The first is named “COVID-19 fake news dataset” which was originally generated in 2020 by Sumit Bank and was published and made freely available on Coronavirus Disease Research Community-COVID-19. It contains 10202 fake news related to coronavirus which different users shared on social media sites. Some of them are gathered from Facebook, some belong to Instagram, and others belong to websites and Twitter blogs. All of them are collected using different keywords including COVID-19, pandemic, corona, and coronavirus. This dataset is basically assembled in two columns. The first contains text, special characters, and attributes which is named as Title, and the second contains binary values as 0 and 1 and named Outcome. Here, 1 presents real news while 0 presents fake news.
The second dataset which is used in this research work is named “extremist-non-extremist dataset” which was developed by [81]. The dataset was generated using Twitter streaming API, and tweets containing more than one extremist word like ISIS, suicide, bomb, etc. are collected. Each review is compared with seed words present in a manually built extremist lexicon and added to the dataset. The final data was stored in a .csv file. The dataset consists of 21,186 tweets in total, of which 12,755 are labeled as extremist and 8,432 are labeled as nonextremist. Extremist tweets are replaced with 1 and nonextremist tweets are replaced with 0.
The third dataset which is used in this research work is named “COVID-19 English tweets” developed by [82]. A research had revealed that data which is shared on social media sites is uninformative. Therefore, they thought that informative data should be highlighted through a shared automated task, where all the participants have to use their developed dataset of COVID-19 English tweets. They collected tweets using Twitter API with ten keywords including “coronavirus”, “covid-19,” “covid_19,” “covid-2019,” “covid19,” “covid2019,” “covid_2019,” “coronaVirusUpdate,” “coronavid19,” and “SARS-CoV-2.” Every tweet in the corpus contains a minimum of one word from the above-mentioned keywords. They collected tweets of four months from March 2020 to June 2020. Then, they applied different filters to tweets like removing tweets containing 9 words and also removing the tweet of a person who has less than 5 followers and removing tweets that are retweeted. They have also labeled them as Informative and Uninformative. Informative tweets must contain suspects, death, affected cases, recovered cases, and a number of tests, etc. Train file contains 4820 tweets, test file contains 1539 tweets, and validation file contains 566 tweets.
4.2. Data Preprocessing
COVID-19 fake news, COVID-19 English tweets, and extremist-non-extremist datasets which are used in the proposed research work need to be cleaned in the very first step of natural language processing which is called preprocessing step. In this step, cleaning methods on both datasets are applied to remove URLs, converting every word to lower case, and lemmatization and punctuation removal are performed. These methods will eliminate special characters, hyperlinks, empty spaces, identifiers, and words that are very short. This step will clean both datasets.
4.3. Encoding
Many efficient and automatic learning models do not accept input in text form. Therefore, the text is converted into digital vectors which rely on the technique of bag-of-words for transfer learning models. We have counted every word's score, and then, feature extraction was also performed. This research work uses the best, efficient, and appropriate classification algorithms based on results from literature reviews and then builds our model accordingly.
4.4. Model Evaluation and Testing
After the training model was evaluated and its performance is measured using different parameters including confusion matrix, accuracy, recall, precision, and F1-score, we have tested our model on both datasets which are unclassified on fake news, COVID-19 tweets, and hate speech.
Performance evaluation is performed using four parameters which are precision, recall, F1-score, and accuracy. Confusion matrix and heat maps are also generated for evaluation purposes. Precision is known as positive values which are gained from prediction. So, it is the fraction of relevant occurrences among gained occurrences. On the other hand, recall is called sensitivity; it is the relevance of gained occurrences. The weighted average of precision and recall is defined as F1-score. It takes false positives and false negatives into account. The total number of rightly predicted values is called accuracy. Performance of any classification model is measured using the N × N matrix which is known as the confusion matrix. It contains true positive, true negative, false positive, and false negative values in the matrix which is used to evaluate the actual values with the values predicted by the classifier. Heat maps are used to observe the data through visualization. It presents different attributes. Visualization helps to find patterns and also gives a perspective of depth. So, heat map is used to explore and observe the data.
This research work uses nine classification models. These classifiers are BERT-base, BERT-large, RoBERTa-base, RoBERTa-large, DistilBERT, ALBERT-base-v2, XLM-RoBERTa-base, Electra-small, and BART-large.
Tables 1–3 present sample tweets for all three datasets. Tables 4–8 show results of nine transfer learning models which are validated using the above-mentioned performance metrics named precision, recall, F1-score, and accuracy on COVID-19 fake news dataset, COVID-19 English tweet dataset, and extremist-non-extremist dataset, respectively. Results clearly show that transfer learning classification models outshine using test datasets obtained from reliable sources. Tables 5, 7, and 9 present a comparative analysis for all three datasets with state-of-the-art approaches. Figure 1 presents the research methodology for the proposed research work. Figures 2–4 show accuracies graph of TL classifiers for COVID-19 fake news, COVID-19 English tweet, and extremist-non-extremist dataset Figures 5–13 present the heat maps for COVID-19 fake news dataset. While Figures 14–22 present the heat maps for transfer learning classifiers and Figures 23–31 show heat maps for extremist-non-extremist dataset.
Table 1.
Tweets | Labels |
---|---|
A chain lists recommendations to prevent and treat coronavirus | 0 |
Australia closing borders in a few hours for 6 months | 1 |
Table 2.
Tweets | Labels |
---|---|
Oh Allah, we are helpless | Nonextremist |
Oh Allah, destroy US and Israel | Extremist |
Table 3.
Tweets | Labels |
---|---|
Bill Maher says coronavirus “overreactions” making him “sick:” “People die! That's what happens in life! | Uninformative |
#Australia Melbourne GP clinic closed after doctor tests positive for #coronavirus after seeing 70 patients this month | Informative |
Table 4.
Model | Accuracy | Precision | Recall | F1 score |
---|---|---|---|---|
BERT-base | 99.56 | 97.21 | 97.77 | 97.53 |
BERT-large | 99.31 | 99.07 | 93.13 | 95.89 |
RoBERTa-base | 99.71 | 99.85 | 96.84 | 98.29 |
RoBERTa-large | 99.66 | 98.78 | 97.32 | 98.04 |
DistilBERT | 99.41 | 96.69 | 96.69 | 96.69 |
ALBERT-base-v2 | 98.68 | 90.83 | 95.30 | 92.94 |
XLM-RoBERTa-base | 99.22 | 96.01 | 95.08 | 95.54 |
Electra-small | 99.17 | 96.85 | 93.56 | 95.14 |
BART-large | 99.31 | 99.07 | 93.13 | 95.89 |
Table 5.
Table 6.
Model | Accuracy | Precision | Recall | F1 score |
---|---|---|---|---|
BERT-base | 98.44 | 98.42 | 98.45 | 98.43 |
BERT-large | 98.44 | 98.45 | 98.42 | 98.43 |
RoBERTa-base | 96.48 | 96.48 | 96.62 | 96.48 |
RoBERTa-large | 97.00 | 96.97 | 97.12 | 97.00 |
DistilBERT | 98.31 | 98.39 | 98.23 | 98.30 |
ALBERT-base-v2 | 97.78 | 97.75 | 97.83 | 97.78 |
XLM-RoBERTa-base | 98.57 | 98.53 | 98.61 | 98.56 |
Electra-small | 94.52 | 94.66 | 94.76 | 94.52 |
BART-large | 98.83 | 98.85 | 98.80 | 98.82 |
Table 7.
Table 8.
Model | Accuracy | Precision | Recall | F1 score |
---|---|---|---|---|
BERT-base | 99.71 | 98.82 | 97.84 | 98.33 |
BERT-large | 99.71 | 98.82 | 97.84 | 98.33 |
RoBERTa-base | 99.66 | 99.29 | 96.82 | 98.02 |
RoBERTa-large | 99.36 | 98.56 | 94.16 | 96.24 |
DistilBERT | 99.51 | 96.80 | 97.74 | 97.27 |
ALBERT-base-v2 | 98.97 | 94.80 | 93.45 | 94.12 |
XLM-RoBERTa-base | 99.56 | 99.77 | 95.26 | 97.40 |
Electra-small | 98.73 | 97.42 | 87.82 | 92.02 |
BART-large | 99.56 | 98.22 | 96.77 | 97.48 |
Table 9.
5. Experimental Results
5.1. Results for COVID-19 Fake News Data
5.1.1. Discussion
The above-mentioned diagrams and tables contain results for transfer learning classifiers for the COVID-19 fake news dataset. These nine transfer learning classifiers showed excellent performance on the COVID-19 fake news dataset. These classifiers are evaluated using different values of precision, recall, accuracy, and F1-score. Transfer learning classification models performed really well and achieved the highest accuracies. Among all transfer learning models, the RoBERTa-base model achieved the highest accuracy of 99.71%. The RoBERTa-large gained the second position and BERT-base achieved the third position among all transfer learning models with 99.68% and 99.56% of accuracy. DistilBERT, BERT-large, BART-large, XLM-RoBERTa, Electra-small, and ALBERT-base-v2 achieved 99.41%, 99.31%, 99.31%, 99.22%, 99.17%, and 98.68% of accuracies, respectively. So, RoBERTa-base leads all seventeen classification models.
5.1.2. Comparative Analysis with State-of-the-Art Approaches
In Table 5, the proposed work is compared with state-of-the-art approaches [83–86] w.r.t to the text classification task. State-of-the-art approaches use machine learning and deep learning techniques including XGBoost, Naïve Bayes, deep neural network, and T1-convolutional neural network. The objective and purpose of experimentation are to perform fake news text classification using state-of-the-art techniques and transfer learning-based proposed fine-tuned approaches. The performance of experimented approaches is compared in terms of accuracy on COVID-19 fake news dataset. XGBoost exhibits the lowest accuracy of 75% on the COVID-19 fake news dataset. The proposed approach achieved the highest accuracy of 99.66% using RoBERTa-base.
5.2. Results for COVID-19 English Tweets Dataset
5.2.1. Discussion
Transfer learning models work outstandingly on all experimented datasets and surprisingly take less time.TL models performed best on the COVID-19 tweet dataset. Bart-large won this time with a 98.83% accuracy score. It also performed well in terms of precision, recall, and F1-score as it has achieved 98.85, 98.80, and 98.82 respectively. XLM-RoBERTa stands in the second position with a 98.57% accuracy score. BERT-base and BERT-large gained the third position with a 98.44% accuracy score. They have achieved 98.42 precision, 98.45 recall, and 98.43 F1-score. DistilBERT, ALBERT-base-v2, RoBERTa-large, RoBERTa-base, and Electra-small achieved 98.31, 97.78, 97.00, 96.48, and 94.52 accuracy score, respectively. So, transfer learning models performed outstandingly on all three experimented datasets.
5.2.2. Comparative Analysis with State-of-the-Art Approaches
In Table 7, the proposed work is compared with state-of-the-art approaches [87, 88] with respect to the text classification task. State-of-the-art approaches use machine learning techniques including multilayer perceptron and support vector machine. The objective and purpose of experimentation are to perform tweets classification using state-of-the-art techniques and transfer learning-based proposed fine-tuned approaches. The performance of experimented approaches is compared in terms of accuracy on the COVID-19 English tweet dataset. MLP exhibits the lowest accuracy of 0.78% on the COVID-19 English tweet dataset. The proposed approach achieved the highest accuracy of 98.83% using BART-large.
5.3. Results for Extremist-Non-Extremist Dataset
5.3.1. Discussion
All these figures and tables clearly show the performance of transfer learning classifiers on the extremist-non-extremist dataset. Evaluation metrics consist of precision, recall, F1-score, and accuracy. Transfer learning classifiers overshine with the highest accuracies. BERT-base and BERT-large shine among other transfer learning models. Both classifiers gained a 99.71% accuracy score. RoBERTa-based gained a 99.6% accuracy score, XLM-RoBERTa and BART-large gained a 99.56% accuracy score, DistilBERT gained a 99.51% accuracy score, RoBERTa-large gained a 99.36% accuracy score, ALBERT-base-v2 achieved a 98.97% accuracy score, and Electra-small gained a 98.73% accuracy score. From all the above-mentioned results, BERT-base and BERT-large both outshine among all nine text classification models.
5.3.2. Comparative Analysis with State-of-the-Art Approaches
In Table 9, the proposed work is compared with state-of-the-art approaches [89–92] with respect to the text classification task. State-of-the-art approaches use machine learning and deep learning-based techniques including Naïve Bayes, CNN, and HDLTex. The objective and purpose of experimentation are to perform tweets classification using state-of-the-art techniques and transfer learning-based proposed fine-tuned approaches. The performance of experimented approaches is compared in terms of accuracy on the extremist-non-extremist dataset. HDLTex achieved the lowest accuracy of 76.5% on the extremist-non-extremist dataset. The proposed approach achieved the highest accuracy of 99.71% using BERT-large and BERT-base.
6. Conclusion
In this research, nine transfer learning models which are BERT-base, BERT-large, RoBERTa-base, RoBERTa-large, DistilBERT, XLM-RoBERTa-base, ALBERT-base-v2, Electra-small, and BART-large are applied on COVID-19 fake news dataset, COVID-19 English tweet dataset, and extremist-non-extremist dataset for binary text classification. The experimentation is performed on these datasets which are taken from reliable repositories. All transfer learning models are evaluated using evaluation metrics: accuracy, precision, recall, and F1-score.
In the future, we aim to do experiments on large and more datasets with multiclass classification. We can also use different language datasets to perform text classification. It would be valuable to include emoticons as they are widely used in social media to represent expressions. Also, we will try to use the Twitter streaming API to retrieve tweets in real time in order to do a real-time sentiment analysis and explore other social networks.
Data Availability
The data are not available until the thesis defense. In case of any queries, the readers can contact the corresponding author.
Disclosure
This work is part of the Ph.D. thesis of the student.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
References
- 1.Zeng Z., Shi H., Wu Y., Hong Z. Survey of natural language processing techniques in Bioinformatics. Computational and Mathematical Methods in Medicine . 2015;2015:10. doi: 10.1155/2015/674296.674296 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Jefferson C., Liu H., Cocea M. Fuzzy approach for sentiment analysis. Proceedings of the 2017 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE); July 2017; Naples, Italy. [DOI] [Google Scholar]
- 3.Mustafa R. U., Saqib Nawaz P. M., viger F. Early detection of controversial Urdu speeches from social media. Data Sci. Pattern Recognit . 2017;1(2):26–42. [Google Scholar]
- 4.Kumar K., Harish B. S., Darshan H. K. Sentiment analysis on IMDb movie reviews using hybrid feature extraction method. International Journal of Interactive Multimedia and Artificial Intelligence . 2019;5(5):p. 109. doi: 10.9781/ijimai.2018.12.005. [DOI] [Google Scholar]
- 5.Wang Y. Stock market forecasting with financial micro-blog based on sentiment and time series analysis. Journal of Shanghai Jiaotong University . 2017;22(2):173–179. doi: 10.1007/s12204-017-1818-4. [DOI] [Google Scholar]
- 6.Xia H., Yang Y., Pan X., Zhang Z., An W. Sentiment analysis for online reviews using conditional random fields and support vector machines. Electronic Commerce Research . 2020;20(2):343–360. doi: 10.1007/s10660-019-09354-7. [DOI] [Google Scholar]
- 7.Krishna A., Akhilesh V., Aich A., Hegde C. Sentiment Analysis of Restaurant Reviews Using Machine Learning Techniques . Vol. 545. Singapore: Springer; 2019. [Google Scholar]
- 8.Schmidt A., Wiegand M. A survey on hate speech detection using natural language processing. Proceedings of the Fifth International Workshop on Natural Language Processing for Social Media; 2017; German. p. 10. [DOI] [Google Scholar]
- 9.Kirilenko A. P., Stepchenkova S. O., Kim H., Li X. Automated sentiment analysis in tourism: comparison of approaches. Journal of Travel Research . 2018;57(8):1012–1025. doi: 10.1177/0047287517729757. [DOI] [Google Scholar]
- 10.Bangyal W. H., Ahmad J., Tayyab H., Pervaiz S. An improved bat algorithm based on novel initialization technique for global optimization problem. International Journal of Advanced Computer Science and Applications . 2018;9(7):158–166. doi: 10.14569/IJACSA.2018.090723. [DOI] [Google Scholar]
- 11.Bangyal W. H., Ahmad J., Rauf H. T. Optimization of neural network using improved bat algorithm for data classification. Journal of Medical Imaging and Health Informatics . 2019;9(4):670–681. doi: 10.1166/jmihi.2019.2654. [DOI] [Google Scholar]
- 12.Bangyal W. H., Ahmad J., Shafi I., Abbas Q. A forward only counter propagation network-based approach for contraceptive method choice classification task. Journal of Experimental & Theoretical Artificial Intelligence . 2012;24(2):211–218. doi: 10.1080/0952813X.2011.639091. [DOI] [Google Scholar]
- 13.Bangyal W. H., Ahmed J., Rauf H. T. A modified bat algorithm with torus walk for solving global optimisation problems. International Journal of Bio-Inspired Computation . 2020;15(1):1–13. doi: 10.1504/IJBIC.2020.105861. [DOI] [Google Scholar]
- 14.Junaid M., Bangyal W. H., Ahmad J. A novel bat algorithm using sobol sequence for the initialization of population. Proceedings of the 2020 IEEE 23rd International Multitopic Conference (INMIC); November 2020; Bahawalpur, Pakistan. [DOI] [Google Scholar]
- 15.Bangyal W. H., Hameed A., Alosaimi W., Alyami H. A new initialization approach in particle swarm optimization for global optimization problems. Computational Intelligence and Neuroscience . 2021;2021:17. doi: 10.1155/2021/6628889.6628889 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Bangyal W. H., Ahmad J., Abbas Q. Recognition of off-line isolated handwritten character using counter propagation network. International Journal of Engineering and Technology . 2013;5(2):227–230. doi: 10.7763/ijet.2013.v5.548. [DOI] [Google Scholar]
- 17.Pervaiz S., Ul-Qayyum Z., Bangyal W. H., Gao L., Ahmad J. A systematic literature review on particle swarm optimization techniques for medical diseases detection. Computational and Mathematical Methods in Medicine . 2021;2021:10. doi: 10.1155/2021/5990999.5990999 [DOI] [PMC free article] [PubMed] [Google Scholar] [Retracted]
- 18.Bangyal W. H., Nisar K., Ibrahim A. A. B., Haque M. R., Rodrigues J. J. P. C., Rawat D. B. Comparative analysis of low discrepancy sequence-based initialization approaches using population-based algorithms for solving the global optimization problems. Applied Sciences . 2021;11(16):p. 7591. doi: 10.3390/app11167591. [DOI] [Google Scholar]
- 19.Bangyal W. H., Ahmad J., Abbas Q. Analysis of learning rate using CPN algorithm for hand written character recognition application. International Journal of Engineering and Technology . 2013;5(2):187–190. doi: 10.7763/ijet.2013.v5.538. [DOI] [Google Scholar]
- 20.Ashraf A., Pervaiz S., Haider Bangyal W., et al. Studying the impact of initialization for population-based algorithms with low-discrepancy sequences. Applied Sciences . 2021;11(17):8190–8241. doi: 10.3390/app11178190. [DOI] [Google Scholar]
- 21.Ashraf A., Ali Almazroi A., Haider Bangyal W., A. Alqarni M. Particle swarm optimization with new initializing technique to solve global optimization problems. Intelligent Automation & Soft Computing . 2022;31(1):191–206. doi: 10.32604/iasc.2022.015810. [DOI] [Google Scholar]
- 22.Bangyal W. H., Ahmad J., Tayyab H., Pervaiz S. An overview of mutation strategies in bat algorithm. International Journal of Advanced Computer Science and Applications . 2018;9(8):523–534. doi: 10.14569/ijacsa.2018.090866. [DOI] [Google Scholar]
- 23.Hameed W. A., Ahmad J., Nisar K., et al. New modified controlled bat algorithm for numerical optimization problem. Computers, Materials & Continua . 2022;70(2):2241–2259. doi: 10.32604/cmc.2022.017789. [DOI] [Google Scholar]
- 24.Haider Bangyal W., Batool H., Ahmed J., Rauf H. T., Bangyal S. A., Pervaiz S. An improved particle swarm optimization algorithm with Chi-Square Mutation Strategy. International Journal of Advanced Computer Science and Applications . 2019;10:3. doi: 10.14569/ijacsa.2019.0100362. [DOI] [Google Scholar]
- 25.Mozafari M., Farahbakhsh R., Crespi N. A BERT-based transfer learning approach for hate speech detection in online social media. Complex Networks and Their Applications VIII . 2020;881:928–940. doi: 10.1007/978-3-030-36687-2_77. [DOI] [Google Scholar]
- 26.Florio K., Basile V., Polignano M., Basile P., Patti V. Time of your hate: the challenge of time in hate speech detection on social media. Applied Sciences . 2020;10(12):4180–12. doi: 10.3390/APP10124180. [DOI] [Google Scholar]
- 27.Davidson T., Warmsley D., Macy M., Weber I. Automated hate speech detection and the problem of offensive language. 2017;arXiv(Icwsm):512–515. [Google Scholar]
- 28.Okoro E. M., Abara B. A., Umagba A. O., Ajonye A. A., Isa Z. S. A hybrid approach to fake news detection on social media. Nigerian Journal of Technology . 2018;37(2):p. 454. doi: 10.4314/njt.v37i2.22. [DOI] [Google Scholar]
- 29.Mishev K., Gjorgjevikj A., Vodenska I., Chitkushev L. T., Trajanov D. Evaluation of sentiment analysis in finance: from lexicons to transformers. IEEE Access . 2020;8:131662–131682. doi: 10.1109/ACCESS.2020.3009626. [DOI] [Google Scholar]
- 30.Bataa E., Wu J. An investigation of transfer learning-based sentiment analysis in Japanese. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics; 2019; Florence, Italy. pp. 4652–4657. [DOI] [Google Scholar]
- 31.Malte A., Ratadiya P. Evolution of transfer learning in natural language processing. 2019, arXiv.
- 32.Yu S., Su J., Luo D. Improving BERT-based text classification with auxiliary sentence and domain knowledge. IEEE Access . 2019;7:176600–176612. doi: 10.1109/ACCESS.2019.2953990. [DOI] [Google Scholar]
- 33.Chatsiou K. Text classification of manifestos and COVID-19 press briefings using bert and convolutional neural networks. 2020. pp. 1–12. arXiv.
- 34.Houlsby N., et al. Parameter-efficient transfer learning for NLP. 2019. https://arxiv.org/abs/1902.00751 .
- 35.Wani A., Joshi I., Khandve S., Wagh V., Joshi R. Evaluating deep learning approaches for Covid19 fake news detection. 2021. [Online]. Available: http://arxiv.org/abs/2101.04012. [DOI]
- 36.Raha T., Indurthi V., Upadhyaya A. Identifying COVID-19 fake news in social media. 2021, [Online]. Available: http://arxiv.org/abs/2101.11954.
- 37.Glazkova A., Glazkov M., Trifonov T. g2tmn at Constraint@AAAI2021: exploiting CT-BERT and ensembling learning for COVID-19 fake news detection. 2020. pp. 1–12. arXiv.
- 38.Gundapu S., Mamidi R. Transformer based automatic COVID-19 fake news detection system. 2021. pp. 1–12. http://arxiv.org/abs/2101.00180 .
- 39.Giachanou A., Zhang G., Rosso P. Multimodal multi-image fake news detection. Proceedings of the 2020 IEEE 7th International Conference on Data Science and Advanced Analytics (DSAA); October 2020; Sydney, Australia. pp. 647–654. [DOI] [Google Scholar]
- 40.Yang S., Shu K., Wang S., Gu R., Wu F., Liu H. Unsupervised fake news detection on social media: a generative approach. Proceedings of the 33rd AAAI Conf. Artif. Intell. AAAI 2019, 31st Innov. Appl. Artif. Intell. Conf. IAAI 2019 9th AAAI Symp. Educ. Adv. Artif. Intell. EAAI 2019; January 2019; Honolulu, Hawaii. pp. 5644–5651. [DOI] [Google Scholar]
- 41.Rizoiu M. A., Wang T., Ferraro G., Suominen H. Transfer learning for hate speech detection in social media. 2019. arXiv.
- 42.Devlin J., Chang M. W., Lee K., Toutanova K. BERT: pre-training of deep bidirectional transformers for language understanding. 2018. pp. 4171–4186. arXiv.
- 43.Kaliyar R. K., Goswami A., Narang P. FakeBERT: fake news detection in social media with a BERT-based deep learning approach. Multimedia Tools and Applications . 2021;80(8):11765–11788. doi: 10.1007/s11042-020-10183-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Guo T., Gao H. Content enhanced BERT-based text-to-SQL generation. 2019. pp. 2–7. arXiv.
- 45.Chen Z., Qian T. Transfer capsule network for aspect level sentiment classification. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics; August 2019; Florence, Italy. pp. 547–556. [DOI] [Google Scholar]
- 46.Sun C., Huang L., Qiu X. Utilizing BERT for aspect-based sentiment analysis via constructing auxiliary sentence. Proceedings of the NAACL HLT 2019 - 2019 Conf. North Am. Chapter Assoc. Comput. Linguist. Hum. Lang. Technol. - Proc. Conf . 2019;1:380–385. [Google Scholar]
- 47.Odel P. R. M., Nalyse T. O. A., On C.-C. O., Müller M., Salathé M. bert: a n. 2020.
- 48.Radivchev V., Nikolov A. Nikolov-radivchev at SemEval-2019 task 6: offensive tweet classification with BERT and ensembles. 2019. pp. 691–695.
- 49.Goldberg Y. Assessing BERT’s syntactic abilities. 2018. pp. 2–5.
- 50.Liu Y., Ott M., Goyal N., et al. RoBERTa: a robustly optimized BERT pretraining approach. 2019;1 http://arxiv.org/abs/1907.11692 . [Google Scholar]
- 51.Warstadt A., Zhang Y., Li X., Liu H., Bowman S. R. Learning which features matter: RoBERTa acquires a preference for linguistic generalizations (eventually). Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP); November 2020; USA. pp. 217–235. [DOI] [Google Scholar]
- 52.Murarka A., Radhakrishnan B., Ravichandran S. Detection and Classification of mental illnesses on social media using RoBERTa. 2020. http://arxiv.org/abs/2011.11226 .
- 53.Dhanalaxmi S., Agarwal R., Sinha A. Detection of COVID-19 informative tweets using RoBERTa. 2020. http://arxiv.org/abs/2010.11238 .
- 54.Kim B., Seo J., Koo M.-W. Randomly wired network based on RoBERTa and dialog history attention for response selection. IEEE/ACM Transactions on Audio, Speech, and Language Processing . 2021;29(c):2437–2442. doi: 10.1109/taslp.2021.3077119. [DOI] [Google Scholar]
- 55.Casola S., Lavelli A. 2020. pp. 101–103. FBK @ SMM4H2020: RoBERTa for detecting medications on Twitter.
- 56.Delobelle P., Winters T., Berendt B. RobBERT: a Dutch RoBERTa-based language model. Findings of the Association for Computational Linguistics: EMNLP 2020 . 2020;1:3255–3265. doi: 10.18653/v1/2020.findings-emnlp.292. [DOI] [Google Scholar]
- 57.Li B., Rudzicz F. TorontoCL at CMCL 2021 shared task: RoBERTa with multi-stage fine-tuning for eye-tracking prediction. 2021. https://github.com/LuminosoInsight/ [DOI]
- 58.Sanh V., Debut L., Chaumond J., Wolf T. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. 2019. pp. 2–6. http://arxiv.org/abs/1910.01108 .
- 59.Büyüköz B., Hürriyetoğlu A., Özgür A. Analyzing ELMo and DistilBERT on socio-political news classification. Proceedings of the Work. Autom. Extr. Socio-political Events from News 2020; May 2020; USA. pp. 9–18. [Google Scholar]
- 60.Staliūnaitė I., Iacobacci I. Compositional and lexical semantics in RoBERTa, BERT and DistilBERT: a case study on CoQA. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP); November 2020; USA. pp. 7046–7056. [DOI] [Google Scholar]
- 61.Abadeer M. Assessment of DistilBERT performance on named entity recognition task for the detection of protected health information and medical concepts. Proceedings of the 3rd Clinical Natural Language Processing Workshop; November 2020; USA. pp. 158–167. [DOI] [Google Scholar]
- 62.Mozafari J., Fatemi A., Moradi P. A method for answer selection using DistilBERT and important words. Proceedings of the 2020 6th International Conference on Web Research (ICWR); April 2020; Tehran, Iran. pp. 72–76. [DOI] [Google Scholar]
- 63.Jayarao P., Sharma A. Retraining DistilBERT for a voice shopping assistant by using universal dependencies. 2021. http://arxiv.org/abs/2103.15737 .
- 64.Lan Z., Chen M., Goodman S., Gimpel K., Sharma P., Soricut R. ALBERT: a lite BERT for self-supervised learning of language representations. 2019. pp. 1–17. http://arxiv.org/abs/1909.11942 .
- 65.Gregory H., Li S., Mohammadi P., Tarn N., Draelos R., Rudin C. A transformer approach to contextual sarcasm detection in Twitter. Proceedings of the Second Workshop on Figurative Language Processing; July 2020; USA. pp. 270–275. [DOI] [Google Scholar]
- 66.Vijjali R., Potluri P., Kumar S., Teki S. Two stage transformer model for COVID-19 fake news detection and fact checking. 2020. http://arxiv.org/abs/2011.13253 .
- 67.Ngai H., Park Y., Chen J., Parsapoor M. Transformer-based models for question answering on COVID19. 2021. pp. 1–7. http://arxiv.org/abs/2101.11432 .
- 68.Al-Garadi M. A., Yang Y.-C., Cai H., et al. Text classification models for the automatic detection of nonmedical prescription medication use from social media. BMC Medical Informatics and Decision Making . 2021;21(1):1–13. doi: 10.1186/s12911-021-01394-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Zhao Y., Tao X. {ZYJ}123@{D}ravidian{L}ang{T}ech-{EACL}2021: offensive Language Identification based on {XLM}-{R}o{BERT}a with {DPCNN}. Proceedings of the First Work. Speech Lang; October 2021; Parkville, Victoria. Technol. Dravidian Lang.; pp. 216–221. [Google Scholar]
- 70.Ou X., Li H. YNU @ dravidian-CodeMix-fire2020: XLM-RoBERTa for multi-language sentiment analysis. 2020. pp. 4–9.
- 71.Ou X., Li H. YNU_OXZ @ HaSpeeDe 2 and AMI: XLM-RoBERTa with ordered neurons LSTM for classification task at EVALITA 2020. EVALITA Evaluation of NLP and Speech Tools for Italian - December 17th, 2020 . 2020;2765:102–109. doi: 10.4000/books.aaccademia.6912. [DOI] [Google Scholar]
- 72.Xie S., Ma J., Yang H., Jiang L., Mo Y., Shen J. PALI at SemEval-2021 task 2: fine-tune XLM-RoBERTa for word in context disambiguation. 2021. http://arxiv.org/abs/2104.10375 . [DOI]
- 73.Clark K., Luong M.-T., Le Q. V., Manning C. D. ELECTRA: pre-training text encoders as discriminators rather than generators. 2020. pp. 1–18. [Online]. Available: http://arxiv.org/abs/2003.10555.
- 74.Das K. A., Baruah A., Barbhuiya F. A., Dey K. Ensemble of ELECTRA for profiling fake news spreaders. 2020;September:22–25. [Google Scholar]
- 75.Shen J., Liu J., Liu T., Yu C., Han J. Training ELECTRA augmented with multi-word selection. 2021;2 doi: 10.18653/v1/2021.findings-acl.219. , http://arxiv.org/abs/2106.00139. [DOI] [Google Scholar]
- 76.Lewis M., Liu Y., Goyal N., et al. BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; 2020; Stroudsburg, PA, USA. pp. 7871–7880. [DOI] [Google Scholar]
- 77.Popa C., Rebedea T. BART}-{TL}: weakly-supervised topic label generation. Proceedings of the 16th Conf. Eur. Chapter Assoc. Comput. Linguist. Main; 2021; Berlin. pp. 1418–1425. https://www.aclweb.org/anthology/2021.eacl-main.121 . [Google Scholar]
- 78.Mustar A., Lamprier S., Piwowarski B. Using BERT and BART for query suggestion. CEUR Workshop Proc . 2020;2621 [Google Scholar]
- 79.Xing Y., Shi Z., Meng Z., Ma Y., Wattenhofer R. KM-BART: knowledge enhanced multimodal BART for visual commonsense generation. 2021. http://arxiv.org/abs/2101.00419 . [DOI]
- 80.de Bruyn M., Lotfi E., Buhmann J., Daelemans W. BART for knowledge grounded conversations. CEUR Workshop Proc . 2020;2666 [Google Scholar]
- 81.Ahmad S., Asghar M. Z., Alotaibi F. M., Awan I. Detection and classification of social media-based extremist affiliations using sentiment analysis techniques. Human-centric Computing and Information Sciences . 2019;9(1) doi: 10.1186/s13673-019-0185-6. [DOI] [Google Scholar]
- 82.Ohashi S., Kajiwara T., Chu C., Takemura N., Nakashima Y., Nagahara H. IDSOU at WNUT-2020 task 2: identification of informative COVID-19 English tweets. Proceedings of the Sixth Workshop on Noisy User-generated Text (W-NUT 2020); January 2020; USA. pp. 428–433. [DOI] [Google Scholar]
- 83.Khanam Z., Alwasel B. N., Sirafi H., Rashid M. Fake news detection using machine learning approaches. IOP Conference Series: Materials Science and Engineering . 2021;1099(1) doi: 10.1088/1757-899x/1099/1/012040.012040 [DOI] [Google Scholar]
- 84.Jain A., Shakya A., Khatter H., Gupta A. K. A smart system for fake news detection using machine learning. Proceedings of the 2019 International Conference on Issues and Challenges in Intelligent Computing Techniques (ICICT) 2019; September 2019; USA. pp. 2019–2022. [DOI] [Google Scholar]
- 85.Kong S. H., Tan L. M., Gan K. H., Samsudin N. H. Fake news detection using deep learning. Proceedings of the 10th Symposium on Computer Applications & Industrial Electronics (ISCAIE); June 2020; Malaysia. [DOI] [Google Scholar]
- 86.Yang Y., Zheng L., Zhang J., Cui Q., Li Z., Yu P. S. TI-CNN: convolutional neural networks for fake news detection. 2018. http://arxiv.org/abs/1806.00749 .
- 87.Monica C., Nagarathna N. Detection of fake tweets using sentiment analysis. SN Computer Science . 2020;1(2):1–7. doi: 10.1007/s42979-020-0110-0. [DOI] [Google Scholar]
- 88.Khalil A., Hajjdiab H., Al-Qirim N. Detecting fake followers in twitter: a machine learning approach. International Journal of Machine Learning and Computing . 2017;7(6):198–202. doi: 10.18178/ijmlc.2017.7.6.646. [DOI] [Google Scholar]
- 89.Azizan S. A., Aziz I. A. Terrorism detection based on sentiment analysis using machine Learning.pdf. Journal of Engineering and Applied Sciences . 2017;12(3):691–698. [Google Scholar]
- 90.Wei Y., Singh L., Martin S. Identification of extremism on twitter. Proceedings of the 2016 IEEE/ACM Int. Conf. Adv. Soc. Networks Anal. Mining, ASONAM 2016; July 2016; USA. pp. 1251–1255. [DOI] [Google Scholar]
- 91.Gambäck B., Sikdar U. K. Using convolutional neural networks to classify hate-speech. Proceedings of the First Workshop on Abusive Language Online; August 2017; Vancouver, BC, Canada. pp. 85–90. [DOI] [Google Scholar]
- 92.Kowsari K., Brown D. E., Heidarysafa M., Jafari Meimandi K., Gerber M. S., Barnes L. E. HDLTex: hierarchical deep learning for text classification. Proceedings of the 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA); December 2017; Cancun, Mexico. pp. 364–371. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data are not available until the thesis defense. In case of any queries, the readers can contact the corresponding author.