Abstract
Most consumers rely on online reviews when deciding to purchase e-commerce services or products. Unfortunately, the main problem of these reviews, which is not completely tackled, is the existence of deceptive reviews. The novelty of the proposed system is the application of opinion mining on consumers' reviews to help businesses and organizations continually improve their market strategies and obtain an in-depth analysis of the consumers' opinions regarding their products and brands. In this paper, the long short-term memory (LSTM) and deep learning convolutional neural network integrated with LSTM (CNN-LSTM) models were used for sentiment analysis of reviews in the e-commerce domain. The system was tested and evaluated by using real-time data that included reviews of cameras, laptops, mobile phones, tablets, televisions, and video surveillance products from the Amazon website. Data preprocessing steps, such as lowercase processing, stopword removal, punctuation removal, and tokenization, were used for data cleaning. The clean data were processed with the LSTM and CNN-LSTM models for the detection and classification of the consumers' sentiment into positive or negative. The LSTM and CNN-LSTM algorithms achieved an accuracy of 94% and 91%, respectively. We conclude that the deep learning techniques applied here provide optimal results for the classification of the customers' sentiment toward the products.
1. Introduction
Web 3.0 has the main features of the semantic web, artificial intelligence, connectivity, etc., allowing people to use social media to communicate and express their opinions about real-world events. In this context, the analysis of users' reviews is essential for companies to grow worldwide. This makes opinion mining a key player in the analysis of reviews and discussions. Nowadays, companies analyze this type of information to improve the quality and performance of the products and, consequently, survive in a competitive market. Opinion mining can be expressed as the reason behind any action or movement that people use to follow the reason [1].
Within the huge amount of data generated on the Internet, important information is hidden. Data mining techniques are used to extract information and solve various problems. Online product reviews have two important aspects under which data are stored on the Internet. Commercial websites are platforms where users express their sentiment or opinion on several topics. Sentiment analysis refers to a broad area of natural language processing (NLP), computational linguistics, and text mining [2]. The use of these techniques leads to the extraction and analysis of the opinion on a given product. Opinion mining defines an opinion as positive or negative, and sentiment analysis defines the polarity value of a user's opinion on a particular product or service. The current approaches of sentiment analysis are mainly [3] machine learning algorithms [4], lexicon-based methods, [5] and hybrid models [6, 7].
Negation is a prevalent morphological development that impacts polarity and, therefore, must be reflected in the assessment of sentiment. Automatic detection of negation in news articles is required for numerous text processing applications, including sentiment analysis. Here, we explored the role and importance of users' reviews concerning particular products on the decision using sentiment analysis. We present experimental results that demonstrate that sentiment analysis is appropriate to this end. The goal was to determine the polarity of the natural language of texts written in product reviews.
The existing straightforward approaches are statistical, based on frequencies of positive and negative words. Recently, researchers discovered new ways to account for other aspects of content, such as structural or semantic features. The present work focuses on the identification of document-level negation by using multiple computational methods. In recent years, with the exponential growth of smartphone use, many people are connected to social networking platforms, like Facebook, Twitter, and Instagram. Social networks have become a field to express beliefs or opinions, emotions, thoughts, personal issues, places, or personalities.
There are numerous studies applying sentiment analysis, some of which used real-time data from Twitter for extracting patterns by employing the Twitter-streaming application programming interface (API) [8, 9]. The sentiment analyzers are divided into two types: SentiWordNet [10] and WordNet [11]. Sentiment analysis uses positive and negative scores to classify opinions. By developing a model to analyze word sequence disambiguation [12], the Twitter-streaming API was used to gather data concerning the Indonesian presidential elections [13]. Needless tweets were removed, and the remaining data were investigated for sentimental aspects by dividing each tweet into numerous sub-tweets and calculating the sentiment polarity of the sub-tweets for predicting the consequence of the elections. The mean absolute error metric was used to evaluate the results, it noted that the prediction error was 0.6 better than the previous study [14]. To predict the Swedish election outcome by using Twitter data, a system was developed [15]. To predict the outcome of the European elections, a new method was designed that studied the similarity of the structure with the outcome of the vote. Another method was created to test Brazilian municipal elections in six cities [16]. In this methodology, sentiment analysis was applied along with a stratified sample [17] of users to compare the characteristics of the findings with the actual voters.
Many researchers have used machine learning and artificial intelligence to analyze the sentiment of tweets [18, 19]. In [20], the Naive Bayes, support vector machine (SVM) [21], and information entropy-based [22] models were applied to classify product reviews. A hybrid machine learning algorithm based on Twitter opinion mining was proposed in [23]. Heydari et al. [24] proposed time series model for fraudulent sentiment reviewer analysis. Hajek et al. [25] developed a deep feedforward neural network and convolution model to detect fake positive and negative review in an Amazon dataset. Long et al. [26] applied LSTM with multi-head attention network for predicting sentiment-based text using China social media dataset. Dong et al. [27] proposed supervised machine linear regression for predicting sentiment of customers presented in online shopping data using sentiment analysis learning approaches.
Researchers have been focusing on developing powerful models to deal with the ever-increasing complexity of big data [28, 29], as well as expanding sentiment analysis to a wide range of applications [30, 31], from financial forecasting to marketing strategies [32] among other areas [33, 34]. However, only a few of them analyzed different deep learning approaches to give real evidence of their performance [35]. Deep learning techniques are becoming increasingly popular. When assessing the performance of a single approach on a single dataset in a specific area, the results suggest that CNN and RNN have relatively good accuracy. Based on AdaBoost combination, Gao et al. [36] proposed CNN model for sentiment analysis in user-generated text. In this vein, Hassan and Mahmood [37] demonstrated that the CNN and RNN models overcame the problem of short texts in deep learning models.
Some traditional approaches, which are assisted by machine learning techniques, are based on aspects of the used language. Using the domain of movie opinions, Pang et al. [18] studied the performance of various machine learning algorithms, including Naive Bayes, maximum entropy, and SVM. By using SVM with unigrams, they achieved an accuracy of 82.9%. NLP is typically used to extract features used by a sentiment classifier. In this aspect, the majority of NLP strategies are centered on the usage of n-grams but the use of a bag-of-words strategy is also common [38, 39]. Numerous studies have demonstrated significant results when employing the bag-of-words as a text representation for item categorization [40–44].
Researchers have taken advantage of NLP themes to develop deep learning models based on neural networks with more than three layers, according to the journal Nature. Most of these studies found that deep learning models accurately detect sentiment in various situations. The CNN [45, 46], RNN [47], deep neural network [48], recursive neural deep model [49], and the attention-based bidirectional CNN-RNN [50] models are some representative examples. Some researchers combine models, which are then referred to as hybrid neural networks. The hierarchical bidirectional RNN is an example of a hybrid neural network [51]. The main issue with sentiment analysis of product reviews in the e-commerce domain is the existence of fake reviews that lead customers to select undesired products [52].
The main contributions of the proposed research are the following:
The generation of a sentiment score using a lexicon-based approach for each product review of the dataset.
Labeling the review texts as negative if the generated sentiment score is <0 or positive if the score is >1.
The combination of all product reviews into a single data frame to obtain more sentiment-related words.
Improving the accuracy by developing a hybrid deep learning model combining the CNN and LSTM models for the product-related sentiment classification.
Comparing the classification performance of the CNN-LSTM and LSTM models.
2. Materials and Methods
The proposed methodology for predicting the review-related sentiments is based on the deep learning algorithms presented here. The phases of the proposed system are the following: dataset collection, data preprocessing, generating the sentiment score, polarity calculation, applying the CNN-LSTM model, evaluation metrics, and analysis of the results. Figure 1 shows the framework of the proposed methodology used in the present study.
2.1. Datasets
To evaluate the proposed system, the dataset [53] was collected from reviews on the Amazon website in JSON file format. Each JSON file comprises a number of reviews (Table 1). The dataset includes reviews of laptops, mobile phones, tablets, televisions, and video surveillance products. The data preprocessing includes various steps, such as lowercase processing with meta-features like the reviewer's ID, the product ID, and the review text.
Table 1.
Product name | Review count |
---|---|
Laptops | 1,946 |
Mobile phones | 1,918 |
Tablets | 1,894 |
Televisions | 1,596 |
Video surveillance products | 2,597 |
2.2. Data Preprocessing
We implemented different preprocessing steps aiming at cleaning the review texts so that they are easy to process. The following preprocessing methods were performed on the dataset as a whole.
2.2.1. Lowercase
It entails converting whole words of the review text into lowercase words.
2.2.2. Stopword Removal
Stopwords are widely used words in a language, such as “the,” “a,” “an,” “is,” and “are”. As these words do not carry any information significant for the model, they were removed from the content of the review.
2.2.3. Punctuation Removal
All punctuation marks in the review texts were removed.
2.2.4. One-Word Review Elimination
Reviews that included only one word were eliminated.
2.2.5. Contraction Removal
This process replaces a word originally written in the short form with the respective full form; for instance, “when've” becomes “when have.”
2.2.6. Tokenization
Each sentence of the review texts was divided into small pieces of words or tokens.
2.2.7. Part-of-Speech Tagging
This step is used to tag each word present in the sentence with a POS tag, for example, “VB” for a verb, “AJJ” for an adjective, and “NN” for a noun.
2.2.8. Score Generation
The review text was evaluated for sentiment, and a score was generated. For calculating the sentiment score, the dataset was matched with opinion lexicon [53] that consists of 5,000 positive words and 4,500 negative words with their respective scores. The sentiment score was calculated for each review text based on the scores of the lexicon. The review text was labeled as positive if the score was >0; otherwise, it was labeled as negative.
2.2.9. Word Embeddings
We calculated numerical vectors with every preprocessed sentence in the product review dataset using the “Word embeddings” method. To create word indices, we first turned all of the review text terms into sequences. The Keras text tokenizer [54] is being used to obtain those indices. We made sure that no term or word gets a zero index in the tokenizer, and that the vocabulary size is adjusted properly. Then, for each single word in the training and testing sets, a distinctive index is generated, which is employed to create numeric vectors of all review texts of the dataset.
2.3. The CNN-LSTM Model
Figure 2 presents the structure of the CNN-LSTM model used for sentiment classification of customers' reviews using an Amazon dataset.
2.3.1. Embedding Layer
This is the initial layer of the CNN-LSTM model that is used to transform each word in the training dataset into an actual-valued vector, meaning that a set of sentiment-related words are constructed and transformed into a numerical form. This process is known as word embedding. The embedding layer consisted of three components: the vocabulary size (maximum features; 15,000 words), the embedding dimensions (50), and the input sequence length (400 words).
2.3.2. Dropout Layer
The main task of this layer is to avoid the overfitting of the model [52]. Here, we assigned the value 0.4 to the dropout rate parameter, where this value has a range between 0 and 1. The main function of the dropout layer is to arbitrarily deactivate a set of neurons in the embedding layer, where every neuron denotes the dense exemplification of a sentiment word in a review text.
CNN is a deep learning technique used in different areas such as natural language preprocessing tasks, computer vision, and medical image processing.
2.3.3. Convolution Layer
The third layer of the CNN-LSTM model is used for the extraction of features from the input matrix. It uses n convolution filters that operate over the elements of the input sequence matrix to find the convolutions for each sequence. We set the number of filters to 64 and the size of the filter kernel to 3 × 3.
2.3.4. Max Pooling Layer
This layer performs downsampling beside the spatial dimensionality of the given input sequences. It considers the maximum value of all input features in the pool of each filter kernel. It has assigned to 5 × 5 kernel.
2.3.5. LSTM Layer
LSTM is a type of RNN capable of learning long-term dependence [52]. We used an LSTM layer and assigned it to 50 hidden units toward the next layer. One of the most notable advantages of employing a convolutional neural network as feature extraction technique beyond a traditional LSTM is the reduction in the aggregating amount of features. Throughout the feature extraction process, a sentiment classification model uses these features (words) for prediction of the product review text as positive or negative sentiment. LSTM executes precalculations for the input sequences before providing an output to the last layer of the network. In every cell, four discrete computations are conducted based on four gates: input (it), forget (ft), candidate (ct), and output (ot). The structure of the LSTM model is presented in Figure 3. The equations for these gates are as follows:
(1) |
where sig and tanh are the sigmoid and tangent activation functions, respectively, X is the input data, W and b represent the weight and bias factor, respectively, Ct is the cell state, c ~ t is the candidate gate, and ht refers to the output of the LSTM cell.
2.3.6. Dense Layer (Fully Connected Layer)
This is a hidden layer in the CNN-LSTM model. It consists of 512 artificial connected neurons that connect all neurons of the network. The function applied to this layer is the rectified linear unit described by the following equation:
(2) |
2.3.7. Sigmoid Activation Function
It is the first layer that detects and classifies the output classes (positive or negative sentiment). The sigmoid function formula is given as follows (Algorithm 1):
(3) |
2.4. Evaluation Metrics
To evaluate the proposed models (CNN-LSTM and LSTM), the accuracy, precision, recall, F1-score, and specificity metrics were used. The performance measurements are presented below:
(4) |
where true positive (TP) represents the total number of samples that are successfully classified as positive sentiment, false positive (FP) is the total number of samples that are incorrectly classified as negative sentiments, true negative (TN) denotes the total number of samples that are successfully classified as negative sentiment, and false negative (FN) represents the total number of samples that are incorrectly classified as positive sentiments.
3. Experimental Results
In this section, we present the experimental results of the application of the CNN-LSTM and LSTM models for the analysis and prediction of sentiment in the e-commerce domain. We used hardware with 4 GB RAM and an i7 2800 CPU and ran the experiments on the Jupyter environment. The evaluation metrics (accuracy, precision, F1-score, recall, and specificity) were employed to examine the proposed system. The word cloud (sentiment words and product names) of the dataset is presented in Figure 4, which shows graphical representations of words (large font words) that give greater importance to that seem more repeatedly in the used product review dataset.
3.1. Data Splitting
In this phase, we divided the dataset that consisted of 13,057 product reviews into 70% training, 10% validation, and 20% testing datasets. Then, the CNN-LSTM and LSTM models were applied to detect and classify the review texts into positive or negative. Table 2 shows the splitting of the dataset.
Table 2.
Total number of reviews | Training set 80% | Validation set 10% | Testing set 20% |
---|---|---|---|
13,057 (11,184 positive; 1,873 negative) | 9,400 | 1,045 | 2,612 |
3.2. Results and Discussion
Table 3 shows the results of the deep learning approaches. The CNN-LSTM model achieved high accuracy (96%).
Table 3.
Models | Specificity | Accuracy (%) | Precision (%) | Recall (%) | F1-score (%) |
---|---|---|---|---|---|
LSTM | 95 | 91.03 | 92.07 | 97.73 | 95.50 |
CNN-LSTM | 96 | 94 | 94 | 99 | 96.03 |
The confusion matrix of the CNN-LSTM and LSTM models is shown in Figure 5. The confusion matrix is used to present the rates of TP, FP, TN, and FN of the sample. Based on these rates, the evaluation metrics (specificity, accuracy, recall, precision, and F1-score) were calculated to evaluate the CNN-LSTM model using unseen data to predict the sentiment of customers. LSTM resulted in 82.24% TP, while CNN-LSTM resulted in 83.54% TP. As for misclassification, LSTM resulted in 6.39% FP and CNN-LSTM in 5.28% FP, indicating that the CNN-LSTM model was slightly better than the LSTM model.
The accuracy performance of LSTM for the training and validation datasets is presented in Figure 6. The LSTM model presented increasing accuracy during the training phase (from 86% to 94%), whereas in the testing phase, it achieved 91% accuracy with 10 epochs. The loss of the LSTM model in the training phase decreased from 5 to 0.35, while in the validation phase, the model loss decreased from 0.3 to 0.27.
The accuracy performance of the CNN-LSTM during the training phase increased from 87.50% to 97%. In the validation phase, the accuracy performance reached 94% (Figure 7(a)). The loss of the CNN-LSTM model in the validation phase was 0.20 (Figure 7(b)).
The dataset developed by Rajkumar et al. [53] proposed SVM and Naive Bayes methods to predict sentiment analysis. They collected data from Amazon concerning mobile phones, tablets, cameras, and televisions. They applied the SVM method to each dataset individually. Here, we applied deep learning models to all the datasets combined. The empirical results of our system were compared with the results of [28] and are shown in Table 4. The CNN-LSTM model achieved an accuracy of 94%.
Table 4.
Models | Datasets | Accuracy (%) |
---|---|---|
Support vector machine [28] | Televisions, tablets, mobile phones, laptops, and video surveillance | 88, 84, 92, 88, and 93 |
Proposed system (CNN-LSTM) | All dataset | 94 |
4. Conclusion
Recently, sentiment analysis has become a valuable tool for the generation and evaluation of different types of data, helping the decision-making processes that lead to the improvement of businesses and companies. Social networking creates a large amount of data that require processing and analysis to obtain relevant insights. In the present study, the experimental dataset was collected from the Amazon website and included reviews of laptops, mobile phones, tablets, televisions, and video surveillance products. The lexicon-based approach was used for the calculation of the sentiment score for each review text. The output of the preprocessed data was classified with the LSTM and CNN-LSTM models. The experimental results showed that our model was satisfactory in all the measurement metrics.
Acknowledgments
The authors deeply acknowledge Taif University for supporting this research through Taif University Researchers Supporting Project Number (TURSP-2020/328), Taif University, Taif, Saudi Arabia. This work was supported by the Deanship of Scientific Research, Vice Presidency for Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia (Project No: GRANT674).
Data Availability
We have collected dataset from authors of research article: https://www.researchgate.net/publication/326985579_ Sentiment_Analysis_ on_Product_Reviews_Using_ Machine _Learning_ Techniques_Proceeding_of_CISC_ 2017
Conflicts of Interest
The authors declare that they have no conflicts of interest.
References
- 1.Cambria E., Das D. S. A. Affective computing and sentiment analysis. A Practical Guide to Sentiment Analysis . 2017;31:1–10. doi: 10.1007/978-3-319-55394-8_1. [DOI] [Google Scholar]
- 2.Jagdale O., Harmalkar V., Chavan S., Sharma N. Twitter mining using R. Int. J. Eng. Res. Adv. Tech. . 2017;3:252–256. [Google Scholar]
- 3.Medhat W., Hassan A., Korashy H. Sentiment analysis algorithms and applications: a survey. Ain Shams Engineering Journal . 2014;5(4):1093–1113. doi: 10.1016/j.asej.2014.04.011. [DOI] [Google Scholar]
- 4.Sebastiani F. Machine learning in automated text categorization. ACM Computing Surveys . 2002;34(1):1–47. doi: 10.1145/505282.505283. [DOI] [Google Scholar]
- 5.Taboada M., Brooke J., Tofiloski M., Voll K., Stede M. Lexicon-based methods for sentiment analysis. Computational Linguistics . 2011;37(2):267–307. doi: 10.1162/coli_a_00049. [DOI] [Google Scholar]
- 6.Prabowo R., Thelwall M. Sentiment analysis: a combined approach. Journal of Informetrics . 2009;3(2):143–157. doi: 10.1016/j.joi.2009.01.003. [DOI] [Google Scholar]
- 7.Dang Y., Zhang Y., Chen H. A lexicon-enhanced method for sentiment classification: an experiment on online product reviews. IEEE Intelligent Systems . 2010;25(4):46–53. doi: 10.1109/mis.2009.105. [DOI] [Google Scholar]
- 8.Jose R., Chooralil V. S. Prediction of election result by enhanced sentiment analysis on twitter data using word sense disambiguation. Proceedings of the 2015 International Conference on Control Communication & Computing India (ICCC); November 2015; Trivandrum, India. pp. 638–641. [DOI] [Google Scholar]
- 9.Esuli A., Sebastiani F. S. A High-Coverage Lexical Resource for Opinion Mining . Pisa, Italy: Institute of Information Science and Technologies (ISTI) of the Italian National Research Council (CNR); 2006. [Google Scholar]
- 10.Miller G. A. WordNet. Communications of the ACM . 1995;38(11):39–41. doi: 10.1145/219717.219748. [DOI] [Google Scholar]
- 11.Navigli R. Word sense disambiguation. ACM Computing Surveys . 2009;41(2):1–69. doi: 10.1145/1459352.1459355. [DOI] [Google Scholar]
- 12.Ibrahim M., Abdillah O., Wicaksono A. F., Adriani M. Buzzer detection and sentiment analysis for predicting presidential election results in a twitter nation. Proceedings of the 2015 IEEE International Conference on Data Mining Workshop (ICDMW); November 2015; Atlantic City, NJ, USA. pp. 1348–1353. [DOI] [Google Scholar]
- 13.Dokoohaki N., Zikou F., Gillblad D., Matskin M. Predicting Swedish elections with twitter: a case for stochastic link structure analysis. Proceedings of the 2015 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM); August 2015; Paris, France. pp. 1269–1276. [Google Scholar]
- 14.Liben-Nowell D., Kleinberg J. The link-prediction problem for social networks. Journal of the American Society for Information Science and Technology . 2007;58(7):1019–1031. doi: 10.1002/asi.20591. [DOI] [Google Scholar]
- 15.Miranda Filho R., Almeida J. M., Pappa G. L. Twitter population sample bias and its impact on predictive outcomes: a case study on elections. Proceedings of the 2015 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM); August 2015; Paris, France. pp. 1254–1261. [Google Scholar]
- 16.Foreman E. Survey Sampling Principles . Boca Raton, FL, USA: CRC Press; 1991. [Google Scholar]
- 17.Alsubari S. N., Deshmukh S. N., Abdullah Alqarni A., et al. Data analytics for the identification of fake reviews using supervised learning. Computers, Materials & Continua . 2022;70(2):3189–3204. [Google Scholar]
- 18.Pang B., Lee L., Vaithyanathan S. Thumbs up?. Proceedings of the ACL-02 conference on Empirical methods in natural language processing—EMNLP ‘02’; July 2002; Philadelphia, PA, USA. pp. 79–86. [DOI] [Google Scholar]
- 19.Gautam G., Yadav D. Sentiment analysis of twitter data using machine learning approaches and semantic analysis. Proceedings of the 2014 Seventh International Conference on Contemporary Computing (IC3); August 2014; Noida, India. pp. 437–442. [DOI] [Google Scholar]
- 20.Joachims T. Text categorization with support vector machines: learning with many relevant features. Proceedings of the 10th European Conference on Machine Learning; April 1998; Chemnitz, Germany. pp. 137–142. [DOI] [Google Scholar]
- 21.Berger A. L., Pietra V. J. D., Pietra S. A. D. A maximum entropy approach to natural language processing Comput. Linguist . 1996;22:39–71. [Google Scholar]
- 22.Khan F. H., Bashir S., Qamar U. Tom: twitter opinion mining framework using hybrid classification scheme. Decision Support Systems . 2014;57:245–257. doi: 10.1016/j.dss.2013.09.004. [DOI] [Google Scholar]
- 23.Mukherjee A., Venkataraman V., Liu B., Glance N. S. What yelp fake review filter might be doing?. Proceedings of the Seventh International AAAI Conference on Weblogs and Social Media; July, 2013; Cambridge, MA, USA. pp. 409–418. [Google Scholar]
- 24.Heydari A., Tavakoli M., Salim N. Detection of fake opinions using time series. Expert Systems with Applications . 2016;58:83–92. doi: 10.1016/j.eswa.2016.03.020. [DOI] [Google Scholar]
- 25.Hajek P., Barushka A., Munk M. Fake consumer review detection using deep neural networks integrating word embeddings and emotion mining. Neural Computing & Applications . 2020;32(23) doi: 10.1007/s00521-020-04757-2.17259 [DOI] [Google Scholar]
- 26.Long F., Zhou K., Ou W. Sentiment Analysis of Text Based on Bidirectional LSTM with Multi-Head Attention. Advanced Optical Imaging for Extreme Environments . 2019;7 doi: 10.1109/ACCESS.2019.2942614.141960 [DOI] [Google Scholar]
- 27.Dong J., Chen Y., Gu A., et al. Potential Trend for Online Shopping Data Based on the Linear Regression and Sentiment Analysis. Mathematical Problems in Engineering . 2020;2020:11. doi: 10.1155/2020/4591260.4591260 [DOI] [Google Scholar]
- 28.Roy S. S., Biba M., Kumar R., Kumar R., Samui P. Handbook of Research on Soft Computing and Nature-Inspired Algorithms . Hershey, Pennsylvania: IGI Global; 2017. A new SVM method for recognizing polarity of sentiments in twitter; pp. 281–291. [DOI] [Google Scholar]
- 29.Keenan M. J. S. Advanced Positioning, Flow, and Sentiment Analysis in Commodity Markets . Hoboken, NJ, USA: Wiley; 2018. [Google Scholar]
- 30.Satapathy R., Cambria E., Hussain A. Sentiment Analysis in the Bio-Medical Domain . Berlin, Germany: Springer; 2017. [Google Scholar]
- 31.Rajput A. Innovation in Health Informatics . Amsterdam, The Netherlands: Elsevier; 2020. natural language processing, sentiment analysis, and clinical analytics; pp. 79–97. [DOI] [Google Scholar]
- 32.Qian J., Niu Z., Shi C. Sentiment analysis model on weather related tweets with deep neural network. Proceedings of the 2018 10th International Conference on Machine Learning and Computing; February 2018; Macau, China. pp. 31–35. [DOI] [Google Scholar]
- 33.Pham D.-H., Le A.-C. Learning multiple layers of knowledge representation for aspect based sentiment analysis. Data & Knowledge Engineering . 2018;114:26–39. doi: 10.1016/j.datak.2017.06.001. [DOI] [Google Scholar]
- 34.Preethi G., Krishna P. V., Obaidat M. S., Saritha V., Yenduri S. Application of deep learning to sentiment analysis for recommender system on cloud. Proceedings of the 2017 International Conference on Computer, Information and Telecommunication Systems (CITS); July 2017; Dalian, China. pp. 93–97. [Google Scholar]
- 35.Tul Q., Ali M., Riaz A., et al. Sentiment analysis using deep learning techniques: a review. International Journal of Advanced Computer Science and Applications . 2017;8(6):p. 424. doi: 10.14569/ijacsa.2017.080657. [DOI] [Google Scholar]
- 36.Gao Y., Rong W., Shen Y., Xiong Z. Convolutional neural network based sentiment analysis using Adaboost combination. Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN); July 2016; Vancouver, BC, Canada. pp. 1333–1338. [Google Scholar]
- 37.Hassan A., Mahmood A. Deep learning approach for sentiment analysis of short texts. Proceedings of the Third International Conference on Control, Automation and Robotics (ICCAR); April 2017; Nagoya, Japan. pp. 705–710. [DOI] [Google Scholar]
- 38.Kraus M., Feuerriegel S. Sentiment analysis based on rhetorical structure theory:Learning deep neural networks from discourse trees. Expert Systems with Applications . 2019;118:65–79. doi: 10.1016/j.eswa.2018.10.002. [DOI] [Google Scholar]
- 39.Li L., Goh T.-T., Jin D. How textual quality of online reviews affect classification performance: a case of deep learning sentiment analysis. Neural Computing & Applications . 2018;32(9):4387–4415. doi: 10.1007/s00521-018-3865-7. [DOI] [Google Scholar]
- 40.Singhal P., Bhattacharyya P. Sentiment Analysis and Deep Learning: A Survey . Bombay, Indian: Center for Indian Language Technology Indian Institute of Technology; 2016. [Google Scholar]
- 41.Abid F., Alam M., Yasir M., Li C. Sentiment analysis through recurrent variants latterly on convolutional neural network of Twitter. Future Generation Computer Systems . 2019;95:292–308. doi: 10.1016/j.future.2018.12.018. [DOI] [Google Scholar]
- 42.Alharbi A. S. M., de Doncker E. Twitter sentiment analysis with a deep neural network: an enhanced approach using user behavioral information. Cognitive Systems Research . 2019;54:50–61. doi: 10.1016/j.cogsys.2018.10.001. [DOI] [Google Scholar]
- 43.Beigi G., Hu X., Maciejewski R., Liu H. Sentiment Analysis and Ontology Engineering . Cham, Switzerland: Springer; 2016. An overview of sentiment analysis in social media and its applications in disaster relief; pp. 313–340. [DOI] [Google Scholar]
- 44.Sikka K., Wu T., Susskind J., Bartlett M. Exploring bag of words architectures in the facial expression domain. Proceedings of the European Conference on Computer Vision; October 2012; Florence, Italy. pp. 250–259. [DOI] [Google Scholar]
- 45.Bekkerman R., Allan J. Using Bigrams in Text Categorization . UMass: Amherst, MA, USA: Center of Intelligent Information Retrieval; 2004. [Google Scholar]
- 46.Krapac J., Verbeek J., Jurie F. Modeling spatial layout with Fisher vectors for image categorization. Proceedings of the International Conference on Computer Vision; November 2011; Barcelona, Spain. pp. 1487–1494. [DOI] [Google Scholar]
- 47.Wang X., McCallum A., Wei X. Topical n-grams: phrase and topic discovery, with an application to information retrieval. Proceedings of the Seventh IEEE International Conference on Data Mining (ICDM); October 2007; Omaha, Nebraska. pp. 697–702. [DOI] [Google Scholar]
- 48.Severyn A., Moschitti A. Twitter sentiment analysis with deep convolutional neural networks. Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval; August 2015; Santiago, Chile. pp. 959–962. [DOI] [Google Scholar]
- 49.Yanmei L., Yuda C. Research on Chinese micro-blog sentiment analysis based on deep learning. Proceedings of the 8th International Symposium on Computational Intelligence and Design (ISCID); December 2015; Hangzhou, China. pp. 358–361. [DOI] [Google Scholar]
- 50.Basiri M. E., Nemati S., Abdar M., Cambria E., Acharya U. R. ABCDM: an attention-based bidirectional CNN-RNN deep model for sentiment analysis. Future Generation Computer Systems . 2021;115:279–294. doi: 10.1016/j.future.2020.08.005. [DOI] [Google Scholar]
- 51.Arras L., Montavon G., Müller K. R., Samek W. Explaining Recurrent Neural Network Predictions in Sentiment Analysis
- 52.Alsubari S. N., Deshmukh S. N., Al-Adhaileh M. H., Alsaade F. W., Aldhyani T. H. Development of Integrated Neural Network Model for Identification of Fake Reviews in E-Commerce Using Multidomain Datasets. Applied Bionics and Biomechanics . 2021;2021:11. doi: 10.1155/2021/5522574.5522574 [DOI] [PMC free article] [PubMed] [Google Scholar] [Retracted]
- 53.Jagdale R. S., Shirsat V. S., Deshmukh S. N. Sentiment analysis on product reviews using machine learning techniques. Cognitive informatics and soft computing. Advances in Intelligent Systems and Computing . 2018;768 doi: 10.1007/978-981-13-0617-4_61. [DOI] [Google Scholar]
- 54. Tensorflow and text preprocessing, Accessed 20 January 2022, https://www.tensorflow.org/api_docs/python/tf/keras/preprocess /text/Tokenize.
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
We have collected dataset from authors of research article: https://www.researchgate.net/publication/326985579_ Sentiment_Analysis_ on_Product_Reviews_Using_ Machine _Learning_ Techniques_Proceeding_of_CISC_ 2017