Skip to main content
PLOS One logoLink to PLOS One
. 2023 Apr 26;18(4):e0283896. doi: 10.1371/journal.pone.0283896

User review analysis of dating apps based on text mining

Qian Shen 1, Siteng Han 1, Yu Han 1, Xi Chen 1,*
Editor: Viacheslav Kovtun2
PMCID: PMC10132638  PMID: 37099495

Abstract

With the continuous development of information technology, more and more people have become to use online dating apps, and the trend has been exacerbated by the COVID-19 pandemic in these years. However, there is a phenomenon that most of user reviews of mainstream dating apps are negative. To study this phenomenon, we have used topic model to mine negative reviews of mainstream dating apps, and constructed a two-stage machine learning model using data dimensionality reduction and text classification to classify user reviews of dating apps. The research results show that: firstly, the reasons for the current negative reviews of dating apps are mainly concentrated in the charging mechanism, fake accounts, subscription and advertising push mechanism and matching mechanism in the apps, proposed corresponding improvement suggestions are proposed by us; secondly, using principal component analysis to reduce the dimensionality of the text vector, and then using XGBoost model to learn the low-dimensional data after oversampling, a better classification accuracy of user reviews can be obtained. We hope These findings can help dating apps operators to improve services and achieve sustainable business operations of their apps.

1 Introduction

Maybe the term ‘online dating’ sounded weird in the 1990s, but today we have become accustomed to it. Mobile phones are ‘virtual bars’ in people’s pockets, allowing singles to socialize wherever they want. At least 200 million people worldwide use digital dating services every month, a study of Smith and Duggan [1] found that one in ten Americans has used online dating websites or mobile dating apps; sixty-six percent of online daters have met someone they know through dating websites or apps, and 23 percent have met spouses or long-term partners through these sites or apps. One of Statista’s report [2] claimed that in 2020, there would be 44.2 million online dating service users in United States. The company’s digital market outlook estimated that number will increase to 53.3 million by 2025. Due to the COVID-19 pandemic since 2020, many activities of people have shifted from offline to online. It has also led to a significant increase in the frequency of online dating app users using them. Chisom’s research work [3] suggested that loneliness and boredom due to adhering to the stay at home policy in the age of COVID-19, there was a rapid increase of online dating apps especially on Tinder and had in so many ways. In other words, dating apps have very good market prospects at present.

However, a good market prospect also means that there will be cruel competition among enterprises behind it. For operators of dating apps, one of the key factors in keeping their apps stable against the competitions or gaining more market share is getting positive reviews from as many users as possible. In order to achieve this goal, operators of dating apps should analyze the reviews of users from Google Play and other channels in a timely manner, and mine the main opinions reflected in the user reviews as an important basis for formulating apps’ improvement strategies. The study of Ye, Law and Gu [4] found significant relationship between online consumer reviews and hotel business performances. This conclusion can also be applied on apps. Noei, Zhang and Zou [5] claimed that for 77% of apps, taking into account the key content of user reviews when updating apps was significantly associated with an increase in ratings for newer versions of apps.

For user reviews of apps presented in a textual state, we believe that text mining models can be used to analyze these reviews. Some researchers such as M Lee, M Jeong and J Lee [6] have studied the impact of online user negative reviews on consumers’ choice when booking a hotel through text mining. Latent Dirichlet Allocation (LDA) was proposed by Blei et al. [7]. Since then, topic models based on LDA have become one of the key research areas of text mining. LDA is very widely used in the commercial fields. For example, Wahyudi and Kusumaningrum [8] have used an LDA-based topic model to perform sentiment analysis on user reviews of online shopping malls in Indonesia in their study.

Most of the sentences that people speak every day contain some kinds of emotions, such as happiness, satisfaction, anger, etc. We tend to analyze the emotions of sentences according to our experience of language communication. Feldman [9] thought that sentiment analysis is the task of finding the opinions of authors about specific entities. Operators of dating apps usually collect user feelings and opinions through questionnaires or other surveys within the websites or apps. For many customers’ opinions in the form of text collected in the surveys, it is obviously impossible for operators to use their own eyes and brains to watch and judge the emotional tendencies of the opinions one by one. Therefore, we believe that a feasible method is to first build a suitable model to fit the existing customer opinions that have been classified by sentiment tendency. In this way, the operators can then obtain the sentiment tendency of the newly collected customer opinions through batch analysis of the existing model, and conduct more in-depth analysis as needed.

At present, many machine learning and deep learning models can be used to analyze text sentiment which is processed by word segmentation. In the study of Abdulkadhar, Murugesan and Natarajan [10], LSA (Latent Semantic Analysis) was firstly used for feature selection of biomedical texts, then SVM (Support Vector Machines), SVR (Support Vactor Regression) and Adaboost were applied to the classification of biomedical texts. Their overall results show that AdaBoost performs better compared to two SVM classifiers. Sun et al. [11] proposed a text-information random forest model, which proposed a weighted voting mechanism to improve the quality of the decision tree in the traditional random forest for the problem that the quality of the traditional random forest is difficult to control, and it was proved that it can achieve better results in text classification. Aljedani, Alotaibi and Taileb [12] have explored the hierarchical multi-label classification problem in the context of Arabic and propose a hierarchical multi-label Arabic text classification (HMATC) model using machine learning methods. The results show that the proposed model was superior to all the models considered in the experiment in terms of computational cost, and its consumption cost is less than that of other evaluation models. Shah et al. [13] constructed a BBC news text classification model based on machine learning algorithms, and compared the performance of logistic regression, random forest and K-nearest neighbor algorithms on datasets. The results show that logistic regression classifier with the TF-IDF Vectorizer feature attains the highest accuracy of 97% for the data set. Jang et al. [14] have proposed an attention-based Bi-LSTM+CNN hybrid model that takes advantage of LSTM and CNN and has an additional attention mechanism. Testing results on Internet Movie Database (IMDB) movie review data showed that the newly proposed model produces more accurate classification results, as well as higher recall and F1 scores, than single multilayer perceptron (MLP), CNN or LSTM models and hybrid models. Lu, Pan and Nie [15] have proposed a VGCN-BERT model that combines the capabilities of BERT with a lexical graph convolutional network (VGCN). In their experiments with several text classification datasets, their proposed method outperformed BERT and GCN alone and was more effective than previous studies reported.

However, in practice when the text contains many words or the numbers of texts are large, the word vector matrix will obtain higher dimensions after word segmentation processing. Therefore, we should consider reducing the dimensions of the word vector matrix first. The research of Vinodhini and Chandrasekaran [16] showed that dimensionality reduction using PCA (principal component analysis) can make text sentiment analysis more effective. LLE (Locally Linear Embedding) is a manifold learning algorithm that can achieve effective dimensionality reduction for high-dimensional data. He et al. [17] believed that LLE is very effective in dimensionality reduction of text data.

Currently, there are fewer text mining studies on user reviews of apps that people use every day, but this field has caught the attention of researchers [18]. Much of the research on dating apps now focuses on psychology and sociology, with minority of studies looking at dating apps from a business perspective. The study by Ranzini, Rosenbaum and Tybur [19] found that Dutch people are more likely to choose Dutch people as potential partners when using dating apps, while Dutch people with higher education are more likely to choose potential partners with higher education backgrounds when using dating apps. Tran et al. [20] found that users of dating apps had significantly higher odds of unhealthy weight-control behaviors than those who had not used dating apps. Rochat et al. [21] used cluster analysis to study the characteristics of Tinder users. The results show that Tinder users participating in the study could be reasonably divided into four groups, and the users of each group were different in gender, marital status, depression and usage patterns. Tomaszewska and Schuster [22] compared perceptions related to sexuality of dating app users and non-dating app users, namely their risky sexual scripts and sexual self-esteem, and their risky and sexually assertive behaviors. Results showed that dating app users had more risky sexual scripts and reported more risky sexual behaviors than non-dating app users. In addition, male dating app users had lower sexual self-esteem and were more accepting of sexual coercion than male non-dating app users. Lenton et al. [23] studied the relationship between social anxiety and depressive symptoms of dating app users and their degree of dating app use, they found that dating app user social anxiety and depressive symptoms were positively correlated with their level of dating app use, and that these symptoms predicted that men were less likely to initiate contact with people matched by dating apps, but not women.

In some research work, researchers have proposed methods or tools to help operators of apps, websites, hotel etc. to analyze user reviews. Considering that user reviews for apps are valuable for app operators to improve user experience and user satisfaction, but manually analyzing large numbers of user reviews to get useful opinions is inherently challenging, Vu et al. [24] proposed MARK, a keyword-based semi-automated review analysis framework that can help app operators analyze user reviews more effectively to get useful input from users. Jha and Mahmoud [25] proposed a novel semantic approach for app review classification, it can be used to extract user needs from application evaluations, enabling a more efficient classification process and reducing the chance of overfitting. Dalal and Zaveri [26] proposed a view mining system for binary and fine-grained sentiment classification that can be used for user reviews, and empirical studies show that the proposed system can perform reliable sentiment classification at different granularity levels. Considering that a large number of user reviews need to be explored, analyzed, and organized to better assist website operators in making decisions, Sharma, Nigam and Jain [27] proposed an aspect-based opinion mining system to classify reviews, and empirically demonstrated the effectiveness of this system. Considering that hotel managers in Bali can gain insight into the perceived state of the hotel through hotel user reviews, Prameswari, Surjandari and Laoh [28] used text mining methods and aspect-based sentiment analysis in their research to capture hotel user opinions in the form of emotions. The results show that the Recursive Neural Tensor Network (RNTN) algorithm performs well in classifying the sentiment of words or aspects. As a result, we wish to applying machine learning models on mining user reviews of dating apps. In this way, operators of apps can better manage their user review data and improve their apps more effectively.

Considering the increasing popularity of dating apps and the unsatisfactory user reviews of major dating apps, we decided to analyze the user reviews of dating apps using two text mining methods. First, we established a topic model based on LDA to mine the negative reviews of mainstream dating apps, analyzed the main reasons why users give negative reviews, and put forward corresponding improvement suggestions. Next, we built a two-stage machine learning model that combined data dimensionality reduction and data classification, hoping to obtain a classification that can effectively classify user reviews of dating apps, so that app operators can process user reviews more effectively.

2 Data acquisition and research design

2.1 Data acquisition

At present, there are several dating apps that are widely used, such as the famous Tinder and Okcupid. Since most users download these apps from Google Play, we believed that app reviews on Google Play can effectively reflect user feelings and attitudes toward these apps. All the data we used are from reviews of users of these six dating apps: Bumble, Coffee Meets Bagel, Hinge, Okcupid, Plenty of Fish and Tinder. The data are published on figshare.com [29], we promise that sharing the dataset on Figshare complies with the terms and conditions of the sites from which data was accessed. Also, we promise that the methods of data collection used and its application in our study comply with the terms of the website from which the data originated. The data include the text of the reviews, the number of likes the reviews get, and the reviews’ ratings of the apps. At the end of May 2022, we have collected a total of 1,270,951 reviews data. First of all, in order to prevent the impact on the results of text mining, we first carried out text cleaning, deleted symbols, irregular words and emoji expressions, etc.

Considering that there may be some reviews from bots, fake accounts or meaningless duplicates among the many reviews, we believed that these reviews can be filtered by the number of likes they get. If a review has no likes, or just a few likes, it can be considered that the content contained in the review is not of sufficient value in the study of user reviews, because it can’t get enough commendations from other users. In order to keep the size of data we finally use not too small, and to ensure the authenticity of the reviews, we compared the two screening methods of retaining reviews with a number of likes greater than or equal to 5 and retaining reviews with a number of likes greater than or equal to 10. Among all the reviews, there are 25,305 reviews with 10 or more likes, and 42,071 reviews with 5 or more likes.

In order to maintain a certain generality and generalizability of the results of the topic model and classification model, it is considered that relatively more data is a better choice. Therefore, we selected 42,071 reviews with a relatively large sample size with a number of likes greater than or equal to 5. In addition, in order to ensure that there are no worthless comments in the filtered comments, such as repeated negative comments from robots, we randomly selected 500 comments for careful reading and found no obvious worthless comments in these reviews. For these 42,071 reviews, we plotted a pie chart of reviewers’ ratings of these apps, and the numbers such as 1,2 on the pie chart means 1 and 2 points for the app’s ratings.

Looking at Fig 1, we find that the 1-point rating, which represents the worst review, accounts for the majority of the reviews on these apps; while all of the percentages of other ratings are all less than 12% of the reviews. Such a ratio is very shocking. Most of the users who reviewed on Google Play were very dissatisfied with the dating apps they were using.

Fig 1. Distribution of ratings for dating apps.

Fig 1

2.2 Research design

Considering the disparity between positive and negative reviews in dating app user reviews, we believed that we need to establish a topic model based on LDA, hoping to mine the customer opinions contained in the negative reviews. At the same time, we thought it is helpful to build a text emotion classification model that can better match dating app user reviews with ratings in Google Play on a scale of 1 to 5. Such a model can help the operators of dating apps to quickly and effectively classify the text of user reviews when conducting market research or collecting various user opinions, so as to mine customers’ opinions which is reflected in positive, negative or moderate reviews with more details.

3 User negative review mining based on topic model

From the previous visualization results, we can clearly find that in the reviews of users on dating apps in Google Play, negative reviews account for the vast majority. This situation may have a certain impact on the operation of the apps, such as the loss of users or the continuous decline of the reputation of the apps. In order to solve such a problem, we have established a topic model using the implicit Dirichlet distribution to mine the information in the bad reviews of dating apps, in order to analyze the reasons behind the bad reviews and put forward corresponding suggestions that can solve the problem.

3.1 Topic model

With the explosive growth of text information, digging its theme composition from massive non-structured text information as the main mode of text information analysis. Topic Models are important type of machine learning algorithm that can discover potential mixed themes from the existing document set, so they are important unsupervised learning tools for inferring knowledge. Topic models are also universal ways to understand collections of unstructured text documents, they can be used to automate the screening of large amounts of text data. Once key topics have been identified, text files can be grouped for further analysis. There are many algorithms that can be applied to topic modeling, common ones are LSA, LDA (Latent Dirichlet Allocation), etc. In many research works that also use topic models such as [3033], the effectiveness of LSA and LDA may be difficult to determine Consider that LDA considers both the distribution of topics by document and the distribution of terms by topic compared to LSA [34], we have applied Latent Dirichlet assignments.

The topic model LDA is a non-supervised machine learning method. It can effectively extract the hidden themes in large-scale document sets and corpus libraries. The dimensional reduction capacity, modeling ability and scalability have made it one of the popular research directions in the field of theme mining in recent years.

3.2 Latent Dirichlet Allocation

Dirichlet distribution is a probability distribution of a diverse continuous random variable, which is an extension of Beta Distribution. In Bayesian learning, Dirichlet distribution is often used as a prior distribution of multinomial distributions. Definition: The probability density function of the multivariate continuous random variableis θ = (θ1, θ2, …, θn) is

p(θ|α)=Γ(i=0kαi)i=1kΓ(αi)i=1kθiαi-1 (1)
i=1kθi=1,θi0,αi0,

Be written as

θ(α)

The gamma function is [35]

Γ(s)=0xx-1e-xdx,s>0 (2)

The underlying Dirichlet assignment aims to represent each document T as a mixed distribution of topics, and each topic as a mixed distribution on dictionary D by modeling, where the input document set T and the number of topics S are given.

In addition to infering the topic distribution of words in a document, LDA makes assumptions about how topics and documents are generated. Let sigma represent the mixing ratio of words in S topics, a matrix of size S by D. The s-row of the matrix corresponds to the distribution σs(α) of words on topic S, which means that each σi is extracted from a Dirichlet distribution with a symmetric hyperparameter α. ϕ is the topic mixing ratio of the document and is a T * S matrix. Where ϕt in line T corresponds to the topic mixing ratio of document T, and ϕt(β) means that each ϕt is extracted from the Dirichlet distribution with symmetric hyperparameter β. When generating the n word in the document t, extract a topic according to the topic distribution ϕt of this document as the corresponding topic of this word, numbered ztn(ϕt) at first. Then, according to the word distribution ωtn of this topic, the word ωtn is extracted and numbered ωtn(σzm). A complete document can be generated by repeated multiple times, while the entire corpus can be generated by repeated extraction of documents [36].

3.3 Determination of hyperparameters

In order for the model to obtain better analysis results, we first need to determine the values of the hyperparameters of the topic model based on LDA. α and β are main hyperparameters, and we first considered the choice of hyperparameter β. As a common setting in Griffiths and Steyvers [37], we set β = 0.01. To the hyperparameter α, we also have considered the method to calculate this hyperparameter from Griffiths and Steyvers [37] that α = 50/t (where t is number of topics) as the baseline method. Since we consider 27,167 reviews with a score of 1 for analysis, the sample size is not very large, we believed the number of topics shouldn’t be greater than 50.

Topic coherence based on word co-occurrence patterns is one of the effective indicators to measure the effect of topic models. Therefore, we set different hyperparameters based on [37], obtained the topic coherence scores corresponding to these hyperparameters, and presented the results in the form of a line chart in Figs 2 and 3.

Fig 2. Line chart of hyperparametric search for topic models with less topics and high α.

Fig 2

Fig 3. Line chart of hyperparametric search for topic models with more topics and high α.

Fig 3

Figs 2 and 3 shows that the baseline coherence scores of the hyperparameters calculated according to the method proposed by Griffiths and Steyvers [37] are around 0.2 with all the choices of topics’ numbers. Since we don’t need to many topics, we considered smaller α. And the topic coherence scores corresponding to smaller α are shown in Figs 4 and 5.

Fig 4. Line chart of hyperparametric search for topic models with less topics and low α.

Fig 4

Fig 5. Line chart of hyperparametric search for topic models with more topics and low α.

Fig 5

According to the results showed in Figs 4 and 5, we believed that when the hyperparameter α = 0.01 and the number of topics is 10, the model can obtain the highest topic coherence score among the four figures. The final calculation results in topic coherence score = 0.469 is around twice the value of the baseline coherence scores. So we finally chose the LDA topic model with α = 0.01, β = 0.01 and 10 topics.

3.4 Analysis of the topic modeling results

Table 1 shows the analysis results of the LDA topic model with the final hyperparameters.

Table 1. Analysis results of the LDA topic model.

Topic No. Keywords Topic Summary
1 swipe, pay, match, gold money, useless, buy, ⋯ Matching system and payment
2 match, open, notification say, download, contact, ⋯ No notable problems
3 wrong, woman, mile, date distance, boost, ⋯ Long distance match women’s rights
4 super, old, ridiculous ad, block, ⋯ Ad push
5 login, error, code, photo, scam, ⋯ Login problem and scams from fake accounts
6 ban, spam, stupid, rate bot, write, report, ⋯ More effective communication
7 fake, profile, account, reply cancel terrible, ⋯ User authenticity and account issues
8 completely, system, help end, everytime, ⋯ Hope the apps are more functional
9 uninstalled, membership, install, screen, joke, face, ⋯ Issues with fake accounts and apps installation
10 pay, subscription, money, bad, charge, refund, purchase, ⋯ Dissatisfaction of payment and refund
  • Topic 1 shows the words ‘swipe’ ‘pay’ ‘match’ ‘gold’ ‘money’ ‘useless’ ‘buy’. There are complaints from users about the matching system and matters in payment.

  • Topic 2 shows the words ‘match’ ‘open’ ‘say’ ‘notification’ ‘download’ ‘contact’. This topic doesn’t focus on a certain problem.

  • Topic 3 shows the words ‘wrong’ ‘woman’ ‘mile’ ‘date’ ‘distance’ ‘boost’. We believe these words indicate user dissatisfaction with long distances matching and women’s rights in dating.

  • Topic 4 shows the words ‘super’ ‘old’ ‘ridiculous’ ‘ad’ ‘block’. These words reflect that users have great opinions on the advertising push mechanism of apps.

  • Topic 5 shows the words ‘login’ ‘error’ ‘code’ ‘photo’ ‘scam’. Users complained about errors when logging in and false information from other users and the resulting scams.

  • Topic 6 shows the words ‘spam’ ‘rate’ ‘write’ ‘report’. From these words, We believe that users expect more effective communication.

  • Topic 7 contains “fake’ ‘profile’ ‘account’ ‘reply’ ‘cancel’ ‘terrible’. Users want the authenticity of users on apps to be guaranteed and are concerned about issues with their accounts.

  • Topic 8 with words such as ‘completely’ ‘system’ ‘everytime’ “end’ ‘help’ reveals that users think the existing functions of dating apps are not complete, and they want dating apps to be more systematic and meet their needs at any time.

  • Topic 9 contains “uninstalled’ ‘membership’ ‘install’ ‘screen’ ‘joke’ ‘face’. Users have complained about fake accounts in dating apps and problems with app downloads and uninstalls.

  • Topic 10 with words like ‘pay’ ‘subscription’ ‘money’ ‘charge’ ‘refund’ ‘purchase’ ‘bad’. It mainly reflects the user’s dissatisfaction with matters related to payment and refund.

In general, the negative reviews of users are mainly concentrated in four aspects. First, they are dissatisfied with the payment amounts in dating apps and problems with the payment and refund processes; secondly, they complained about fake accounts in apps, including fake photos, personal information, and the resulting scams. This question is raised when people use many dating apps, Duguay [38] has pointed out that this problem exists in Tinder and has caused a certain negative impact; thirdly, they were dissatisfied with the advertising push mechanism and subscription mechanism in apps; finally, they also felt that the matching mechanism provided by apps needed to be improved.

4 Customer review classification based on two-stage machine learning

In order to establish a text sentiment classification model that can distinguish different reviews in a timely and effective manner, we first cleaned the comment text and perform term frequency–inverse document frequency (TF-IDF) processing on the data. The processed data are divided into training dataset and test dataset according to the ratio of 80% and 20%. Looking at the dimensionalities of the training data set, we find that the training dataset is higher than 25,000 dimensionalities, which is close to the number of samples in the training data set itself. We believed that the high dimensionality of the training dataset would interfere with the accuracy of the classification model. At the same time, as shown in Fig 1, the proportion of each category in the dataset is very uneven, which may also interfere with the accuracy of the classification model. In addition, we considered the methods of applying XGBoost and LightGBM to learn only oversampled data as baseline models to evaluate the effect of applying dimensionality reduction models and the final effect of two-stage models.

4.1 Data dimensionality reduction and oversampling

4.1.1 Principal component analysis

Principal Component Analysis (PCA) is a famous data dimensionality reduction model in multivariate statistical analysis. It extracts the amount of information contained in the original high-dimensional data by linear combination, which is called principal components. According to the variance contribution rate of the extracted principal components, the principal component analysis selects the amount of information required in practical applications to achieve the purpose of reducing the data dimension.

➀we normalize the data. For the matrix formed by the word vector:

X=[x11x12x1px21x22x2pxn1xn2xnp]

Then we can standardize the data according to formula 3:

xij*=xij-x^jVar(xj) (3)

With x^j=1ni=1nxij,Var(xJ)=1n-1i=1n(xij-x^j)2,(j=1,2,,p)

➁Let X* represent the matrix after data normalization, then the expression of the correlation coefficient matrix R between these data is:

R=[r11r12r1pr21r22r2prp1rp2rpp]

With rij=Cov(xi*,xj*)Var(x1*)Var(x2*),(n>1).

➂Calculate the eigenvalues(λ1, λ2, ⋯, λp) and eigenvectors of the correlation coefficient matrix R:

ai=(ai1,ai2,,aip)(i=1,2,,p) (4)

➃Next, we should select the required principal components and derive the expression for the principal components. The variance of each principal component obtained by the principal component analysis and the information contained in the principal component are decreasing from large to small. We follow the variance contribution rate of each principal component:

variancecontributionrate=λii=1pλi (5)

Here we select the principal components with variance contribution rate in the top 90% to build a new machine learning classification model.

➄Finally, calculate the specific values of the principal components we need. Substitute the normalized data for calculating the principal components:

Fij=aj1xi1*+aj2xi2*++ajpxip* (6)
(i=1,2,,n;j=1,2,,k)

Finally, we get the required dimensionality-reduced data, which is expressed in matrix form as [39]:

[F11F12F1kF21F22F2kFp1Fp2Fpp]

4.1.2 Locally linear embedding

Locally Linear Embedding (LLE) is a method of data dimensionality reduction for the data of nonlinear signal feature vector dimensions. The idea of this dimensionality reduction is not just to reduce the number of dimensions, but to maintain the mathematical properties of the original data. Under the condition that the characteristics remain unchanged, LLE maps the data in the high-dimensional space to the low-dimensional space, and realizes the second extraction of eigenvalues in the data.

For each training instance x(i), LLE analyzes the k nearest neighbors of the instance, and then tries to reconstruct x(i) as a linear combination of these neighbors. More specifically, LLE makes the Euclidean distance between x(i) and j=1mwi,jx(j) as small as possible by finding the weight wi,j. Meanwhile LLE assumes wi,j = 0 if x(j) is not one of the k nearest neighbors of x(i). Therefore, LLE first constructs an unconstrained optimization, where W is a weight matrix containing all weights wi,j, and the second constraint requires that the weights of each training data x(i) be normalized.

W^=argmini=1m(x(i)-j=1mwi,jx(j))2 (7)
s.t.{wi,j=0,x(j)c.n.ofx(i)j=1mwi,j=1,i=1,2,,m (8)

Then, the weight matrix W^ (containing the weights w^i,j) will encode the local linear relationship between the training data.

Next, LLE will map the training data to a d-dimensional space (where dn), while preserving these local relationships as much as possible. Considering that z(i) is the dimensionality reduction result of x(i) in this d-dimensional space, we hope that the Euclidean distance between z(i) and j=1mW^i,jz(j) is as small as possible, thereby updating the previous objective function. LLE doesn’t take the approach of finding the best weights while keeping the data fixed here. In contrast, LLE keeps the weights constant and finds the best location for the data in a low-dimensional space. Here Z is a matrix containing all z(i). [40]

Z^=argmini=1m(z(i)-j=1mW^i,jx(j))2 (9)

4.1.3 Oversampling

Before building a classifier to fit the dimensionality-reduced data, we should pay attention to the serious imbalance of classification labels in the data. It can be intuitively found from the previous pie chart Fig 1 that the number of samples with classification labels 2, 3 and 4 is much less than the number of samples with labels 1 and 5. In general, such a situation can easily affect the performance of the classification model. The research of Chawla et al. [41] proposed the SMOTE oversampling method and confirmed that this method can avoid overfitting caused by traditional oversampling. In order to solve this problem, we have used the SMOTE method to oversample the data when building the classification model, so that the sample size corresponding to each classification label tended to be balanced.

4.2 Classification models and results

4.2.1 Model introduction

Both XGBoost and LightGBM are powerful ensemble learning models further improved on the basis of Gradient Boosting Decision Tree (GBDT). XGBoost was proposed by Chen and Guestrin [42]. Compared with GBDT, XGBoost adds a regularization part to the loss function to achieve better generalization ability. At the same time, the loss function of XGBoost performs second-order Taylor expansion on the error part, which is more accurate than the first-order Taylor expansion on the error part in GBDT. XGBoost performs parallel selection on the establishment process of each weak learner in the algorithm, which improves the efficiency of the algorithm. Afifah [43] performed sentiment analysis on reviews of a telemedicine application in Indonesia using XGBoost and achieved over 96% classification accuracy. In Zhang’s research [44], XGBoost can perform the task of criminal text classification more effectively than other models.

LightGBM was proposed by Ke et al. [45]. Its internal decision trees use a leaf-wise growth strategy and sets the max_depth hyperparameter to prevent overfitting. LightGBM also uses the histogram algorithm. According to the data binning strategy, the nodes of the decision tree can improve the calculation speed when splitting. The two learning methods it supports feature parallelism and data parallelism further reduce the computational cost. In addition, the Gradient-based One-Side Sampling and Exclusive Feature Bundling algorithms included in LightGBM have both contributed to improving the classification accuracy. He et al. [46] believed that the LightGBM works well on social media prediction tasks. And the research of Zvonarev and Bilyi [47] showed that LightGBM has good performance in sentiment analysis of Russian texts.

4.2.2 Model evaluation

The results of our model are presented in Table 2. Data dimensionality reduction and oversampling using principal component analysis did not improve the classification effect of the LightGBM model much. But for the XGBoost model with its hyperparameters max depth = 10, n estimators = 50 and learning rate = 0.1, such data processing improved the accuracy of the classification model to 0.883. After data dimensionality reduction of PCA and oversampling, the areas under the ROC curve of the LightGBM and XGBoost models at Figs 6 and 7 perform better on each class. In particular, the area under the ROC curve of the XGBoost model exceeds 0.9 for each class. The confusion matrices in Figs 8 and 9 are also strong evidence that the XGBoost model achieves good results.

Table 2. Model accuracy before and after oversampling.
XGBoost LightGBM
Original 0.697 0.688
PCA and oversampling 0.883 0.725
LLE and oversampling 0.718 0.701
Fig 6. ROC curve for PCA and LightGBM.

Fig 6

Fig 7. ROC curve for PCA and XGBoost.

Fig 7

Fig 8. Confusion matrix curve for PCA and LightGBM.

Fig 8

Fig 9. Confusion matrix for PCA and XGBoost.

Fig 9

On the other hand, 0.883 is close to 0.19 more accurate than using XGBoost’s baseline model. Moreover, as the data has a total of 5 categories was considered, we believed that this level of accuracy is high enough to be acceptable. We suggest that dating app operators can use this method to establish an XGBoost model on the data after dimensionality reduction by principal component analysis and oversampling, so as to facilitate batch text sentiment classification for customer opinions collected in the future.

5 Conclusion and outlook

Regarding the problems analyzed from the user’s negative reviews on the current mainstream dating apps, we make the following suggestions for the operators of the apps:

At first, we suggest that the operators of apps can conduct a questionnaire survey on users to understand the user’s expectations for the payment amount, and may try to charge according to the user’s needs. At the same time, the current payment and refund system should be improved to ensure the timely completion of payment and refund for users.

Secondly, we recommend that operators of apps develop more effective ways to identify false information about users. For example, the authenticity of user information can be analyzed by combining multi-modal data such as user photos, IP addresses, and login times.

Then, we suggest that app operators improve the current advertising push mechanism and subscription mechanism, and they can develop more effective recommendation system algorithms to meet different user needs.

Futhermore, we recommend that the operators of apps improve the dating object matching mechanism, which can start from the distance between users, the ages and living habits of users, etc., and also increase the dominance of female users in matching.

Finally, we believe that it is more effective to use principal component analysis to reduce the dimensionality of the text vectors of user reviews of dating apps, and then use the XGBoost model to learn the oversampled low-dimensional data. Such a two-stage machine learning model can efficiently classify these reviews so that operators of apps can batch process the collected user review data.

Supporting information

S1 Graphical abstract

(TIF)

Data Availability

All the data are available on figshare.com (https://figshare.com/articles/dataset/Text_of_user_reviews_of_dating_apps/21895827).

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Smith AW, Duggan M. Online dating & relationship. Pew Research Center Washington, DC; 2013.
  • 2.Statista. Number of online dating users in the United States from 2017 to 2024;. https://www.statista.com/statistics/417654/us-online-dating-user-numbers/.
  • 3. Chisom OB, et al. Effects of modern dating applications on healthy offline intimate relationships during the COVID-19 pandemic: A review of the tinder dating application. Advances in Journalism and Communication. 2021;9(01):12. doi: 10.4236/ajc.2021.91002 [DOI] [Google Scholar]
  • 4. Ye Q, Law R, Gu B. The impact of online user reviews on hotel room sales. International Journal of Hospitality Management. 2009;28(1):180–182. doi: 10.1016/j.ijhm.2008.06.011 [DOI] [Google Scholar]
  • 5. Noei E, Zhang F, Zou Y. Too many user-reviews! What should app developers look at first? IEEE Transactions on Software Engineering. 2019;47(2):367–378. doi: 10.1109/TSE.2019.2893171 [DOI] [Google Scholar]
  • 6. Lee M, Jeong M, Lee J. Roles of negative emotions in customers’ perceived helpfulness of hotel reviews on a user-generated review website: A text mining approach. International Journal of Contemporary Hospitality Management. 2017;. doi: 10.1108/IJCHM-10-2015-0626 [DOI] [Google Scholar]
  • 7. Blei DM, Ng AY, Jordan MI. Latent dirichlet allocation. Journal of machine Learning research. 2003;3(Jan):993–1022. [Google Scholar]
  • 8.Wahyudi E, Kusumaningrum R. Aspect based sentiment analysis in E-commerce user reviews using Latent Dirichlet Allocation (LDA) and Sentiment Lexicon. In: 2019 3rd International Conference on Informatics and Computational Sciences (ICICoS). IEEE; 2019. p. 1–6.
  • 9. Feldman R. Techniques and applications for sentiment analysis. Communications of the ACM. 2013;56(4):82–89. doi: 10.1145/2436256.2436274 [DOI] [Google Scholar]
  • 10. Abdulkadhar S, Murugesan G, Natarajan J. Classifying protein-protein interaction articles from biomedical literature using many relevant features and context-free grammar. J King Saud Univ Comput Inf Sci. 2017;32:553–560. [Google Scholar]
  • 11.Sun Y, Li Y, Zeng Q, Bian Y. Application Research of Text Classification Based on Random Forest Algorithm. 2020 3rd International Conference on Advanced Electronic Materials, Computers and Software Engineering (AEMCSE). 2020; p. 370–374.
  • 12. Aljedani N, Alotaibi RM, Taileb M. HMATC: Hierarchical multi-label Arabic text classification model using machine learning. Egyptian Informatics Journal. 2020;. [Google Scholar]
  • 13. Shah K, Patel H, Sanghvi DJ, Shah M. A Comparative Analysis of Logistic Regression, Random Forest and KNN Models for the Text Classification. Augmented Human Research. 2020;5:1–16. doi: 10.1007/s41133-020-00032-0 [DOI] [Google Scholar]
  • 14. Jang B, Kim M, Harerimana G, Kang S, Kim JW. Bi-LSTM Model to Increase Accuracy in Text Classification: Combining Word2vec CNN and Attention Mechanism. Applied Sciences. 2020;. [Google Scholar]
  • 15. Lu Z, Du P, Nie J. VGCN-BERT: Augmenting BERT with Graph Embedding for Text Classification. Advances in Information Retrieval. 2020;12035:369–382. doi: 10.1007/978-3-030-45439-5_25 [DOI] [Google Scholar]
  • 16.Vinodhini G, Chandrasekaran R. Sentiment classification using principal component analysis based neural network model. In: International Conference on Information Communication and Embedded Systems (ICICES2014). IEEE; 2014. p. 1–6.
  • 17.He C, Dong Z, Li R, Zhong Y. Dimensionality reduction for text using LLE. In: 2008 International Conference on Natural Language Processing and Knowledge Engineering. IEEE; 2008. p. 1–7.
  • 18. Genc-Nayebi N, Abran A. A systematic literature review: Opinion mining studies from mobile app store user reviews. Journal of Systems and Software. 2017;125:207–219. doi: 10.1016/j.jss.2016.11.027 [DOI] [Google Scholar]
  • 19. Ranzini G, Rosenbaum JE, Tybur JM. Assortative (online) dating: Insights into partner choice from an experimental dating app. Computers in Human Behavior. 2022;127:107039. doi: 10.1016/j.chb.2021.107039 [DOI] [Google Scholar]
  • 20. Tran A, Suharlim C, Mattie H, Davison K, Agénor M, Austin SB. Dating app use and unhealthy weight control behaviors among a sample of US adults: A cross-sectional study. Journal of Eating Disorders. 2019;7(1):1–12. doi: 10.1186/s40337-019-0244-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Rochat L, Bianchi-Demicheli F, Aboujaoude E, Khazaal Y. The psychology of “swiping”: A cluster analysis of the mobile dating app Tinder. Journal of Behavioral Addictions. 2019;8(4):804–813. doi: 10.1556/2006.8.2019.58 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Tomaszewska P, Schuster I. Comparing sexuality-related cognitions, sexual behavior, and acceptance of sexual coercion in dating app users and non-users. Sexuality research and social policy. 2020;17(2):188–198. doi: 10.1007/s13178-019-00397-x [DOI] [Google Scholar]
  • 23. Lenton-Brym AP, Santiago VA, Fredborg BK, Antony MM. Associations between social anxiety, depression, and use of mobile dating applications. Cyberpsychology, Behavior, and Social Networking. 2021;24(2):86–93. doi: 10.1089/cyber.2019.0561 [DOI] [PubMed] [Google Scholar]
  • 24.Vu PM, Nguyen TT, Pham HV, Nguyen TT. Mining User Opinions in Mobile App Reviews: A Keyword-Based Approach (T). 2015 30th IEEE/ACM International Conference on Automated Software Engineering (ASE). 2015; p. 749–759.
  • 25.Jha N, Mahmoud A. Mining User Requirements from Application Store Reviews Using Frame Semantics. In: Requirements Engineering: Foundation for Software Quality; 2017.
  • 26. Dalal MK, Zaveri MA. Opinion Mining from Online User Reviews Using Fuzzy Linguistic Hedges. Appl Comput Intell Soft Comput. 2014;2014:735942:1–735942:9. [Google Scholar]
  • 27.Sharma R, Nigam S, Jain R. Mining of product reviews at aspect level. ArXiv. 2014;abs/1406.3714.
  • 28.Prameswari P, Surjandari I, Laoh E. Opinion mining from online reviews in Bali tourist area. 2017 3rd International Conference on Science in Information Technology (ICSITech). 2017; p. 226–230.
  • 29.Shen Q. Text of user reviews of dating apps. 2023. [DOI] [PMC free article] [PubMed]
  • 30.Tseng SC, Lu Y, Chakraborty G, Chen LS. Comparison of Sentiment Analysis of Review Comments by Unsupervised Clustering of Features Using LSA and LDA. 2019 IEEE 10th International Conference on Awareness Science and Technology (iCAST). 2019; p. 1–6.
  • 31.Bellaouar S, Bellaouar MM, Ghada IE. Topic Modeling: Comparison of LSA and LDA on Scientific Publications. 2021 4th International Conference on Data Storage and Data Engineering. 2021;.
  • 32.Kakkonen T, Myller N, Sutinen E. Applying Latent Dirichlet Allocation to Automatic Essay Grading. In: FinTAL; 2006.
  • 33.Williams T, Betak JF. A Comparison of LSA and LDA for the Analysis of Railroad Accident Text. In: ANT/SEIT; 2019.
  • 34.Lee S, Baker J, Song J, Wetherbe JC. An Empirical Comparison of Four Text Mining Methods. 2010 43rd Hawaii International Conference on System Sciences. 2010; p. 1–10.
  • 35.Ng KW, Tian GL, Tang ML. Dirichlet and related distributions: Theory, methods and applications. 2011;.
  • 36. Jelodar H, Wang Y, Yuan C, Feng X, Jiang X, Li Y, et al. Latent Dirichlet allocation (LDA) and topic modeling: models, applications, a survey. Multimedia Tools and Applications. 2019;78(11):15169–15211. doi: 10.1007/s11042-018-6894-4 [DOI] [Google Scholar]
  • 37. Griffiths TL, Steyvers M. Finding scientific topics. Proceedings of the National academy of Sciences. 2004;101(suppl_1):5228–5235. doi: 10.1073/pnas.0307752101 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Duguay S. Dressing up Tinderella: Interrogating authenticity claims on the mobile dating app Tinder. Information, communication & society. 2017;20(3):351–367. doi: 10.1080/1369118X.2016.1168471 [DOI] [Google Scholar]
  • 39. Hastie T, Tibshirani R, Friedman JH, Friedman JH. The elements of statistical learning: data mining, inference, and prediction. vol. 2. Springer; 2009. [Google Scholar]
  • 40.Géron A. Hands-on machine learning with Scikit-Learn, Keras, and TensorFlow: Concepts, tools, and techniques to build intelligent systems. “O’Reilly Media, Inc.”; 2019.
  • 41. Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP. SMOTE: synthetic minority over-sampling technique. Journal of artificial intelligence research. 2002;16:321–357. doi: 10.1613/jair.953 [DOI] [Google Scholar]
  • 42.Chen T, Guestrin C. Xgboost: A scalable tree boosting system. In: Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining; 2016. p. 785–794.
  • 43.Afifah K, Yulita IN, Sarathan I. Sentiment Analysis on Telemedicine App Reviews using XGBoost Classifier. In: 2021 International Conference on Artificial Intelligence and Big Data Analytics. IEEE; 2021. p. 22–27.
  • 44.Qi Z. The text classification of theft crime based on TF-IDF and XGBoost model. In: 2020 IEEE International Conference on Artificial Intelligence and Computer Applications (ICAICA). IEEE; 2020. p. 1241–1246.
  • 45. Ke G, Meng Q, Finley T, Wang T, Chen W, Ma W, et al. Lightgbm: A highly efficient gradient boosting decision tree. Advances in neural information processing systems. 2017;30. [Google Scholar]
  • 46.He Z, He Z, Wu J, Yang Z. Feature construction for posts and users combined with lightgbm for social media popularity prediction. In: Proceedings of the 27th ACM International Conference on Multimedia; 2019. p. 2672–2676.
  • 47.Zvonarev A, Bilyi A. A comparison of machine learning methods of sentiment analysis based on Russian language twitter data. In: The 11th Majorov International Conference on Software Engineering and Computer Systems; 2019.

Decision Letter 0

Viacheslav Kovtun

4 Dec 2022

PONE-D-22-30174Analysis of Dating Apps' User Reviews Based on Text MiningPLOS ONE

Dear Dr. Chen,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jan 18 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Viacheslav Kovtun, Dr.Sc., Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. In your Methods section, please include additional information about your dataset and ensure that you have included a statement specifying whether the collection method of the collected dataset, and its use for research, complied with the terms and conditions for the website.

3. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code and data that underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse. New software must comply with the Open Source Definition. If data cannot be openly shared for ethical or legal restrictions, please provide details for accessing these data from the original data holder.

4. Thank you for stating the following in the Acknowledgments Section of your manuscript: 

"The work was financialy supported by Chinese National Natural Science Foundation with Grant No.12101473, the Basic Research Program of Natural Science in Shaanxi Province with Grant Nos.2021JQ-764 and 2021JQ-766, and Education Department of Shaanxi Province with Grant No.21JK0640."

We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. 

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows: 

"Xi Chen was financialy supported by Chinese National Natural Science Foundation with Grant No.12101473, the Basic Research Program of Natural Science in Shaanxi Province with Grant Nos.2021JQ-764 and 2021JQ-766, and Education Department of Shaanxi Province with Grant No.21JK0640.

The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript."

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

5. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

""Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access.

We will update your Data Availability statement to reflect the information you provide in your cover letter.

6. PLOS requires an ORCID iD for the corresponding author in Editorial Manager on papers submitted after December 6th, 2016. Please ensure that you have an ORCID iD and that it is validated in Editorial Manager. To do this, go to ‘Update my Information’ (in the upper left-hand corner of the main menu), and click on the Fetch/Validate link next to the ORCID field. This will take you to the ORCID site and allow you to create a new iD or authenticate a pre-existing iD in Editorial Manager. Please see the following video for instructions on linking an ORCID iD to your Editorial Manager account: https://www.youtube.com/watch?v=_xcclfuvtxQ

7. We note you have included a table to which you do not refer in the text of your manuscript. Please ensure that you refer to Table 1 in your text; if accepted, production will need this reference to link the reader to the Table. 

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: No

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: It's possible that I'll suggest giving this article a presentation at a conference so that You may have an audience and continue the conversation about the issues. My primary justification is on the fact that "users" are the "unit of analysis" linked with this study; hence, there needs to be a concrete subjective evaluation to generate an accurate construct that is suitable for quantifying the impact of "User Reviews." However, while employing machine learning, the assumption is that a new course of action "or knowledge" can emerge from the analysis of the machine learning, sadly the work has not showed any improvement on the problems linked with "User Reviews" on dating apps. Generally speaking, user-based aspects used in this study is faulty.

You don't need any sort of conceptualised form of an advanced thought in order to gain something new through "Negative Review Mining." Instead, all you need is to have an open mind.

To begin, you have to get a firm grasp on the fact that ratings and reviews are inextricably linked, and it is your responsibility to ensure that this connection is maintained.

The users review and rating is important because it has led to understanding of the rating approached mapped to reviews.

Rating is numeric and reviews are reflections. The numeric values of rating will indicate if DATIN APPS is Good or BAD, whereas reviews will indicate the user’s perceptions. Previous research reveals that “Bad ratings are trustworthy regardless of the number of reviews”, that is users tend to believe reported bad rating. On the other hand, “Good ratings are trustworthy only when they come along with a high number of reviews”

The proposed four facets of "bad reviews of users" are not utilised in any novel way by Section 3, which does not produce anything that is formulated.

The fourth section is a routine analysis, and it does not introduce anything novel that contributes to the advancement of the research field. The XGBoost and LightGBM models are applied to the dataset that is already in existence without any additional information being connected to them.

,

Reviewer #2: In this paper, the authors are proposed “Analysis of Dating Apps’ User Reviews Based on Text Mining”.

The strengths of the paper are that it is well structured, the description of the related work is well done and that results are extensively compared to results of the similar research.

Minor revisions:

1. Authors should draw a graphical abstract of the proposed approach

2. Authors should mention the names of the all the six dating apps

3. Why authors used LDA instead of LSA.

4. Why authors used XGBoost justify it

5. Proofread the entire manuscript

6. Authors follow the recently published paper “Abdulkadhar, S., Murugesan, G., & Natarajan, J. (2020). Classifying protein-protein interaction articles from biomedical literature using many relevant features and context-free grammar. Journal of King Saud University-Computer and Information Sciences, 32(5), 553-560.”

7. Proofread the entire manuscript.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2023 Apr 26;18(4):e0283896. doi: 10.1371/journal.pone.0283896.r002

Author response to Decision Letter 0


14 Feb 2023

#Responses to Academic Editor:

Thank you very much for your guidance and comments on our work. Please find our itemized responses below and our corrections in the re-submitted files.

1. Comments: Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf.

Response: We used the latex template provided by PLOS ONE for our writing and followed the format of the paper required by PLOS ONE.

2. Comments: In your Methods section, please include additional information about your dataset and ensure that you have included a statement specifying whether the collection method of the collected dataset, and its use for research, complied with the terms and conditions for the website.

Response: We have uploaded the dataset to figshare.com and made it publicly available, referenced the dataset in the Data Acquisition section of the article, and added a statement of compliance with data collection. The URI of our data is:

https://figshare.com/articles/dataset/Text_of_user_reviews_of_dating_apps/21895827

3. Comments: Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code and data that underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse. New software must comply with the Open Source Definition. If data cannot be openly shared for ethical or legal restrictions, please provide details for accessing these data from the original data holder.

Response: The code we used has been publicly posted on GitHub, and the URL of the code (https://github.com/Qian0214Shen/code-of-text-mining-paper) has been filled in the relevant field when submitted.

4. Comments: Thank you for stating the following in the Acknowledgments Section of your manuscript:

"The work was financialy supported by Chinese National Natural Science Foundation with Grant No.12101473, the Basic Research Program of Natural Science in Shaanxi Province with Grant Nos.2021JQ-764 and 2021JQ-766, and Education Department of Shaanxi Province with Grant No.21JK0640."

We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

"Xi Chen was financialy supported by Chinese National Natural Science Foundation with Grant No.12101473, the Basic Research Program of Natural Science in Shaanxi Province with Grant Nos.2021JQ-764 and 2021JQ-766, and Education Department of Shaanxi Province with Grant No.21JK0640.

The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript."

Please include your amended statements within your cover letter; we will change the online submission form on your behalf..

Response: We have removed the section on funding information from the paper.

5. Comments: In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

""Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access.

We will update your Data Availability statement to reflect the information you provide in your cover letter.

Response: We have uploaded the dataset to figshare.com and made it publicly available, referenced the dataset in the Data Acquisition section of the article, and added a statement of compliance with data collection. The URI of our data is:

https://figshare.com/articles/dataset/Text_of_user_reviews_of_dating_apps/21895827

6. Comments: PLOS requires an ORCID iD for the corresponding author in Editorial Manager on papers submitted after December 6th, 2016. Please ensure that you have an ORCID iD and that it is validated in Editorial Manager. To do this, go to ‘Update my Information’ (in the upper left-hand corner of the main menu), and click on the Fetch/Validate link next to the ORCID field. This will take you to the ORCID site and allow you to create a new iD or authenticate a pre-existing iD in Editorial Manager. Please see the following video for instructions on linking an ORCID iD to your Editorial Manager account: https://www.youtube.com/watch?v=_xcclfuvtxQ

Response: We have added ORCID information for the corresponding author in her account.

7. Comments: We note you have included a table to which you do not refer in the text of your manuscript. Please ensure that you refer to Table 1 in your text; if accepted, production will need this reference to link the reader to the Table

Response: We have already referenced Table 1 below it.

#Responses to Reviewer 1:

Thank you very much for taking your time to review this manuscript. Please find our itemized responses below.

1. Comments:  It's possible that I'll suggest giving this article a presentation at a conference so that You may have an audience and continue the conversation about the issues. My primary justification is on the fact that "users" are the "unit of analysis" linked with this study; hence, there needs to be a concrete subjective evaluation to generate an accurate construct that is suitable for quantifying the impact of "User Reviews." However, while employing machine learning, the assumption is that a new course of action "or knowledge" can emerge from the analysis of the machine learning, sadly the work has not showed any improvement on the problems linked with "User Reviews" on dating apps. Generally speaking, user-based aspects used in this study is faulty.

Response: The starting point of our research is based on the perspective of enterprise information management, that is, by mining the user reviews of the apps, analyzing how the operators of the apps can improve the apps based on the opinions of users, and trying to develop a method for the operators of the apps to quickly classify the unmarked user reviews collected. Regarding the impact of quantifying user reviews that you mentioned, this may require a large-scale market research to understand how a large number of users perceive these reviews in Google Play and how these reviews influence user choices. In fact, this is indeed a very interesting and exciting topic, and we hope to have the opportunity to collect more suitable data for such research in further research.

2. Comments: You don't need any sort of conceptualised form of an advanced thought in order to gain something new through "Negative Review Mining." Instead, all you need is to have an open mind.

Response: We strongly agree that an open mind is indeed a necessary condition for people to accept bad reviews, but in the current era of big data, it is difficult for app operators to rely only on an open mind to mine information from a large number of reviews. Therefore, we hope to efficiently mine information from massive user reviews by applying machine learning models.

3. Comments: Rating is numeric and reviews are reflections. The numeric values of rating will indicate if DATIN APPS is Good or BAD, whereas reviews will indicate the user’s perceptions. Previous research reveals that “Bad ratings are trustworthy regardless of the number of reviews”, that is users tend to believe reported bad rating. On the other hand, “Good ratings are trustworthy only when they come along with a high number of reviews”

Response: In the current reviews of apps, malicious low-scoring reviews or worthless reviews from bots are always inevitable, and these reviews are difficult to express the general thoughts of app users. There are other comments, perhaps due to their short publication time, that do not receive enough likes, and the value of these comments is difficult to determine in batches. Therefore, in order to control the mining value of the data and ensure that the size of data in the dataset is not too small to affect the fit of the machine learning model, we select comments with more than or equal to 5 likes for analysis.

4. Comments: The fourth section is a routine analysis, and it does not introduce anything novel that contributes to the advancement of the research field. The XGBoost and LightGBM models are applied to the dataset that is already in existence without any additional information being connected to them.

Response: We should admit that the fourth part of our study is only an application to existing methods. First of all, LightGBM and XGBoost are actually very good machine learning classification models and are widely used in machine learning research in various fields. And in the process of application, we obtained 88.3% good accuracy for machine learning classification tasks with 5 classes, which is close to 19% higher than our baseline model, so we think this result is acceptable.

#Responses to Reviewer 2:

Thank you very much for your guidance and comments on our work. Please find our itemized responses below and our corrections in the re-submitted files.

1. Comments: Authors should draw a graphical abstract of the proposed approach

Response: We have produced the Graphic Abstract and submitted it as an attachment.

2. Comments: Authors should mention the names of the all the six dating apps

Response: We have declared the names of 6 apps in the section of data acquisition in Page 4.

3. Comments: Why authors used LDA instead of LSA

Response: Comparing some previous studies, we think it is difficult to say which is better than LSA or LDA. But considering that LDA considers both the distribution of topics by document and the distribution of topics over LSA, we chose LDA, as detailed at the end of the first paragraph of section 3.1.

4. Comments: Why authors used XGBoost justify it

Response: In the process of application XGBoost, we obtained 88.3% good accuracy for machine learning classification tasks with 5 classes, which is close to 19% higher than our baseline model, so we think this result is acceptable.

5. Comments: Proofread the entire manuscript

Response: We have proofread the full text and corrected some grammatical errors.

6. Comments: Authors follow the recently published paper “Abdulkadhar, S., Murugesan, G., & Natarajan, J. (2020). Classifying protein-protein interaction articles from biomedical literature using many relevant features and context-free grammar. Journal of King Saud University-Computer and Information Sciences, 32(5), 553-560.”

Response: We have already referred to the paper you mentioned and cited it in the introduction section.

Attachment

Submitted filename: Response to reviewers.pdf

Decision Letter 1

Viacheslav Kovtun

20 Mar 2023

User Review Analysis of Dating Apps based on Text Mining

PONE-D-22-30174R1

Dear Dr. Chen,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Viacheslav Kovtun, Dr.Sc., Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: After giving the piece my complete attention, I can say that the authors have addressed each and every one of my concerns.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

Acceptance letter

Viacheslav Kovtun

30 Mar 2023

PONE-D-22-30174R1

User Review Analysis of Dating Apps based on Text Mining

Dear Dr. Chen:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Professor Viacheslav Kovtun

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Graphical abstract

    (TIF)

    Attachment

    Submitted filename: Response to reviewers.pdf

    Data Availability Statement

    All the data are available on figshare.com (https://figshare.com/articles/dataset/Text_of_user_reviews_of_dating_apps/21895827).


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES