Skip to main content
PLOS One logoLink to PLOS One
. 2024 Mar 4;19(3):e0298977. doi: 10.1371/journal.pone.0298977

Don’t judge a book or health app by its cover: User ratings and downloads are not linked to quality

Maciej Hyzy 1,2,*, Raymond Bond 1, Maurice Mulvenna 1, Lu Bai 3, Anna-Lena Frey 2, Jorge Martinez Carracedo 1, Robert Daly 2, Simon Leigh 2,4
Editor: Kehinde Kazeem Kanmodi5
PMCID: PMC10911617  PMID: 38437233

Abstract

Objective

To analyse the relationship between health app quality with user ratings and the number of downloads of corresponding health apps.

Materials and methods

Utilising a dataset of 881 Android-based health apps, assessed via the 300-point objective Organisation for the Review of Care and Health Applications (ORCHA) assessment tool, we explored whether subjective user-level indicators of quality (user ratings and downloads) correlate with objective quality scores in the domains of user experience, data privacy and professional/clinical assurance. For this purpose, we applied spearman correlation and multiple linear regression models.

Results

For user experience, professional/clinical assurance and data privacy scores, all models had very low adjusted R squared values (< .02). Suggesting that there is no meaningful link between subjective user ratings or the number of health app downloads and objective quality measures. Spearman correlations suggested that prior downloads only had a very weak positive correlation with user experience scores (Spearman = .084, p = .012) and data privacy scores (Spearman = .088, p = .009). There was a very weak negative correlation between downloads and professional/clinical assurance score (Spearman = -.081, p = .016). Additionally, user ratings demonstrated a very weak correlation with no statistically significant correlations observed between user ratings and the scores (all p > 0.05). For ORCHA scores multiple linear regression had adjusted R-squared = -.002.

Conclusion

This study highlights that widely available proxies which users may perceive to signify the quality of health apps, namely user ratings and downloads, are inaccurate predictors for estimating quality. This indicates the need for wider use of quality assurance methodologies which can accurately determine the quality, safety, and compliance of health apps. Findings suggest more should be done to enable users to recognise high-quality health apps, including digital health literacy training and the provision of nationally endorsed “libraries”.

Introduction

According to a report from 2021 there were more than 350,000 health apps available in the iOS and Android stores, with an estimated 250 health apps added every day [1]. Moreover, searches for digital health products within app stores have also increased [2]. A potential catalyst for this could have been the COVID-19 pandemic and restricted access to incumbent services. Nevertheless, these findings clearly indicate that the public has an interest in health apps.

However, given the large number of health apps on offer, it can be difficult for users to identify high-quality apps that meet their needs. Notably, selecting the low-quality app can be associated with substantial opportunity costs and/or risks. For example, a systematic assessment of suicide prevention and deliberate self-harm mobile health apps found that some apps encouraged risky behaviours such as the uptake of drugs [3]. Moreover, reviews across different disease areas have shown that many health apps do not comply with data privacy, sharing, and security standards [47], have safety concerns [8], provide incomplete or misleading medical information [9,10], lack evidence-based components [11], and/or have not been supported by efficacy/effectiveness studies [5,6,12]. Also, health experts have largely avoided formally recommending apps, which forces users to obtain recommendations from other sources [13]. Therefore, if not sufficiently informed, user’s app choices can result in poor health benefits if ineffective apps are chosen and/or pose significant risks to user’s health and privacy.

Notably, in the absence of guidance, users are likely to select health apps based on metrics that they perceive to be proxies for quality, such as prior purchases/downloads and user ratings. For instance, a study from 2020 [14], found that besides price, in-app purchase options, and presence of in-app advertisements, user ratings were impactful predictors of user downloads, and the number of downloads increased with average user ratings. However, while metrics such as user ratings may be useful when selecting many other goods and services, they may not accurately reflect the value and risks associated with the use of health apps [15], as these aspects are complex to assess and often not immediately apparent to (prior) users of the app.

In line with this, previous studies have shown that app quality ratings are often not significantly positively associated with user ratings. For instance, user ratings were found not to be significantly correlated with Mobile Application Ratings Scale (MARS) scores [16,17] or ‘PsyberGuide credibility ratings scale’ (PGCRS) scores [18]. A study from 2022 [19], found a weak but significant negative correlation between their criteria and scores and user ratings for women with anxiety during pregnancy.

These findings suggest that user ratings and downloads are not a good proxy for overall app quality. However, most frameworks are not all-encompassing [2023], for example, the MARS doesn’t include privacy questions. Hence, from the previous findings, it is unclear whether user ratings and download rates may be associated with compliance with quality components, such as user experience (UX), professional/clinical assurance (PCA) and data privacy (DP). The current study aimed to examine this relationship.

Specifically, this study’s objective is to analyse the relationship between health app quality scores (UX, PCA and DP) with user ratings and the number of downloads of corresponding health apps. This study has one hypothesis, user ratings and number of downloads are inadequate predictors of user experience, professional/clinical assurance, and data privacy of health apps.

Materials and methods

The dataset provenance

The dataset used for this study was provided by the Organisation for the Review of Care and Health Applications (ORCHA). ORCHA is a United Kingdom (UK) based digital health compliance company that specialises in the assessment of health apps. ORCHA provides an ‘ORCHA library’ that contains information about health apps that have been assessed regarding professional/clinical assurance, data privacy and user experience, allowing consumers and clinical professionals to make informed decisions whether to use or recommend these health apps. ORCHA is currently working with 70% of the National Health Service (NHS) organisations within England.

ORCHA has provided a dataset comprising 2127 health app assessments which were assessed using the ORCHA Baseline Review tool, Version 6 (OBR V6) [24]. For this study 881 Android health apps have been used, the steps involved in the inclusion of health apps can be found in Fig 1 of S1 Appendix. The OBR V6 tool is the latest version of the ‘ORCHA assessment tool’ which consists of ~300 objective assessment questions (where most questions are objective dichotomous–yes/no questions). OBR V6 assesses three aspects of a health app, namely 1) professional/clinical assurance (PCA), 2) data privacy (DP), and 3) user experience (UX) (also referred to as ‘usability and accessibility’). Each of these three domains is scored individually on a scale from 0 to 100 and these three domain scores are combined into an overall ORCHA score.

The dataset consists of the aggregated user ratings, number of downloads and quality scores (UX, PCA and DP scores) for each health app. Each assessment of the 881 health apps has been carried out by at least 2 trained reviewers, where in the case of a dispute, a third reviewer would resolve it. All reviewers have undergone the same training to use the OBR V6 assessment tool. The dataset used included health app assessments that were published between 18th January 2021 and 6th January 2022.

Statistical analysis and modelling

Data was accessed and analysed between July and December 2022. We carried out secondary data analyses of this ORCHA dataset, using R studio and R programming language. Spearman correlations were used to examine how correlated ORHCA, UX, PCA and DP scores are with user ratings (a 1–5 ratings) and number of downloads. The number of downloads variable was converted into download levels, as only download ranges, not exact numbers of downloads, were available. There were 20 ranges of downloads, and each was assigned a download level going from 1 (the smallest) to 20 (the highest). For the analysis the smallest value in each of the 20 ranges was also used as an alternative to the download levels. This was done to improve rigour of the analysis by using two approaches to estimate number of downloads from the available range of downloads.

Multiple linear regression (MLR) was used to model the relationship between app quality scores and the apps’ user ratings and downloads. R squared and adjusted R squared metrics were used to measure the fitness of the models. For all statistical tests, a p-value < .013 (Bonferroni-corrected for multiple hypothesis testing) was considered statistically significant. If there are any links among user ratings and downloads, and quality scores they should be revealed by spearman correlations and/or MLR.

Ethical approval

This secondary data analytics study was approved by Ulster University (ethics filter committee for Faculty of Computing, Engineering and the Built Environment). The process undertaken by ORCHA ensures that health app developers are aware of their score and are given time to contest findings of the assessment which may be amended if developers provide additional relevant information. All reviews, unless explicitly asked to be removed by the developer, are covered as suitable for research in ORCHA’s privacy policy.

Results

There was a total of 881 Android health apps used for this study. The categories of health apps and sample size (n) used in this study are depicted in Table 1 in descending order of sample size. Each health app has been assigned to one or multiple categories.

Table 1. Categories and sample size.

Category Sample size (n)
1 Healthy living 319
2 Mental health 253
3 Medicines and clinical reference 88
4 Neurological 68
5 Diabetes 49
6 Pregnancy 47
7 Respiratory 44
8 Women’s health 38
9 Children’s health 37
10 Musculoskeletal disorders 31
11 Cancer 28
12 Sexual health 33
13 Utilities and Administration 30
14 Neurodiverse 27
15 Ophthalmology 24
16 Pain management 22
17 Cardiology 21
18 Dental 19
19 Dermatology 17
20 Gastrointestinal 15
21 Ear/Nose/Throat/Mouth 14
22 First aid 9
23 Social support network 8
24 Urology 7
25 Allergy 7
26 Older adult 5

Table 2 depicts sample size, median and interquartile range (IQR) for each score (ORCHA, UX, PCA and DP), most common download level when separated by user ratings (in the intervals of > = 1 and <2, > = 2 and <3, > = 3 and <4, > = 4 and < = 5). Table 3 depicts ORCHA recorded number of downloads, sample size, median and interquartile range (IQR) for each score (ORCHA, UX, PCA and DP) when separated by assigned download level (1–20). The sample size for download levels varied from 0 to 177 health apps.

Table 2. User ratings details.

User ratings Sample size (Number of apps) ORCHA score–median (IQR) UX score–median (IQR) PCA score–median (IQR) DP score–median (IQR) Most common download level
> = 1 and <2 8 59(7.5) 72.19(8.96) 48.16(23.30) 65(14.36) 8,10
> = 2 and <3 43 59(22) 76.67(11.27) 46.66(43.50) 67.44(21.04) 8
> = 3 and <4 222 63.5(21) 75.21(9.42) 55.67(44.28) 65.23(17.51) 10
> = 4 and < = 5 608 60(24) 74.55(9.58) 46.93(46.30) 65.36(18.77) 12

Table 3. Downloads details.

Recorded downloads Download levels Sample size (# apps) ORCHA score–median (IQR) UX score–median (IQR) PCA score–median (IQR) DP score–median (IQR)
0+ 1 0 0 0 0 0
1+ 2 0 0 0 0 0
5+ 3 0 0 0 0 0
10+ 4 2 47(12) 64.10(0.90) 26.85(17.68) 65.75(6.45)
50+ 5 3 54(9) 70(7.19) 31.94(10.90) 69.30(15.08)
100+ 6 30 65(19.75) 73.35(9.61) 59.30(40.72) 59.30(16.08)
500+ 7 28 66.5(27.25) 72.20(9.94) 54.93(46.01) 63.64(20.06)
1,000+ 8 139 61(22) 74.55(9.95) 50.89(43.99) 65(22.53)
5,000+ 9 70 62(20.5) 75.21(8.18) 53.84(42.59) 65(21.23)
10,000+ 10 177 60(25) 74.54(10.88) 47.74(47.05) 65.39(18.22)
50,000+ 11 70 60(22.5) 77.51(9.94) 46.72(48.52) 63.89(15.30)
100,000+ 12 134 62(23.75) 75.21(9.54) 46.31(45.51) 66.02(20.98)
500,000+ 13 61 66(22) 74.41(7.73) 55.67(47.41) 67.44(21.11)
1,000,000+ 14 72 60(19.5) 75.52(10.11) 47.95(40.23) 68.62(13.64)
5,000,000+ 15 31 53(20) 72.95(10.56) 33.50(44.30) 66.76(11.34)
10,000,000+ 16 53 63(22) 76.67(8.13) 50.06(40.49) 65.74(13.02)
50,000,000+ 17 6 61(19.25) 74.88(8.57) 41.92(32.56) 79.05(12.39)
100,000,000+ 18 4 53.5(7) 72.60(2.09) 24.59(18.71) 74.64(2.99)
500,000,000+ 19 0 0 0 0 0
1,000,000,000+ 20 1 62(0) 76.67(0) 41.06(0) 75.15(0)

Table 4 depicts Spearman correlation that user ratings and downloads had with each other and with each of the quality scores (ORCHA, UX, PCA and DP). For user ratings and downloads, all correlations were weak (<0.2) and not significant with the quality scores. User ratings had weak negative correlations with PCA and DP scores, and weak positive correlation with UX score. UX and DP scores were weakly positively correlated with downloads, while PCA scores were weakly negatively correlated with downloads. User ratings and downloads correlation was .190 and it was statistically significant when adjusted for multiple hypothesis testing with Bonferroni corrected alpha (p < .001).

Table 4. Spearman correlations.

Bonferroni corrected alpha value .05/9 ≈ .006.

User ratings
(p-value)
Downloads*
(p-value)
ORCHA score -.024
(p = .473)
-.011
(p = .747)
UX score .010
(p = .759)
.084
(p = .012)
PCA score -.043
(p = .207)
-.081
(p = .016)
DP score -.020
(p = .545)
.088
(p = .009)
User ratings NA .190
(p < .001)

*Download levels and ORCHA recorded downloads (plus removed) have the same correlations with other variables.

Table 5 shows the results of MLR, predicting all the assessment scores (separately) with user ratings and download levels. Adjusted R squared was very small for all the scores; however, F-test p-values were statistically significant for UX (p = .005) and DP (p = .003) scores. To make examination of the data more rigorous, the smallest value in the range of values recorded by ORCHA (ORCHA recorded downloads–with plus removed) were also used for comparison.

Table 5. MLR results, using both download levels and ORCHA recorded downloads (removed plus).

Bonferroni corrected alpha value .05/4 ≈ .013.

Scores R squared Adjusted R squared Model p-value (F-test) Independent variable
ORCHA .001 -.002 .759 User ratings
Download level
UX .012 .010 .005 User ratings
Download level
PCA .006 .004 .065 User ratings
Download level
DP .013 .011 .003 User ratings
Download level
ORCHA .001 -.002 .782 User ratings
ORCHA downloads
UX .001 -.001 .704 User ratings
ORCHA downloads
PCA .002 -.0003 .421 User ratings
ORCHA downloads
DP .003 .001 .287 User ratings
ORCHA downloads

Figs 1 and 2 depict how scores’ medians vary with user ratings and download levels. Independent scores UX, PCA and DP are represented with green, blue and purple lines colours respectively and the dependent ORCHA score is depicted with a red line. The download levels of ‘1, 2, 3 and 19’ are not included since the sample size was 0. Fig 1 in S2 Appendix depicts boxplots for each score per user ratings in the intervals of > = 1 and <2, > = 2 and <3, > = 3 and <4, > = 4 and < = 5. Figs 2–5 in S2 Appendix depicts each score per download level. Sample size is above each boxplot.

Fig 1. Median score for each user ratings.

Fig 1

Fig 2. Median score for each download level.

Fig 2

Discussion

Principal findings

This study shows that user ratings and number of downloads are inadequate at predicting the quality of health apps. User ratings and download levels demonstrated weak correlations with all scores (ORCHA, UX, PCA and DP) and each other, as shown in Table 4 (only user ratings and downloads achieved statistically significant correlation with each other when using Bonferroni corrected alpha). Most scores showed a negative correlation with user ratings; UX was the only score that had a positive correlation–albeit weak and not significant. UX and DP scores were positively correlated with download levels, whilst ORCHA and PCA showed a negative correlation with the latter.

The MLR models had low R squared values (< .02), as shown in Table 5, meaning that a lot of the variance in the model is unexplained by the model. This further indicates the inadequacy of user ratings and downloads at predicting scores (ORCHA, UX, PCA and DP).

Our findings indicate that user ratings and download levels are not accurate predictors of objective app quality. This suggests that users have difficulty determining, as a basis for their ratings and download decisions, key aspects that contribute to app quality and safety. A potential contributing factor to this may be a lack of digital health literacy. A study from 2021 described digital health literacy and internet connectivity as “super social determinants of health” [25], because they have implications for the wider social determinants of health. A study from 2017, found that individuals who were younger, had more education, reported excellent health, and had a higher income were the main users of health apps [26].

Moreover, our findings are in line with a study from 2022, which provided evidence of a gap between the user ratings and expert ratings from a curated library of over 1,200 apps that covered physical and mental health [27]. Our results suggest that the cause of this gap may be that health experts look for evidence of clinical quality, utility, privacy, and security that is not considered by users when they rate apps on the iOS and Android app stores. Moreover, users who get their health information, from the internet, rely on search engine results, that may come from unaccredited sources [28]. This indicates that a trusted objective way to judge the quality of health apps is needed.

The study conducted in this paper highlights the need for quality assurance methodologies/tools to accurately determine the quality, safety and compliance of health apps. Our results are in line with the hypothesis that “user ratings and number of downloads are inadequate predictors of user experience, professional/clinical assurance, and data privacy of health apps”. The lack of correlation observed between quality assessment tools and user ratings and downloads of health apps suggest that many users use harmful and unsafe health apps, which may partly be due to poor digital health literacy. These issues need to be addressed as departments of health, for example the Food and Drug Administration of the United States [29] or Health and Social Care Northern Ireland [30], are moving towards embracing digital health technologies such as health apps.

Limitations

This study was limited to Android health apps only, therefore, inclusion of iOS apps, while not expected to be systematically different, may have yielded different findings. User ratings and the number of downloads of health apps included in this study could have changed by the time this study has been published. Additionally, as with any study in digital health, these technologies are highly flexible and subject to change, with updates occurring on a regular basis. Therefore, it is entirely possible that either or both objective compliance of the apps and the number of downloads or user ratings, may have changed since the study began, stressing the need for follow up studies.

OBR is performed by humans and therefore it is entirely possible, although unlikely, that errors can occur in the objective assessment of health apps. The sample size for user ratings ranges (from 8 to 608) and download levels (from 0 to 177) varied widely. Only range of downloads as shown in Table 2 was available for analysis; the exact number of downloads for each health app was unavailable for this study. Which means that precision was not possible, leading to overestimation of download Figs for some and under estimation for others, a natural side effect of transforming continuous data into categorical variables.

Conclusion

This study shows that online user app ratings and the number of app downloads are inadequate predictors of the quality of the health apps in terms of their user experience, professional/clinical assurance, and data privacy. This indicates the need for quality assurance methodologies/tools to accurately determine the quality, safety and compliance of health apps. It also suggests that the success and uptake of a health app is not based on its quality, which is a worrying prospect given the need for high quality health apps and given the need for digital health literacy amongst citizens. It is important that users self-select high quality health apps as opposed to being misled by user ratings and the popularity of an app.

Supporting information

S1 Appendix

(DOCX)

pone.0298977.s001.docx (81.3KB, docx)
S2 Appendix

(DOCX)

pone.0298977.s002.docx (517.2KB, docx)
S1 Data

(CSV)

pone.0298977.s003.csv (66.4KB, csv)

Acknowledgments

We would like to acknowledge the contribution of the many health apps reviewers and developers who worked with ORCHA that allowed for the review of health apps and consented for their data to be used for the purposes of research. Without their contribution and consent this research would not have been possible.

Data Availability

All relevant data are within the manuscript and its Supporting information files.

Funding Statement

This research has been funded by the department for the economy (DfE) in the UK and by company called ORCHA.

References

  • 1.Kern J, Skye A, Krupnick M, Pawley S, Pedersen A, Preciado K, et al. written consent of IQVIA and the IQVIA Institute. Digital Health Trends 2021.
  • 2.Leigh S, Daly R, Stevens S, Lapajne L, Clayton C, Andrews T, et al. Web-based internet searches for digital health products in the United Kingdom before and during the COVID-19 pandemic: a time-series analysis using app libraries from the Organisation for the Review of Care and Health Applications (ORCHA). BMJ Open 2021;11:e053891. doi: 10.1136/bmjopen-2021-053891 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Larsen ME, Nicholas J, Christensen H. A Systematic Assessment of Smartphone Tools for Suicide Prevention. PLoS One 2016;11. doi: 10.1371/journal.pone.0152285 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Alfawzan N, Christen M, Spitale G, Biller-Andorno N. Privacy, Data Sharing, and Data Security Policies of Women’s mHealth Apps: Scoping Review and Content Analysis. JMIR Mhealth Uhealth 2022;10(5):E33735 Https://MhealthJmirOrg/2022/5/E33735 2022;10:e33735. doi: 10.2196/33735 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Melcher J, Torous J. Smartphone Apps for College Mental Health: A Concern for Privacy and Quality of Current Offerings. Psychiatric Services 2020;71:1114–9. doi: 10.1176/appi.ps.202000098 [DOI] [PubMed] [Google Scholar]
  • 6.Sander LB, Schorndanner J, Terhorst Y, Spanhel K, Pryss R, Baumeister H, et al. ‘Help for trauma from the app stores?’ A systematic review and standardised rating of apps for Post-Traumatic Stress Disorder (PTSD). Eur J Psychotraumatol 2020;11. doi: 10.1080/20008198.2019.1701788 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Tangari G, Ikram M, Ijaz K, Kaafar MA, Berkovsky S. Mobile health and privacy: cross sectional study. BMJ 2021;373. doi: 10.1136/bmj.n1248 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Akbar S, Coiera E, Magrabi F. Safety concerns with consumer-facing mobile health applications and their consequences: a scoping review. Journal of the American Medical Informatics Association 2020;27:330–40. doi: 10.1093/JAMIA/OCZ175 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Faessen JPM, Lucassen DA, Buso MEC, Camps G, Feskens EJM, Brouwer-Brolsma EM. Eating for 2: A Systematic Review of Dutch App Stores for Apps Promoting a Healthy Diet during Pregnancy. Curr Dev Nutr 2022;6. doi: 10.1093/cdn/nzac087 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.van Galen LS, Xu X, Koh MJA, Thng S, Car J. Eczema apps conformance with clinical guidelines: a systematic assessment of functions, tools and content. Br J Dermatol 2020;182:444–53. doi: 10.1111/bjd.18152 [DOI] [PubMed] [Google Scholar]
  • 11.MacPherson M, Bakker AM, Anderson K, Holtzman S. Do pain management apps use evidence-based psychological components? A systematic review of app content and quality. Canadian Journal of Pain 2022;6:33–44. doi: 10.1080/24740527.2022.2030212 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Simon L, Reimann J, Steubl LS, Stach M, Spiegelhalder K, Sander LB, et al. Help for insomnia from the app store? A standardized rating of mobile health applications claiming to target insomnia. J Sleep Res 2022:e13642. doi: 10.1111/jsr.13642 [DOI] [PubMed] [Google Scholar]
  • 13.Singh K, Drouin K, Newmark LP, Jae HL, Faxvaag A, Rozenblum R, et al. Many mobile health apps target high-need, high-cost populations, but gaps remain. Health Aff 2016;35:2310–8. doi: 10.1377/hlthaff.2016.0578 [DOI] [PubMed] [Google Scholar]
  • 14.Biviji R, Vest JR, Dixon BE, Cullen T, Harle CA. Factors Related to User Ratings and User Downloads of Mobile Apps for Maternal and Infant Health: Cross-Sectional Study. JMIR Mhealth Uhealth 2020;8(1):E15663 Https://MhealthJmirOrg/2020/1/E15663 2020;8:e15663. doi: 10.2196/15663 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Leigh S, Ouyang J, Mimnagh C. Effective? Engaging? Secure? Applying the ORCHA-24 framework to evaluate apps for chronic insomnia disorder. Evid Based Ment Health 2017;20:e20–e20. doi: 10.1136/eb-2017-102751 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Selvaraj SN, Sriram A. The Quality of Indian Obesity-Related mHealth Apps: PRECEDE-PROCEED Model–Based Content Analysis. JMIR Mhealth Uhealth 2022;10(5):E15719 Https://MhealthJmirOrg/2022/5/E15719 2022;10:e15719. doi: 10.2196/15719 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Bustamante LA, Ménard CG, Julien S, Romo L. Behavior Change Techniques in Popular Mobile Apps for Smoking Cessation in France: Content Analysis. JMIR Mhealth Uhealth 2021;9(5):E26082 Https://MhealthJmirOrg/2021/5/E26082 2021;9:e26082. doi: 10.2196/26082 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.About One Mind PsyberGuide | One Mind PsyberGuide n.d. https://onemindpsyberguide.org/about-psyberguide/ (accessed September 24, 2022).
  • 19.Evans K, Donelan J, Rennick-Egglestone S, Cox S, Kuipers Y. Review of Mobile Apps for Women With Anxiety in Pregnancy: Maternity Care Professionals’ Guide to Locating and Assessing Anxiety Apps. J Med Internet Res 2022;24(3):E31831 Https://WwwJmirOrg/2022/3/E31831 2022;24:e31831. doi: 10.2196/31831 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Hensher M, Cooper P, Dona SWA, Angeles MR, Nguyen D, Heynsbergh N, et al. Scoping review: Development and assessment of evaluation frameworks of mobile health apps for recommendations to consumers. Journal of the American Medical Informatics Association 2021;28:1318–29. doi: 10.1093/jamia/ocab041 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Henson P, David G, Albright K, Torous J. Deriving a practical framework for the evaluation of health apps. Lancet Digit Health 2019;1:e52–4. doi: 10.1016/S2589-7500(19)30013-5 [DOI] [PubMed] [Google Scholar]
  • 22.Torous JB, Chan SR, Yee-Marie Tan Gipson S, Kim JW, Nguyen TQ, Luo J, et al. A hierarchical framework for evaluation and informed decision making regarding smartphone apps for clinical care. Psychiatric Services 2018;69:498–500. doi: 10.1176/APPI.PS.201700423 [DOI] [PubMed] [Google Scholar]
  • 23.Lagan S, Sandler L, Torous J. Evaluating evaluation frameworks: a scoping review of frameworks for assessing health apps. BMJ Open 2021;11:e047001. doi: 10.1136/bmjopen-2020-047001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Hunt Sophie. Review Documentation—Review Development & Resources | Exte n.d. https://confluence.external-share.com/content/b6055aac-83e4-4947-be0e-ebb8c39559ef (accessed March 13, 2022).
  • 25.Sieck CJ, Sheon A, Ancker JS, Castek J, Callahan B, Siefer A. Digital inclusion as a social determinant of health. Npj Digital Medicine 2021 4:1 2021;4:1–3. doi: 10.1038/s41746-021-00413-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Carroll JK, Moorhead A, Bond R, LeBlanc WG, Petrella RJ, Fiscella K. Who Uses Mobile Phone Health Apps and Does Use Matter? A Secondary Data Analytics Approach. J Med Internet Res 2017;19(4):E125 Https://WwwJmirOrg/2017/4/E125 2017;19:e5604. doi: 10.2196/jmir.5604 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.de Chantal PL, Chagnon A, Cardinal M, Faieta J, Guertin A. Evidence of User-Expert Gaps in Health App Ratings and Implications for Practice. Front Digit Health 2022;4:16. doi: 10.3389/fdgth.2022.765993 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Quinn S, Bond R, Nugent C. Quantifying health literacy and eHealth literacy using existing instruments and browser-based software for tracking online health information seeking behavior. Comput Human Behav 2017;69:256–67. doi: 10.1016/J.CHB.2016.12.032 [DOI] [Google Scholar]
  • 29.Device Software Functions Including Mobile Medical Applications | FDA n.d. https://www.fda.gov/medical-devices/digital-health-center-excellence/device-software-functions-including-mobile-medical-applications (accessed November 22, 2022).
  • 30.Digital Strategy—HSC Northern Ireland 2022–2030 | Department of Health n.d. https://www.health-ni.gov.uk/digitalstrategy (accessed November 22, 2022).

Decision Letter 0

Kehinde Kazeem Kanmodi

19 Nov 2023

PONE-D-23-32569Don’t judge a book or health app by its cover: user ratings and downloads are not linked to qualityPLOS ONE

Dear Dr. Hyzy,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jan 03 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Kehinde Kazeem Kanmodi, BDS

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf.

2. Thank you for providing the following Funding Statement: 

“I have read the journal's policy and the authors of this manuscript have the following competing interests: This study is funded by a DfE Cast award and ORCHA. Simon Leigh, Anna Frey and Robert Daly are/were employees at ORCHA when this study was conducted. Other authors have no conflict of interest to declare.”

We note that one or more of the authors is affiliated with the funding organization, indicating the funder may have had some role in the design, data collection, analysis or preparation of your manuscript for publication; in other words, the funder played an indirect role through the participation of the co-authors.

If the funding organization did not play a role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript and only provided financial support in the form of authors' salaries and/or research materials, please review your statements relating to the author contributions, and ensure you have specifically and accurately indicated the role(s) that these authors had in your study in the Author Contributions section of the online submission form. Please make any necessary amendments directly within this section of the online submission form.  Please also update your Funding Statement to include the following statement: “The funder provided support in the form of salaries for authors [insert relevant initials], but did not have any additional role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. The specific roles of these authors are articulated in the ‘author contributions’ section.”

If the funding organization did have an additional role, please state and explain that role within your Funding Statement.

Please also provide an updated Competing Interests Statement declaring this commercial affiliation along with any other relevant declarations relating to employment, consultancy, patents, products in development, or marketed products, etc. 

Within your Competing Interests Statement, please confirm that this commercial affiliation does not alter your adherence to all PLOS ONE policies on sharing data and materials by including the following statement: "This does not alter our adherence to  PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests). If this adherence statement is not accurate and  there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

3. Thank you for stating the following in the Acknowledgments Section of your manuscript:

‘This research is done in partnership with ORCHA, a UK-based digital health compliance company. This work is supported by a Northern Ireland DfE CAST award / PhD scholarship. We would like to acknowledge the contribution of the many health apps reviewers and developers who worked with ORCHA that allowed for the review of health apps and consented for their data to be used for the purposes of research”

We note that you have provided additional information within the Acknowledgements Section that is not currently declared in your Funding Statement. Please note that funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

“This research has been funded by the department for the economy (DfE) in the UK and by company called ORCHA.”

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

4. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

5. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments (if provided):

Nil.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Thank you for sharing this manuscript.

I found your conclusions sound and a valuable addition to the literature.

I do think you need to make some adjustments to your statistical analysis. Primarily, you perform several tests of correlation and several MLRs, and therefore you should be correcting your p-values for multiple hypothesis testing. As this will only serve to make these values less statistically significant, this will not change the conclusions of your study (and indeed will likely make some of the statistically significant correlations no longer significant).

Reviewer #2: Thank you for the opportunity to review your manuscript comparing user ratings to objective assessments of mobile apps for health.

Overall:

Some awkward sentence structures and minor grammatical errors that can easily be corrected.

Introduction

Page 4, lines 88-90: In the sentence: “Therefore, if not sufficiently informed, user’s app choices can result in foregone health benefits if ineffective apps are chosen and/or can pose significant risks to user’s health and privacy.” I wonder if you mean poor health benefits instead of foregone.

Methods

Could you please give a bit more detail about how you determined the download levels from the ranges?

It may also be useful to your readers to have a brief explanation about why you chose the analyses described.

Results

Page 11, Lines 224-225: You may want to consider using a different method for differentiating between your scores on Figures 1 and 2 as these colours may be difficult for someone who is colour-blind to read.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Mar 4;19(3):e0298977. doi: 10.1371/journal.pone.0298977.r002

Author response to Decision Letter 0


19 Dec 2023

Dear Editor and Reviewers,

Thank you for the feedback and the opportunity to improve our manuscript. We have now addressed reviewers’ comments. Below is our point-by-point response to the comments.

Reviewer #1: Thank you for sharing this manuscript.

I found your conclusions sound and a valuable addition to the literature.

I do think you need to make some adjustments to your statistical analysis. Primarily, you perform several tests of correlation and several MLRs, and therefore you should be correcting your p-values for multiple hypothesis testing. As this will only serve to make these values less statistically significant, this will not change the conclusions of your study (and indeed will likely make some of the statistically significant correlations no longer significant).

Reviewer 1 response A1: We acknowledge that the p-values should be corrected for multiple hypothesis testing. We have now revised the analysis and included Bonferroni corrected alpha values.

Reviewer #2: Thank you for the opportunity to review your manuscript comparing user ratings to objective assessments of mobile apps for health. Overall: Some awkward sentence structures and minor grammatical errors that can easily be corrected.

Introduction: Page 4, lines 88-90: In the sentence: “Therefore, if not sufficiently informed, user’s app choices can result in foregone health benefits if ineffective apps are chosen and/or can pose significant risks to user’s health and privacy.” I wonder if you mean poor health benefits instead of foregone.

Reviewer 2 response A1: We have now reworded the sentence to “Therefore, if not sufficiently informed, user’s app choices can result in poor health benefits if ineffective apps are chosen and/or pose significant risks to user’s health and privacy.”

Methods: Could you please give a bit more detail about how you determined the download levels from the ranges? It may also be useful to your readers to have a brief explanation about why you chose the analyses described.

Reviewer 2 response A2: The following paragraphs were modified in the manuscript.

“Data was accessed and analysed between ‎July and December 2022. We carried out secondary data analyses of this ORCHA dataset, using R studio and R programming language. Spearman correlations were used to examine how correlated ORHCA, UX, PCA and DP scores are with user ratings (a 1-5 rating) and number of downloads. The number of downloads variable was converted into download levels, as only download ranges, not exact numbers of downloads, were available. There were 20 ranges of downloads, and each was assigned a download level going from 1 (the smallest) to 20 (the highest). For the analysis the smallest value in each of the 20 ranges was also used as an alternative to the download levels. This was done to improve rigour of the analysis by using two approaches to estimate number of downloads from the available range of downloads.”

“Multiple linear regression (MLR) was used to model the relationship between app quality scores and the apps’ user ratings and downloads. R squared and adjusted R squared metrics were used to measure the fitness of the models. For all statistical tests, a p-value <.013 (Bonferroni-corrected for multiple hypothesis testing) was considered statistically significant. If there are any links among user ratings and downloads, and quality scores they should be revealed by spearman correlations and/or MLR.”

Results: Page 11, Lines 224-225: You may want to consider using a different method for differentiating between your scores on Figures 1 and 2 as these colours may be difficult for someone who is colour-blind to read.

Reviewer 2 response A3: We acknowledge that people with colour-blindness may not be able to differentiate the scores in the figures. Hence, in addition to using colours, we have now labelled figures 1 and 2 with text.

Attachment

Submitted filename: Reviewer response.docx

pone.0298977.s004.docx (16.4KB, docx)

Decision Letter 1

Kehinde Kazeem Kanmodi

2 Feb 2024

Don’t judge a book or health app by its cover: user ratings and downloads are not linked to quality

PONE-D-23-32569R1

Dear Dr. Hyzy,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Kehinde Kazeem Kanmodi, BDS

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Nil.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #2: (No Response)

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #2: (No Response)

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: (No Response)

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #2: (No Response)

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #2: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: No

**********

Acceptance letter

Kehinde Kazeem Kanmodi

22 Feb 2024

PONE-D-23-32569R1

PLOS ONE

Dear Dr. Hyzy,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Kehinde Kazeem Kanmodi

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix

    (DOCX)

    pone.0298977.s001.docx (81.3KB, docx)
    S2 Appendix

    (DOCX)

    pone.0298977.s002.docx (517.2KB, docx)
    S1 Data

    (CSV)

    pone.0298977.s003.csv (66.4KB, csv)
    Attachment

    Submitted filename: Reviewer response.docx

    pone.0298977.s004.docx (16.4KB, docx)

    Data Availability Statement

    All relevant data are within the manuscript and its Supporting information files.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES