Abstract
Most people with mental health disorders cannot receive timely and evidence-based care despite billions of dollars spent by healthcare systems. Researchers have been exploring using digital health technologies to measure behavior in real-world settings with mixed results. There is a need to create accessible and computable digital mental health datasets to advance inclusive and transparently validated research for creating robust real-world digital biomarkers of mental health. Here we share and describe one of the largest and most diverse real-world behavior datasets from over two thousand individuals across the US. The data were generated as part of the two NIMH-funded randomized clinical trials conducted to assess the effectiveness of delivering mental health care continuously remotely. The longitudinal dataset consists of self-assessment of mood, depression, anxiety, and passively gathered phone-based behavioral data streams in real-world settings. This dataset will provide a timely and long-term data resource to evaluate analytical approaches for developing digital behavioral markers and understand the effectiveness of mental health care delivered continuously and remotely.
Subject terms: Health care, Research data
Measurement(s) | depression • real-world behavior • anxiety • Depressed Mood • physical mobility • daily smartphone usage |
Technology Type(s) | Smartphone Application • GPS navigation system |
Factor Type(s) | depression level, socioeconomic indicators, demographics |
Sample Characteristic - Organism | Homo sapiens |
Sample Characteristic - Environment | real world setting |
Sample Characteristic - Location | contiguous United States of America |
Background & Summary
Although effective treatments exist, depression is one of the leading causes of disability worldwide1. A challenge in eradicating this burdensome and costly illness is the poor access to timely diagnosis and treatment. Individuals with depression often lack adequate and timely care, and even when care is available, it is often not evidence-based, and outcomes are not measured consistently1–3. In addition, there are known barriers such as time commitment, regular follow-ups in behavioral therapy4, and limited availability of highly trained professionals, especially in rural and lower-income areas5. Some populations also do not access care because of the stigma associated with mental health problems6. In particular, there are known socio-technical challenges in recruiting underserved communities7 such as the Hispanic/Latino population, in traditional research studies8,9 in the US. Asian, Hispanic, and Black people are less likely to receive access to and utilize mental healthcare services than their White counterparts10–12. As a result, our understanding of optimal support and treatment options for underserved and minoritized populations remains limited.
In response to such challenges, there is broad interest in using digital health technology to provide individuals with novel preventative and therapeutic solutions in a timely and scalable manner compared to traditional mental health services13. Current applications of digital technology in mental health include the remote deployment14 of guided interventions (e.g., cognitive behavioral therapy15). Due to heterogeneity in diagnoses of mental illnesses, researchers are beginning to explore the use of smartphones and wearables16 in understanding personalized day-to-day factors and their severity impacting individuals’ mental health. Multimodal data gathered from connected devices can help understand individualized real-world behavior, aiding the development of digital phenotypes17. Once validated, these digital signatures can ultimately help track individual physiological symptoms that can be used to tailor and augment the treatment of mental illnesses18 as well as guide the development of novel treatments for mental health disorders.
In the last ten years, technology developers and researchers have begun to explore the potential of smartphone apps as an economical and scalable means for assessing and delivering mental health care. Over 10,000 smartphone apps for mental health and well-being are available on the App and Play stores. However, there is little information about their utilization, efficacy, and effectiveness19. When assessing the quality and efficacy of smartphone-based interventions as a supplement to treatments or compared to other therapeutic interventions, there have been varying results across treatment groups20. There have also been notable concerns regarding the use of mobile apps, potentially excluding groups who may not have access to technology due to cost, digital literacy, location, or availability21. Many individuals also express concerns over potential privacy and consent issues regarding access to personal mental health-related data22 and apps storing sensitive information such as account information and demographics, which may discourage their use23. Essential questions regarding the utility of digital mental health apps are (1) who downloads these apps, (2) do people who use these apps use them as designed, (3) are they effective in collecting valid and long-term behavioral data, (4) for whom are they effective and for how long, and (5) are these tools easy to access and as effective in under-represented minorities?
The limited availability of real-world behavioral datasets curated using FAIR24 data sharing and management principles is a critical bottleneck that hampers digital mental health research advancement. A recent systematic review25 of real-world behavior sensing studies highlighted substantial challenges such as small sample sizes and inconsistent reporting of analytical results that affect the reproducibility of findings. And as a result, it impacts the development of robust and transparently validated digital biomarkers of mental health using real-world data (RWD). Irreproducibility of research findings remains the top challenge stymieing the real-world application of biomedical research. Recent meta-studies have shown that up to 50% of reviewed studies were not reproducible, costing taxpayers billions of dollars every year26,27. NIH has recently instituted a mandate requiring researchers to make their research data accessible for all NIH-funded research starting January 202328.
We describe and share one of the most extensive longitudinal real-world behavioral datasets to enable a broader ecosystem of digital health datasets accessible to the biomedical research community and curated using FAIR principles24. We believe this dataset will provide a significant long-term value for the researchers interested in digital mental health across various use cases, such as (1) cross-study data integration and comparisons of findings, (2) benchmarking analytical approaches for predicting functional and behavioral outcomes, (3) specific app use signatures and their relationship to outcomes, (4) modeling individualized behavioral trajectories over time, (5) the impact of being in traditional treatment while using the app on outcomes, (6) discovery depression subtypes based on baseline characteristics and how they relate to: i) adherence with care, ii) changes in mood and sleep patterns, and iii) stability of outcomes.
Methods
Brighten V1 and V2 studies
The two randomized clinical trials, Brighten (Bridging Research Innovations for Greater Health in Technology, Emotion, and Neuroscience) V1 study29, (henceforth V1) and Brighten V27, (henceforth V2), have been used to investigate: 1) feasibility of conducting fully remote decentralized clinical trials7,29, 2) the relative effectiveness of digital health apps in improving mood30, 3) utility of real-world behavioral data collected passively from smartphones in predicting changes in mood over time at the population and individual level31 and 4) factors impacting participant recruitment and retention in decentralized health research32. Both studies deployed three app-based remote interventions and gathered real-world data (active: remote surveys and passive: smartphone usage) related to depression shared and described here.
Overall, 7,850 participants from across the US were screened, with 2,193 participants meeting the eligibility criteria and consenting to enroll in the two studies (Figs. 1, 2; Table 1). The individual-level engagement data analysis showed significant variation in participant retention linked to demographic and socioeconomic factors further described within the two studies7,32. Study participants’ average depression severity (assessed using PHQ-9 survey33) at onboarding indicated moderate depression symptoms (V1:14 ± 5.1, V2:14.3 ± 5.4). A notable proportion of the cohort indicated suicidal ideation (non-zero response to 9th item of PHQ-9) at onboarding (V1:29.4%, V2:35.6%).
Fig. 1.
Overall schematic showing participant onboarding process from online recruitment and prospective data collection using smartphones to random allocation to one of the study interventions.
Fig. 2.
US map showing the location of participants* who were screened (blue) and enrolled (yellow) in the two Brighten studies. *Based on participants who shared at least the first three numbers of their zipcode.
Table 1.
Summary of all surveys and passive data contributed by participants enrolled in the two studies along with data citation.
Administration Frequency | Activity Type | Content | Unique Participants | Total Records | Data Citation |
---|---|---|---|---|---|
Once | Baseline Characteristics | Age, Gender, State, Zip Code, Income, Race Education, Study Arm, Device | 2193 | 2193 | 10.7303/syn2708259747 |
Baseline PHQ-9 | A clinically validated screener of depression was used to screen participants for eligibility of joining the study | 1919 | 1919 | 10.7303/syn2708281148 | |
GAD-7 | A clinically validated screener of generalized anxiety disorder | 820 | 820 | 10.7303/syn1702265549 | |
AUDIT-C | A three-item alcohol screener | 832 | 835 | 10.7303/syn1702128050 | |
IMPACT Mania and Psychosis Screening | A 5-question self-assessment survey used to screen for issues of mania and/or psychosis collected during the onboarding week | 818 | 825 | 10.7303/syn2705127651 | |
Weekly | PHQ-9 | Same as baseline PHQ-9 survey administered weekly for the first four weeks and then every 2 weeks | 934 | 4875 | 10.7303/syn2720235552 |
SDS | A clinically validated tool to assess functional impairment administered weekly for the first four weeks and then every 2 weeks | 1016 | 4761 | 10.7303/syn1702265853 | |
Patients Global Impression of Change Scale | Reflects a patient’s belief about the efficacy of treatment | 915 | 2856 | 10.7303/syn1702331354 | |
Sleep | A three-question survey to assess the participant’s personal sleep characteristics. | 948 | 2319 | 10.7303/syn1702265955 | |
Mental Health Services Used | Five-question survey to assess the use of mental health services | 947 | 2324 | 10.7303/syn1702266056 | |
Other health-related apps used | Survey to assess other health-related apps used by study participants administered during week 1,4,8,12 | 818 | 2022 | 10.7303/syn1702505857 | |
Study App Satisfaction | A four-question survey to probe participants satisfaction with the assigned intervention app deployed at week 4,8 and 12 | 514 | 1175 | 10.7303/syn1702520258 | |
Daily | PHQ-2 | This survey contains the first two questions of PHQ-9 focused on assessing mood and anhedonia and was administered daily during the 12 weeks of the study period | 1073 | 47,976 | 10.7303/syn1702085559 |
Passive data | Tracking sensor-based data such as phone usage and GPS. For Brighten-v2 raw GPS data was collected | 900 | 60,470 | See Table 3 |
There were significant differences in race/ethnicity between the two study cohorts (Hispanic/Latino V1:13%, V2:38%) primarily driven by underlying differences in recruitment approaches between studies. Using targeted recruitment approaches (further described here7), we enriched the V2 study cohort for Hispanics/Latinos to help investigate the feasibility of using digital mental health tools in this target population. See Table 2 for further details and a descriptive summary of the overall cohort stratified by V1 and V2 studies.
Table 2.
Sociodemographic summary of enrolled participants in Brighten V1 and V2 studies.
Overall | Brighten-V1 | Brighten-V2 | |
---|---|---|---|
Screened(N) | 7850 | 3348 | 4502 |
Enrolled (N) | 2193 | 1110 | 1083 |
Gender (%) | |||
Female | 1638 (74.7) | 870 (78.4) | 768 (70.9) |
Male | 547 (24.9) | 235 (21.2) | 312 (28.8) |
Age Group (%) | |||
18–30 | 1039 (48.4) | 583 (53.6) | 456 (43.1) |
31–40 | 563 (26.2) | 243 (22.4) | 320 (30.2) |
41–50 | 350 (16.3) | 151 (13.9) | 199 (18.8) |
51–60 | 147 (6.9) | 84 (7.7) | 63 (6.0) |
61–70 | 38 (1.8) | 20 (1.8) | 18 (1.7) |
71+ | 8 (0.4) | 6 (0.6) | 2 (0.2) |
Device (%) | |||
iPhone | 1489 (67.9) | 573 (51.6) | 916 (84.6) |
Android | 551 (25.1) | 384 (34.6) | 167 (15.4) |
Race/Ethnicity (%) | |||
Non-Hispanic White | 1144 (52.2) | 646 (58.2) | 498 (46.0) |
Hispanic/Latino | 555 (25.3) | 144 (13.0) | 411 (38.0) |
African-American/Black | 222 (10.1) | 138 (12.4) | 84 (7.8) |
Asian | 149 (6.8) | 94 (8.5) | 55 (5.1) |
More than one | 70 (3.2) | 70 (6.3) | 0 (0.0) |
American Indian/Alaskan Native / Pacific Islander | 25 (1.1) | 13 (1.2) | 12 (1.1) |
Income Last year (%) | |||
<$20,000 | 668 (30.5) | 266 (24.0) | 402 (37.1) |
20,000–40,000 | 417 (19.0) | 154 (13.9) | 263 (24.3) |
40,000–60,000 | 273 (12.4) | 92 (8.3) | 181 (16.7) |
60,000–80,000 | 122 (5.6) | 44 (4.0) | 78 (7.2) |
80,000–100,000 | 67 (3.1) | 18 (1.6) | 49 (4.5) |
100,000+ | 130 (5.9) | 31 (2.8) | 99 (9.1) |
Marital status (%) | |||
Single | 1199 (54.7) | 665 (59.9) | 534 (49.3) |
Married/Partner | 723 (33.0) | 310 (27.9) | 413 (38.1) |
Separated/Widowed/Divorced | 255 (11.6) | 130 (11.7) | 125 (11.5) |
Education (%) | |||
University | 865 (39.4) | 425 (38.3) | 440 (40.6) |
High School | 493 (22.5) | 231 (20.8) | 262 (24.2) |
Community College | 465 (21.2) | 237 (21.4) | 228 (21.1) |
Graduate Degree | 343 (15.6) | 207 (18.6) | 136 (12.6) |
*See Supplementary Table 1 for details on participants with missing data for one or more sociodemographic categories
Study design
Brighten studies (V1 and V2) aimed to assess the feasibility of conducting fully remote smartphone-based randomized clinical trials to evaluate and deliver digital interventions to individuals with depression. While both trials were open to any individual meeting the eligibility criteria, Brighten V2 focused on increasing the participation of Hispanic/Latino populations. Ethical oversight and approval for both trials (V1:NCT00540865, V2:NCT01808976) were granted by the Institutional Review Board of the University of California, San Francisco. Participants in both studies were randomized to one of three intervention apps and prospectively observed for 12 weeks. The three intervention apps were: 1.) iPST - an app based on problem-solving therapy34,35, an evidence-based treatment for depression, 2.) Project Evo - a therapeutic video game meant to improve cognitive skills associated with depression36, and 3.) Health Tips - an app that provides information about strategies to improve mood.
Participant recruitment & onboarding
Participants had to speak English or Spanish (in V2), be 18 years or older, and own either an iPhone with Wi-Fi or 3 G/4 G/LTE capabilities or an Android phone with an Apple iPad version 2.0 or newer device. In both studies, participants had to experience clinically significant symptoms of depression, as indicated by either a score of 5 or higher on PHQ-9 or a score of 2 or greater on PHQ item 10.
Participants were recruited using a combination of three different approaches - traditional methods (ads placed in local media), social media-based advertising (ads on Facebook and Craigslist), and online targeted recruitment campaigns (in partnership with Trialspark.com for the V2 study). The onboarding process involved learning about the study through the study website. A video highlighting the research study’s goals and procedures and the risks and benefits of participation was also made available.
After viewing the video and reading the online study consent, participants had to correctly answer three questions to confirm their understanding that participation was voluntary, not a substitute for treatment, and that they were randomized to treatment conditions. Each question had to be answered correctly before a participant could consent to participate, at which point they could move on to the baseline assessment and randomization. Participants’ consent to participate in the study was asked before their enrollment eligibility could be established. This was done to assess their baseline depression state (via PHQ-9 survey), which was required to meet the study eligibility criteria and could not be collected without the participant’s consent. In the consent form signed by participants, there was a clause describing that other organizations and researchers may use these data for research and quality assurance. This follows the National Institute of Mental Health (NIMH) data sharing policy (NOT-MH-19-033), designed to share data at the end of the award period or upon the research findings being published.
Once confirmed, participants were sent a link to download the study intervention app they were randomly assigned. The V2 study translated the recruitment material into Spanish for Spanish speakers. Participants were compensated for their time in both studies. Each participant received up to $75 gift vouchers for completing all assessments over the 12-week study period: US $15 after completing the initial baseline assessment and US $20 for each subsequent assessment at the 4-, 8-, and 12-week time points.
Data collection included participant self-reported data using a combination of validated survey instruments and short ecological momentary assessments. We also collected passive sensor data from smartphones related to participants’ daily activities as described below.
Baseline assessment
All consented participants in V1 and V2 studies provided the same baseline information. The demographic survey included age, race/ethnicity, marital and employment status, income, and education. To assess the mental health status of participants at baseline, we used the following scales: The Patient Health Questionnaire (PHQ-9)33 for measuring depression symptoms, Sheehan disability scale (SDS)37 for assessing functional impairment, and Generalized anxiety disorder (GAD-7) for anxiety. The enrollment system flagged any participant indicating suicidal ideation (non-zero response on the 9th item of the PHQ-9 survey). Such participants were suggested that this study may not be the best fit for them and that they should consider reaching out to their healthcare provider. We also provided links to online resources, including a 24-hour suicide help hotline, to receive immediate help.
All participants were evaluated for a history of mania, psychosis, and alcohol consumption using a four-item mania and psychosis screening instrument38 and the four-item National Institute on Alcohol Abuse and Alcoholism (NIAAA) Alcohol Screening Test. We also asked study participants about smart device ownership, use of health apps, and mental health services, including the use of medications and psychotherapy. Participants were asked to rate their health on a scale of excellent to poor. All baseline assessments were collected from eligible participants before interventions were provided, and all questions were optional for participants to answer. For further descriptions of the clinical assessments and other surveys administered in the V1 and V2 studies, please visit the Brighten study data portal (www.synapse.org/brighten).
Longitudinal data collection
The participants answered daily and weekly questions using the study app and continual passive data collection. The study app did not collect private information such as the content of text messages, emails, or voice calls.
Active tasks
Participants completed primary outcome assessments using PHQ-9 and SDS surveys once a week for the first four weeks and then every other week for the duration of the study (12 weeks). To assess participants’ mood and anhedonia (core symptoms of depression), we used the PHQ-2 survey administered daily in the morning. Participants also completed other secondary measures (described below) at daily, weekly, or biweekly intervals. These included sleep assessment with three questions - time taken to fall asleep, duration of sleep, and awake time during the night. Each question used a multiple-choice Likert scale of time (e.g., 0–15 minutes, 16–30 minutes, 31–60 minutes, etc.) to answer these sleep questions. Participants were automatically notified every 8 hours, for 24 hours, if they had not completed a survey within 8 hours of its original delivery. An assessment was considered missing if it was not completed within this 24-hour time frame.
Passive data collection
For the V1 study, we deployed an app in collaboration with Ginger.io that collected passive data. The application automatically launched the background data gathering processes once the user logged in. After the initial logon, the process was automatically launched whenever it was not already running (on phone restart or other events that terminate the process). The data collected through the device include communication metadata (e.g., time of call/SMS, call duration, SMS length) and mobile data such as activity type and distance traveled. The study app did not collect personal and identifiable communication data such as the actual text of SMS or the phone numbers for incoming and outgoing calls.
In the V2 study, we developed an in-house app named Survetory to help collect more fine-grained behavioral data. The passive data streams included a daily descriptive summary of phone-based communications (e.g., number of calls and texts) and intraday GPS-based locations. If location services (GPS) were available, the study app recorded the latitude and longitude every ten minutes or a movement more than 100 meters from the previously recorded spot. While the raw GPS data from participants cannot be shared broadly due to data privacy, we are sharing granular intraday level summaries of features derived from the raw GPS data. To featurize the raw GPS data, we developed an open-source pipeline, gSCAP-Geospatial Context Analysis Pipeline (https://github.com/aid4mh/gSCAP). Briefly, the gSCAP pipeline uses raw and sensitive individual-level GPS data to generate individual-level geospatial semantics that is not identifiable in the same way as GPS coordinates. Some examples of these contextual features include overall daily mobility (time and distance walking, biking, etc.), average daily weather (temperature, dew, cloud) at the participant’s location, and type of places visited in a day (coffee shop, park, shopping mall, etc.). However, the pipeline does not extract and store the location of the different places visited).
Further details about study design, participant recruitment, onboarding, and data collection are available in Brighten V129 and V27 papers, respectively.
Data storage and security
The initial raw data from each study were collected and stored on HIPAA-compliant servers located in the Department of Neurology at the University of California, San Francisco. The study servers were configured to UCSF Minimum Security Standards for Electronic Information Resources standards and policies.
Data Records
The data gathered from study participants during the 12-week observation period (as described above) can be accessed by qualified researchers39 via the study data portal (www.synapse.org/brighten). The study portal has detailed data descriptions for each active and passive data stream. The detailed steps for accessing the study data can be found on the study portal under the “Accessing the Brighten Study data” page and highlighted in the usage notes below.
Please refer to Table 1 for a record of surveys completed by the participants and Table 3 for passive behavioral data. Participants’ demographics and clinical characteristics at onboarding are summarized in Table 2.
Table 3.
Summary of Passive Data contributed by participants enrolled in the two studies along with data citation.
Activity | Content | Data Citation |
---|---|---|
Passive Features Brighten v1 | The passive features collected in Brighten V1 study through a propriety third-party app developed by Ginger.IO. | 10.7303/syn1702550060 |
Passive Cluster Entries Brighten v2 | Daily visited location categories | 10.7303/syn1711669561 |
Passive Mobility Features Brighten v2 | Daily aggregates of the mobility features created by gSCAP pipeline (see Methods) | 10.7303/syn1711466262 |
Passive Phone Communication Features Brighten v2 | Daily aggregates of the phone usage metadata such the number of incoming and outgoing calls, messages (length, sent, received), etc. | 10.7303/syn1706050263 |
Passive Weather Features Brighten v2 | The weather features were gathered from the Dark Sky API. For each day that GPS data existed, the API was queried for 24 hourly metrics which were aggregated into the summary statistics. | 10.7303/syn1706128464 |
Technical Validation
The two Brighten studies collected multimodal mental health and behavioral data in fully remote real-world settings of over two thousand participants and should be treated in that context. All shared data has been fully described on the study data portal as part of the data annotation and curation process. During the data cleaning and harmonization process, column names in the data files were aligned. The study portal also stores the provenance of upstream data cleaning and curation steps.
To preserve participants’ privacy and in consultation with data governance experts, we removed the self-reported zip code of study participants and the raw GPS data gathered in the V2 study. However, granular intraday GPS-derived features40 including the open-source pipeline gSCAP used to featurize the GPS data is freely available41. To optimize the parameters of the gSCAP pipeline for clustering GPS data, we took an iterative and visual approach to select parameters. Further details on pipeline parameters are available through the GitHub page (https://github.com/aid4mh/gSCAP).
Since both studies had monetary incentives for study participants to engage with the study apps over the 12 weeks, we paid particular attention to mitigating the possibility of individuals looking to take advantage of this study to acquire the research payment in several ways. First, the maximum compensation of $75 was spread over 12 weeks, reducing the motivation to “game the system.” Second, participants had to have both a valid email and phone number to successfully enroll in the study, reducing the ability to create multiple accounts. The study team also monitored for any duplicate records. Additionally, if an individual made multiple attempts to enroll using the same email address and each time did not properly answer the enrollment questions, the individual would not be enrolled. This prevents the individual from receiving access to any of the study tools or compensation. And finally, the link to download subject-specific study apps was only valid for a single user, as a unique password was required to view the download page.
Limitations
There is active ongoing research to efficiently develop and deploy large-scale, fully remote research studies to collect health data at scale. While the early evidence shows substantial potential42–44 for the use of digital health technology in healthcare settings, several challenges42,43,45 have also emerged, such as the underrepresentation of certain populations who may lack access to the technology or the technological literacy to use the apps. These issues will need to be addressed so that the collected health-related data is representative of the target population and remains balanced over time.
The Brighten data shows similar limitations. First, potential participants with iOS and Android phones could join the study. Participants with android devices were also required to have an iOS-based tablet to use one of the interventional apps. This additional requirement impacted the cohort’s diversity, particularly in the V2 study, where significantly more participants had an iPhone. Second, both studies recruited a population not representative of the general population. For example, close to three-fourths of the enrolled participants were females and are known to be twice as likely depressed as males46. Third, the V2 study tried to enrich the study cohort for Hispanics/Latinos to investigate the use of digital mental health tools within that target population. This led to significant differences in the underlying race/ethnicity between the two studies and should be factored in the downstream analysis. Fourth, despite both studies providing participation incentives, we saw significant participant attrition rates, especially in the V2 study that has been further discussed in our previous work7,32. And finally, the passive data gathered from smartphones was impacted by underlying system-level differences in iOS and Android operating systems, along with different apps used to collect passive data in V1 and V2 studies. We suggest any downstream analysis account for such underlying differences in real-world data collection.
Usage Notes
Brighten data have been donated by many thousands of research participants to advance our understanding of mental health. Due to the sensitive nature of the data, we implemented the following data access governance structure that respects the desire of participants to share their data while also protecting the confidentiality and privacy of research participants. Researchers interested in accessing these data need to complete the following steps:
Become a Synapse Certified User with a validated user profile
Submit a 1–3 paragraph Intended Data Use statement to be posted publicly on Synapse
Agree to comply with the data-specific Conditions for Use when prompted.
These include a commitment to keep the data secure and confidential, not attempt to re-identify research participants for any reason, abide by the guiding principles for responsible research use and the Synapse Awareness and Ethics Pledge(to only use the data as intended, not use the data for any commercial advertisement), and commit to reporting any misuse or data release, intentional or inadvertent to the Synapse Access Compliance Team within five business days. In addition, users must agree to publish findings in open access publications and acknowledge the research participants as data contributors and the study investigators on all publications or presentations resulting from using these data as follows: “These data were contributed by the participants of the Brighten Studies, which were funded by the National Institute of Mental Health with scientific oversight from Pat Arean, Ph.D. and Joaquin A. Anguera, Ph.D., as described in Synapse: https://www.synapse.org/brighten. See the full instructions for requesting data access on the Accessing the Brighten Data page (https://www.synapse.org/brighten)
Supplementary information
Acknowledgements
Support for this research was provided by the National Institute of Mental Health (PAA R34MH100466, T32MH0182607, K24MH074717). The authors would also like to thank all the study participants for their time and willingness to share their data. We also acknowledge the contributions of all our industry partners, namely Ginger.IO and Akili Interactive. We would also like to thank Wow Labz for helping build the data collection app for the V2 study.
Author contributions
P.A. and J.A.A. obtained funding from NIMH and provided overall oversight, including the scientific design of Brighten V1 and V2 studies. AP wrote the first manuscript draft. L.W., A.H. and P.A. made significant contributions in editing the manuscript. J.V. and J.A.A. were involved in the data collection. A.P., L.W. performed the data cleaning, curation, and creation of the study data portal. L.W., A.P. and C.H. developed the gSCAP pipeline used for featurizing the raw GPS data. C.S. Developed the data governance process for sharing the dataset with the broader research community. All authors provided editorial feedback and contributed to the final approval of the manuscript.
Code availability
The studies used a combination of open-source and proprietary apps to collect observational data and deploy remote interventions via three apps (described above). The source code for tools used for the data collection, such as the passive data collection app (Survetory - used in the V2 study), iPST, and HealthTips (two of the three randomized interventions), is available on the study Github archive (https://github.com/aid4mh/Brighten_Studies_Archive). The gSCAP pipeline used to featurize the raw GPS data from the V2 study is available here (https://github.com/aid4mh/gSCAP). The data cleaning and curation pipeline is also available on GitHub (https://github.com/apratap/BRIGHTEN-Data-Release).
Competing interests
The authors declare no competing interests.
Footnotes
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
The online version contains supplementary material available at 10.1038/s41597-022-01633-7.
References
- 1.Depression. at https://www.who.int/news-room/fact-sheets/detail/depression.
- 2.Scott K, Lewis CC. Using Measurement-Based Care to Enhance Any Treatment. Cogn. Behav. Pract. 2015;22:49–59. doi: 10.1016/j.cbpra.2014.01.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Lewis CC, et al. Implementing Measurement-Based Care in Behavioral Health: A Review. JAMA Psychiatry. 2019;76:324–335. doi: 10.1001/jamapsychiatry.2018.3329. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Cuijpers P, Quero S, Dowrick C, Arroll B. Psychological Treatment of Depression in Primary Care: Recent Developments. Curr. Psychiatry Rep. 2019;21:129. doi: 10.1007/s11920-019-1117-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Alavi N, et al. Feasibility and Efficacy of Delivering Cognitive Behavioral Therapy Through an Online Psychotherapy Tool for Depression: Protocol for a Randomized Controlled Trial. JMIR Res. Protoc. 2021;10:e27489. doi: 10.2196/27489. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Andrade LH, et al. Barriers to mental health treatment: results from the WHO World Mental Health surveys. Psychol. Med. 2014;44:1303–1317. doi: 10.1017/S0033291713001943. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Pratap A, et al. Using Mobile Apps to Assess and Treat Depression in Hispanic and Latino Populations: Fully Remote Randomized Clinical Trial. J. Med. Internet Res. 2018;20:e10130. doi: 10.2196/10130. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Arevalo M, et al. Mexican-American perspectives on participation in clinical trials: A qualitative study. Contemp Clin Trials Commun. 2016;4:52–57. doi: 10.1016/j.conctc.2016.06.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Mental Health: Culture, Race, and Ethnicity: Executive Summary: a Supplement to Mental Health: a Report of the Surgeon General. (2001). [PubMed]
- 10.Cook BL, Trinh N-H, Li Z, Hou SS-Y, Progovac AM. Trends in Racial-Ethnic Disparities in Access to Mental Health Care, 2004–2012. Psychiatr. Serv. 2017;68:9–16. doi: 10.1176/appi.ps.201500453. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Thomeer, M. B., Moody, M. D. & Yahirun, J. Racial and Ethnic Disparities in Mental Health and Mental Health Care During The COVID-19 Pandemic. J Racial Ethn Health Disparities, 10.1007/s40615-022-01284-9 (2022). [DOI] [PMC free article] [PubMed]
- 12.Harris KM, Edlund MJ, Larson S. Racial and ethnic differences in the mental health problems and use of mental health care. Med. Care. 2005;43:775–784. doi: 10.1097/01.mlr.0000170405.66264.23. [DOI] [PubMed] [Google Scholar]
- 13.Wies B, Landers C, Ienca M. Digital Mental Health for Young People: A Scoping Review of Ethical Promises and Challenges. Front Digit Health. 2021;3:697072. doi: 10.3389/fdgth.2021.697072. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Lattie EG, Stiles-Shields C, Graham AK. An overview of and recommendations for more accessible digital mental health services. Nature Reviews Psychology. 2022;1:87–100. doi: 10.1038/s44159-021-00003-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Nordh M, et al. Therapist-Guided Internet-Delivered Cognitive Behavioral Therapy vs Internet-Delivered Supportive Therapy for Children and Adolescents With Social Anxiety Disorder: A Randomized Clinical Trial. JAMA Psychiatry. 2021;78:705–713. doi: 10.1001/jamapsychiatry.2021.0469. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Griffin B, Saunders KEA. Smartphones and Wearables as a Method for Understanding Symptom Mechanisms. Front. Psychiatry. 2019;10:949. doi: 10.3389/fpsyt.2019.00949. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Jain SH, Powers BW, Hawkins JB, Brownstein JS. The digital phenotype. Nat. Biotechnol. 2015;33:462–463. doi: 10.1038/nbt.3223. [DOI] [PubMed] [Google Scholar]
- 18.Moshe I, et al. Digital interventions for the treatment of depression: A meta-analytic review. Psychol. Bull. 2021;147:749–786. doi: 10.1037/bul0000334. [DOI] [PubMed] [Google Scholar]
- 19.Larsen ME, et al. Using science to sell apps: Evaluation of mental health app store quality claims. NPJ Digit Med. 2019;2:18. doi: 10.1038/s41746-019-0093-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Goldberg, S. B., Lam, S. U., Simonsson, O., Torous, J. & Sun, S. Mobile phone-based interventions for mental health: A systematic meta-review of 14 meta-analyses of randomized controlled trials. PLOS Digit Health1 (2022). [DOI] [PMC free article] [PubMed]
- 21.Seabrook EM, Nedeljkovic M. The evolving landscape of digital mental health: implications for research and practice. Clin. Psychol. 2021;25:121–123. doi: 10.1080/13284207.2021.1977157. [DOI] [Google Scholar]
- 22.Huckvale K, Torous J, Larsen ME. Assessment of the Data Sharing and Privacy Practices of Smartphone Apps for Depression and Smoking Cessation. JAMA Netw Open. 2019;2:e192542. doi: 10.1001/jamanetworkopen.2019.2542. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Lustgarten SD, Garrison YL, Sinnard MT, Flynn AW. Digital privacy in mental healthcare: current issues and recommendations for technology use. Curr Opin Psychol. 2020;36:25–31. doi: 10.1016/j.copsyc.2020.03.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Wilkinson MD, et al. The FAIR Guiding Principles for scientific data management and stewardship. Sci Data. 2016;3:160018. doi: 10.1038/sdata.2016.18. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.De Angel V, et al. Digital health tools for the passive monitoring of depression: a systematic review of methods. NPJ Digit Med. 2022;5:3. doi: 10.1038/s41746-021-00548-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Freedman LP, Cockburn IM, Simcoe TS. The Economics of Reproducibility in Preclinical Research. PLoS Biol. 2015;13:e1002165. doi: 10.1371/journal.pbio.1002165. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Stupple A, Singerman D, Celi LA. The reproducibility crisis in the age of digital medicine. NPJ Digit Med. 2019;2:2. doi: 10.1038/s41746-019-0079-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Kozlov M. NIH issues a seismic mandate: share data publicly. Nature. 2022;602:558–559. doi: 10.1038/d41586-022-00402-1. [DOI] [PubMed] [Google Scholar]
- 29.Anguera JA, Jordan JT, Castaneda D, Gazzaley A, Areán PA. Conducting a fully mobile and randomised clinical trial for depression: access, engagement and expense. BMJ Innov. 2016;2:14–21. doi: 10.1136/bmjinnov-2015-000098. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Arean PA, et al. The Use and Effectiveness of Mobile Apps for Depression: Results From a Fully Remote Clinical Trial. J. Med. Internet Res. 2016;18:e330. doi: 10.2196/jmir.6482. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Pratap A, et al. The accuracy of passive phone sensors in predicting daily mood. Depress. Anxiety. 2019;36:72–81. doi: 10.1002/da.22822. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Pratap A, et al. Indicators of retention in remote digital health studies: a cross-study evaluation of 100,000 participants. NPJ Digit Med. 2020;3:21. doi: 10.1038/s41746-020-0224-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Kroenke K, Spitzer RL, Williams JB. The PHQ-9: validity of a brief depression severity measure. J. Gen. Intern. Med. 2001;16:606–613. doi: 10.1046/j.1525-1497.2001.016009606.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Mynors-Wallis LM, Gath DH, Day A, Baker F. Randomised controlled trial of problem solving treatment, antidepressant medication, and combined treatment for major depression in primary care. BMJ. 2000;320:26–30. doi: 10.1136/bmj.320.7226.26. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Arean P, Hegel M, Vannoy S, Fan M-Y, Unuzter J. Effectiveness of problem-solving therapy for older, primary care patients with depression: results from the IMPACT project. Gerontologist. 2008;48:311–323. doi: 10.1093/geront/48.3.311. [DOI] [PubMed] [Google Scholar]
- 36.Anguera JA, Gunning FM, Areán PA. Improving late life depression and cognitive control through the use of therapeutic video game technology: A proof-of-concept randomized trial. Depress. Anxiety. 2017;34:508–517. doi: 10.1002/da.22588. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Sheehan DV, Harnett-Sheehan K, Raj BA. The measurement of disability. Int. Clin. Psychopharmacol. 1996;11(Suppl 3):89–95. doi: 10.1097/00004850-199606003-00015. [DOI] [PubMed] [Google Scholar]
- 38.Unützer J, et al. Improving primary care for depression in late life: the design of a multicenter randomized trial. Med. Care. 2001;39:785–799. doi: 10.1097/00005650-200108000-00005. [DOI] [PubMed] [Google Scholar]
- 39.Grayson, S., Suver, C., Wilbanks, J. & Doerr, M. Open Data Sharing in the 21st Century: Sage Bionetworks’ Qualified Research Program and Its Application in mHealth Data Release, 10.2139/ssrn.3502410 (2019).
- 40.Bionetworks, S. Synapse. at https://www.synapse.org/#!Synapse:syn10848316/wiki/588047.
- 41.aid4mh. GitHub - aid4mh/gSCAP: Create contextual geospatial features from longitudinal geolocation data. GitHub at https://github.com/aid4mh/gSCAP.
- 42.Omberg, L. et al. Remote smartphone monitoring of Parkinson’s disease and individual response to therapy. Nat. Biotechnol.10.1038/s41587-021-00974-9 (2021). [DOI] [PubMed]
- 43.Pratap A, et al. Evaluating the Utility of Smartphone-Based Sensor Assessments in Persons With Multiple Sclerosis in the Real-World Using an App (elevateMS): Observational, Prospective Pilot Digital Health Study. JMIR Mhealth Uhealth. 2020;8:e22108. doi: 10.2196/22108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Onnela J-P, Rauch SL. Harnessing Smartphone-Based Digital Phenotyping to Enhance Behavioral and Mental Health. Neuropsychopharmacology. 2016;41:1691–1696. doi: 10.1038/npp.2016.7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Omberg L, Chaibub Neto E, Mangravite LM. Data Science Approaches for Effective Use of Mobile Device-Based Collection of Real-World Data. Clin. Pharmacol. Ther. 2020;107:719–721. doi: 10.1002/cpt.1781. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Albert PR. Why is depression more prevalent in women? J. Psychiatry Neurosci. 2015;40:219. doi: 10.1503/jpn.150205. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Pratap A, 2021. Synapse. [DOI]
- 48.Pratap A, 2021. Synapse. [DOI]
- 49.Pratap A, 2021. Synapse. [DOI]
- 50.Pratap A, 2021. Synapse. [DOI]
- 51.Pratap A, 2021. Synapse. [DOI]
- 52.Pratap A, 2021. Synapse. [DOI]
- 53.Pratap A, 2021. Synapse. [DOI]
- 54.Pratap A, 2021. Synapse. [DOI]
- 55.Pratap A, 2021. Synapse. [DOI]
- 56.Pratap A, 2021. Synapse. [DOI]
- 57.Pratap A, 2021. Synapse. [DOI]
- 58.Pratap A, 2021. Synapse. [DOI]
- 59.Pratap A, 2021. Synapse. [DOI]
- 60.Pratap A, 2021. Synapse. [DOI]
- 61.Pratap A, 2021. Synapse. [DOI]
- 62.Pratap A, 2021. Synapse. [DOI]
- 63.Pratap A, 2021. Synapse. [DOI]
- 64.Pratap A, 2021. Synapse. [DOI]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Citations
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
- Pratap A, 2021. Synapse. [DOI]
Supplementary Materials
Data Availability Statement
The studies used a combination of open-source and proprietary apps to collect observational data and deploy remote interventions via three apps (described above). The source code for tools used for the data collection, such as the passive data collection app (Survetory - used in the V2 study), iPST, and HealthTips (two of the three randomized interventions), is available on the study Github archive (https://github.com/aid4mh/Brighten_Studies_Archive). The gSCAP pipeline used to featurize the raw GPS data from the V2 study is available here (https://github.com/aid4mh/gSCAP). The data cleaning and curation pipeline is also available on GitHub (https://github.com/apratap/BRIGHTEN-Data-Release).