Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2025 May 29.
Published before final editing as: IEEE J Biomed Health Inform. 2023 Nov 29;PP:10.1109/JBHI.2023.3337649. doi: 10.1109/JBHI.2023.3337649

The ChAMP App: A Scalable mHealth Technology for Detecting Digital Phenotypes of Early Childhood Mental Health

Bryn C Loftness 1, Julia Halvorson-Phelan 2, Aisling O’Leary 3, Carter Bradshaw 4, Shania Prytherch 5, Isabel Berman 6, John Torous 7, William E Copeland 8, Nick Cheney 9, Ryan S McGinnis 10, Ellen W McGinnis 11
PMCID: PMC11133764  NIHMSID: NIHMS1983583  PMID: 38019617

Abstract

Childhood mental health problems are common, impairing, and can become chronic if left untreated. Children are not reliable reporters of their emotional and behavioral health, and caregivers often unintentionally under- or over-report child symptoms, making assessment challenging. Objective physiological and behavioral measures of emotional and behavioral health are emerging. However, these methods typically require specialized equipment and expertise in data and sensor engineering to administer and analyze. To address this challenge, we have developed the ChAMP (Childhood Assessment and Management of digital Phenotypes) System, which includes a mobile application for collecting movement and audio data during a battery of mood induction tasks and an open-source platform for extracting digital biomarkers. As proof of principle, we present ChAMP System data from 101 children 4–8 years old, with and without diagnosed mental health disorders. Machine learning models trained on these data detect the presence of specific disorders with 70–73% balanced accuracy, with similar results to clinical thresholds on established parent-report measures (63–82% balanced accuracy). Features favored in model architectures are described using Shapley Additive Explanations (SHAP). Canonical Correlation Analysis reveals moderate to strong associations between predictors of each disorder and associated symptom severity (r = .51–.83). The open-source ChAMP System provides clinically-relevant digital biomarkers that may later complement parent-report measures of emotional and behavioral health for detecting kids with underlying mental health conditions and lowers the barrier to entry for researchers interested in exploring digital phenotyping of childhood mental health.

Index Terms—: Digital health, digital biomarkers, pediatric mental health, mobile health, machine learning, anxiety, adhd, depression, anxiety

I. Introduction

CHILDHOOD mental health problems are common and can become chronic and impairing if left untreated. Children eight and younger are not reliable reporters of their emotional symptoms, making it particularly challenging to assess their emotional and behavioral health. Moreover, caregivers often unintentionally under-report their children’s emotional health, or can exhibit reporting bias based on their own mental health [1]. Because of these challenges in assessment, early childhood researchers have started to explore objective measures of emotional and behavioral health captured directly from a child’s physiology and behaviors.

Successful systems [2], [3] for capturing child physiology and behavior typically leverage mood induction tasks, or behavioral paradigms that “press” for a specific emotional or behavioral response [4]. A child’s response is quantified via a physiological metric (i.e., galvanic skin response, heart rate) measured using specialized equipment (e.g., [5], [6]) and/or via behavioral coding (i.e., tracking facial expressions/body movements in task videos [3], [7]) which requires specialized computer programs (e.g., [8]) and trained personnel [4], [9]. In some cases, quantifiable features responsive to the behavioral paradigms and related to the child’s emotional and behavioral health [10] are then used to supplement clinical assessment [4], [11], [12].

Despite the plethora of research (i.e., [4], [9], [10], [12], [13], [14]) and some clinical utility [4] of objective physiological and behavioral measures during mood induction tasks and their links with early childhood emotional and behavioral health, there are important challenges that limit wide use outside of the original authors’ laboratories. Most of these systems require a) specialized measurement equipment which may be expensive, require expert knowledge to use and interpret, and are cumbersome to wear for young children and/or b) video coding which require personnel effort, and time to train and code reliably. There is an unmet need for a system that can capture objective measures of a child’s physiological and behavioral response to mood inductions that is accessible for patients, researchers, and providers, and could enable the broad use, validation, and iterative development of these approaches.

Digital phenotyping, which leverages metrics available from wearable sensors and smartphones for detecting symptoms and presence of underlying health conditions, holds great promise for addressing this unmet need [15]. While there are some great early examples (e.g., smartphone metrics to detect autism in young children [16]), there is still much to learn about digital phenotypes, particularly in early childhood.

Our prior work has demonstrated that digital phenotyping derived from wearable sensor and audio data during mood induction paradigms can identify young children with internalizing disorders with good accuracy (75%–81%) [17], [18], [19], [20], [21], [22], [23]. However, it can be difficult for researchers to replicate this paradigm as an interdisciplinary team with expertise in wearable sensors, biomechanics, signal processing, data science, and clinical psychology is required to meet the important requirements for stringent validation [24]. We also note that these previous models did not classify specific diagnoses or severity, which are important outputs to consider in a future complementary clinical tool that may leverage similar data streams.

In this work, our team presents the ChAMP (Childhood Assessment and Management of digital Phenotypes) System, which is an accessible smartphone app and open-source platform for collecting and interpreting motion and audio data during mood induction tasks. In so doing, we aim to make standardized, validated digital phenotyping tools accessible to other researchers to enable efficacy and effectiveness testing across diverse populations and use-cases [15], [25], [26]. This approach ultimately allows evidence-based complementary digital mental health tools, such as this system and accompanying models, to get into the hands of patients and providers more swiftly. The technological advances and findings in this work build considerably on past work. Namely, we 1) collect a new sample of over 100 children and their families to evaluate our newly developed ChAMP System, 2) and explore its ability to detect specific diagnoses such as anxiety, depression, and ADHD via machine learning approaches deployed on the data it collects and 3) extend our past work evaluating individual mood-induction tasks by considering data from a combination of tasks and for predicting individual diagnoses, as opposed to general internalizing disorders.

II. Methods

A. The ChAMP System

The ChAMP System standardizes the administration of behavioral mood induction tasks for children (Table I) by providing a repeatable and replicable digital framework, enabling collection of multimodal physiological data at scale and facilitating the discovery of digital phenotypes. As illustrated in Fig. 1, the system is composed of a mobile health application and an open-source interactive data processing pipeline. To initiate the ChAMP System, the research administrator opens the ChAMP app and clicks start for each mood induction task when prompted. The ChAMP app is designed with simple screens indicating expected administrator actions (e.g., start, restart, save) and, large-font countdowns while within active tasks. The app collects continuous motion (via the onboard inertial measurement unit which captures acceleration and angular velocity at a sampling rate of 300 Hz) and audio data (via the onboard microphone at a sampling rate of 8 kHz), makes sounds to signal procedural requirements to the administrator and/or child, ends tasks, and stores organized data for researchers.

TABLE I.

ChAMP Behavior Battery Tasks

Task Approach Speech Bubbles
RDoC Construct Potential Threat Acute Threat Reward
Duration 30 s 3 m 3 m
Description Child walks into darkened room toward covered object Child tells story and is told it will be judged Child plays with bubbles from a machine
Stimuli Rapid beeping from 19 to 25 s to induce anxiety Buzzer sounds at 90 and 160 s to induce startle None
Output Motion data as CSV Audio data as mp3 Motion data as CSV

ChAMP app behavioral battery is composed of approach, speech and bubbles mood-in-duction tasks which can be completed in under seven minutes. Each task addresses a research domain criteria (RDoC) construct associated with childhood mental health.

Fig. 1.

Fig. 1.

To facilitate scalable validation and reproduction of the kid study we make the critical components to using the ChAMP mobile application widely accessible, including: A process for researchers to directly request ChAMP app download to enable their own data collection (left), the protocol for administering the ChAMP behavioral battery (middle), and an interactive open-source analysis script for reproducing biomarker feature sets (right).

The ChAMP App, written in Kotlin, was designed to be downloaded onto an Android smartphone and secured to the child’s lower back using a custom waist belt (belt instructions on the ChAMP Website [27]). The interactive data processing pipeline is implemented in Python in a Jupyter Notebook and automatically extracts features that quantify how a child moves and speaks during each task. Features are extracted from theory-driven phases within the task and provide a summary of movement and vocal data during each phase. Digital biomarkers are defined as features that demonstrate clinical relevance and can be combined to form digital phenotypes of childhood emotional and behavioral health. The platform enables seamless collaboration, so researchers with relevant skills can contribute new algorithms for the community to use.

B. The ChAMP Assessment Battery

Each of the three mood induction tasks (see Table I: Approach [18], [19], [28], Speech [20], [29], [30], and Bubbles [2], [3], [21]) induces an emotionally salient stimuli representing multiple constructs of daily functioning [31], associated with emotional and behavioral problems [6], [10], [32], [33]. Each task is described in detail in the full protocol (see [27]).

C. Theory-Driven Feature Engineering

The Approach task was cut into six 5-second segments that capture each protocol sequence (i.e., walking into room; standing in front of covered object) and separate theorized threat response phases (i.e., potential threat, startle, and response modulation) [18], [34]. The Bubbles task was cut into twelve 15-second segments to explore potential changes in response to a single, repeated reward over time. As time until reward satiation is largely unexplored (outside of food reward) it was unknown whether or when reward satiation would be reached and so a span of seconds (similar, but somewhat longer than methodologies used in MRI reward studies [35]) was chosen, yet was largely exploratory. For these two tasks, angular velocity magnitude and acceleration magnitude were derived from the smartphone’s 3-axis gyroscope and accelerometer. For every segment, the median, sum, standard deviation, variance, and 5th, 25th, 75th, and 95th percentiles of angular velocity and acceleration magnitudes were extracted to summarize the child’s movement within each segment. Additional details are available in [27].

The Speech task was cut into three segments: prior to the first buzzer (0–90 sec), between the first and second buzzer (90–150 sec), and after the second buzzer (150–180 sec). These segments attempt to capture differential aspects of fear including proactive avoidance of upcoming threat (being judged pre buzzers), and reactive escape from startle [36], as well as any changes due to habituation of a repeated stimuli [37] (after the second buzzer). Five seconds of audio at the beginning and end of the task and before and after each buzzer were removed to reduce the likelihood of any administrator audio being captured (i.e., giving instruction and/or time left). The WebRTC voice activity detector (VAD) was used to isolate the child’s voice in the recording for analysis [38] following which recording volume was increased by 15 dB to ensure we were accommodating for the posterior positioning of the phone. Audio features were selected from our previous work (e.g., (zero crossing rate (ZCR), Mel frequency cepstral coefficients (MFCC), spectral centroid [20]) and expanded to include proportion of time spent speaking in each task segment and throughout the whole task, chromagram (chroma stft), root mean square, spectral bandwidth, and the approximate minimum and maximum frequency. We also assessed the variance and mean pitch intensity, as well as the ranked list of dominant pitch intensities, all extracted using chroma stft [39]. We developed the ranked-dominant pitch intensity feature by summing the intensity of each pitch across the segment, and then ranking the pitches from 0–1 by the pitch intensity contribution, 0 being more dominant (e.g., mean ranked-dominant pitch intensity of C♯ in diagnosed children was .727 versus for undiagnosed children it was .454, indicating the dominant pitch for undiagnosed children was more likely ranked higher for C♯ than diagnosed children). Descriptive statistics (variance and/or mean) were computed for each feature. Additional details are available in [27].

D. Study Recruitment and Experimental Protocol

Children and their primary caregiver were recruited as participants in Fall 2021–2022 from a state-wide sample via advertisements (n = 104) as well as in collaboration with the Vermont Center for Children, Youth and Families (VCCYF) in the Division of Child Psychiatry at the University of Vermont (n = 6). To meet inclusion criteria, children had to be aged 4–8 years, with no suspected or diagnosed developmental disorders, serious medical conditions, or cognitive impairments, and primary caregivers had to be at least 18 years old and English-speaking.

For this analysis, 104 children between ages 4 and 8 years were included (M = 81.42 months, SD = 14.50), 61.5% (n = 62) were male, and 89.4% (n = 93) were White, non-Hispanic. Most reporting caregivers (80.2% biological mothers, n = 77) were married (82.6%, n = 71, 18 missing), had attained a bachelor’s degree or higher (88% n = 75, 19 missing) and 81% (n = 69, 19 missing) had average yearly household income of $75,000 or more. During the study lab visit, children were assessed for mental health diagnoses of which 47.6% (n = 49) met criteria for at least one disorder of anxiety, depression or ADHD. Overall, 33% (n = 34) had an anxiety disorder (inclusive of generalized anxiety disorder, separation anxiety disorder, social phobia, anxiety-unspecified, and post-traumatic stress disorder); 13.6%, (n = 14) had a depressive disorder (inclusive of major depressive disorder, persistent depressive disorder and depressive disorder-unspecified); 19.4% (n = 20) had a diagnosis of ADHD (inattentive, hyperactive or combined), including 11 children (10%) with an anxiety or depressive disorder (Fig. 3). The range of disorders in this sample fit recommendations of recruiting participants with less specific clinical criteria, and without a “super-normal” control group to better represent general populations [40].

Fig. 3.

Fig. 3.

Breakdown of participant diagnostic results by sex.

Eligible children and their primary caregivers attended a 2.5-hour laboratory visit. During the visit, the child’s caregiver was consented, administered a clinical interview about their child’s mental health (Schedule for Affective Disorders and Schizophrenia Preschool Version, present and lifetime) [41], [42] and completed several questionnaires about their child’s mental health and development. Diagnoses were derived via clinical consensus with a supervising licensed clinical psychologist using the best-estimate procedures based on parent-report symptom checklists and family history.

Simultaneously, the child was outfitted with sensors and a mobile phone (Fig. 2(a)). The child was led to a novel room by an administrator to engage in the ChAMP Behavior Battery (Fig. 2(b)). Afterward the child completed a structured intelligence test and played with toys until their caregiver completed their tasks. Families were compensated with a $60 gift card and a $5 toy. Families were called by a licensed clinical psychologist to discuss the results of their child’s mental health assessment and referral recommendations within two weeks of their visit. Study procedures were approved by the University of Vermont Institutional Review Board (CHRBSS 00001218).

Fig. 2.

Fig. 2.

Participating children and their primary caregiver participate in a ~2.5-hour study visit. The caregiver completes a clinical interview to assess their child’s mental health status ((a), bottom left) while the child is outfitted with several wearable sensors including 5 BioStamp sensors, an E4 Empatica wrist sensor, and the belt-worn android smartphone with the downloaded app, ChAMP ((a), right) and completes the ChAMP behavioral assessment (b).

E. Analytic Methods

Raw signal data needs context to interpret, and the theory driven phases used to segment each task provides this needed context. However, prior to assessment of theory driven phases, a determination of inclusion or exclusion of demographic information is needed for model feature selection. We performed t-tests to determine associations between age and disorder presence. We also used t-testing to assess if child sex was related to symptom severity for children who endorsed at least one symptom. We performed Chi squares testing to determine relationships between child sex and disorder presence. We then evaluated for correlations between age and symptom severity using Spearman correlations.

Preparing to dig deeper into understanding the relationship between the movement and audio features extracted from these tasks and emotional and behavioral health, we partitioned the dataset by task and by availability of data. This process produced 5 separate feature sets from the original 104 subjects, each were evaluated separately through model building efforts; Approach Task only (94 subjects), Bubbles Task only (99 subjects), Speech Task only (99 subjects), all three completed tasks (94 subjects), and 5-nearest neighbors imputation for tasks for subjects who were missing one or more tasks (101 subjects). Three subjects were not included in analysis as they had no sensor data. While in previous work we considered task-specific models, this is the first dataset we have considered that combined multiple valance tasks into one cohesive model.

To assess the potential for the ChAMP System to detect underlying mental health conditions in young children, we trained several binary classification models. Models were trained to identify children with depression, anxiety, or ADHD based on age, gender, and the 5 differing feature sets from the ChAMP System. To begin, we examine relative performance of Gaussian Naïve Bayes (NB), Support Vector Machine with Stochastic Gradient Descent Booster (SVM), Decision Tree (DT), XGBoost (XGB), and Logistic Regression (LR) classification models with varying infrastructures and feature selection methods. Specifically, we considered each classifier, with and without hyperparameter tuning, with and without decision thresholding weighted on balanced accuracy, trained on either 10 or 20 of the top predictors from 4 different feature importance methods. The feature importance methods considered ranked and selected predictors considering entropy-based Shannon information gain (DT), GINI impurity (GINI), permutation importance (PERM), and XGBoost feature importance (XGB). Models were trained and evaluated using stratified 2-fold shuffled cross-validation and characterized by the mean balanced accuracy (B-Accuracy), mean area under curve for the receiver operating curve (ROC AUC), sensitivity, specificity, false positive rate (FPR), and true positive rate (TPR) averaged across the two folds’ test cases. The primary metric defining performance ranking was balanced accuracy. Two folds were selected to maintain sufficient training instances for diagnoses that have small sample sizes (Fig. 4(c)). For models predicting ADHD, there were 9–10 diagnosed individuals per fold. For anxiety, 15–16 diagnosed children per fold. For depression, there were 6 diagnosed children per fold. Evaluations of feature importance occurred within each fold of the cross validation. The feature set sizes of 10 and 20 were selected to limit model overfitting caused by too many considered features and to narrow down the modalities yielding the most influential biomarkers.

Fig. 4.

Fig. 4.

A. Highest performing binary diagnostic classification models for true clinical consensus prediction using ChAMP features and demographic features (sex and age in months). B. Parent report questionnaire (CBCL) concordance with true clinical consensus C. Breakdown of subject counts by task and diagnostic classification for the approach task (top, green door), speech task middle, (middle, blue speech bubble) and the bubbles task (bottom, pink bubbles icon). Top 10 predictors as defined by shapley additive explanations for overall model contributions in each binary diagnostic model for one fold of the highest performing model for ADHD ((d) left), anxiety ((e) left), and depression ((f) top). Canonical correlation plots for 1-component CCA comparing the canonical variable for the diagnostic predictors in each of the top moels compared to diagnostic symptom count for ADHD ((d) right), anxiety ((e) left), and depression ((f) bottom). NB = naïve bayes, XGB = XGBoost, PERM = permutation feature importance, DT = entropy-based shannon information gain. FPR = false positive rate. TPR = true positive rate. D = diagnosed. ND = not diagnosed. N = number of features selected. D-threshold = decision threshold value. HP tuning = hyperparameter tuning. ANX = anxiety, DEP = depression, B-accuracy = balanced accuracy. KNN = K-nearest neighbors.

For models that were tuned, we selected a subset of parameters for each classifier to consider. For SVM, parameters of learning_rate, loss, penalty, fit_intercept, and alpha were tuned. For the DT classifier, max_features, max_depth, criterion, min_samples_leaf, and min_samples_split, were considered. For the NB classifier, we selected only var_smoothing. For XGB, max_depth, min_child_weight, gamma, subsample, and reg_alpha. For LR, penalty, C, and class_weight were tuned. Top performing models following model evaluations were then explored using SHAP (SHapley Additive exPlanations) [43] and compared to foundational work. Features identified in top performing models were then evaluated using one-component Canonical Correlation Analysis to assess the relationship between these top predictors and diagnostic symptom counts, a measure of symptom severity, for children who had at least one symptom.

III. Results

A. Demographic Associations With Diagnosis and Severity

Child age was not related to the presence of Anxiety (t(101) = .−1.88, p = .064) or ADHD (t(101) = .−.738, p = .462), but related to Depression (t(101) = .−2.88, p = .005) such that children with a diagnosis were older than those without. Child sex was unrelated to the presence of Anxiety (x(1,103) = .117, p = .732) ADHD (x(1,103) = .154, p = .695), and Depression (x(1,103) = .719, p = .397). For children with at least one symptom within diagnosis, child age was unrelated to symptom severity for Anxiety (r(41) = .089, p = .578), or ADHD (r(29) = −.018, p = .927), but was related to depression (r(38) = .356, p = .028). Child sex was related to symptom severity for Anxiety (t(39) = 2.178, p = .036), such that females exhibited more symptoms than males, but child sex was unrelated to ADHD (t(27) = −.160, p = .874), or depression (t(37) = −.912, p = .368) symptoms. Based on several significant relationships found here, we included child sex and age in months in all model datasets.

B. Predictive Modeling of Mental Health Diagnosis

We build several binary models classifying each diagnosis (anxiety, depression, ADHD) using only objective child physiology and behavior which would not require caregiver reporting that can be dependent on caregiver-child relationship, contexts for observations, caregiver mental health, mental health literacy, and societal stigma. All of the top performing models for each diagnostic classification (Fig. 4(a)) favored feature sets that included features across all three behavioral tasks. Notably, each of these top models favored infrastructures that included decision thresholding for optimizing balanced accuracy and two of the three included hyperparameter tuning. All top performing models were able to correctly classify at least 50% of the true positives within the dataset and all models achieved balanced accuracy rates between 70 and 73%.

The top performing thresholds were all found to be lowered from the default rate (.5) to a threshold ranging from .12 to .32. This finding is interesting, as typically lowering a threshold so dramatically would cause an increased risk of false positives. However, the models we present still exhibit false positive rates in similar ranges to parent report (.08–.35 vs .1 to .26).

In previous work, we built several models considering the ability of features derived from each of the individual tasks in the ChAMP mood induction battery to classify the presence of an internalizing diagnosis (depression and/or anxiety) [21]. Sensitivities for top performing models for each task in previous work are similarly or less than the sensitivities achieved for classifying specific disorders (ADHD, depression, and anxiety) herein (.39–.67 vs .5–.77). For example, the model we present for anxiety has better sensitivity than any of the internalizing disorder models we have presented in prior work. Specificities of single-task models for internalizing disorders ranged from .64 to .93 in our prior work. In this work on multi-task diagnostic-specific models, have specificities ranging from .65 to .92.

C. Evaluating Potential as a Complementary System

Consistent with our previous work, we compared psychometric evaluations of the questionnaire-based parent-reported CBCL with our smartphone-based movement and vocal feature models on child internalizing and externalizing diagnosis as determined via K-SADS-PL with clinical consensus. Clinical thresholds on the parent-report CBCL Externalizing Score detect ADHD with lower balanced accuracy (.69 vs. .73) and specificity (.84 vs. .92), but slightly higher sensitivity (.55 vs. .53) than models based on movement and vocal biomarkers (Fig. 4(a) and (b)). Clinical thresholds on the parent-reported CBCL Internalizing Score for detecting depression exhibit lower balanced accuracy (.63 vs. .71), sensitivity (.36 versus .5), and specificity (.9 vs .91), higher false positive rate (.1 vs .09) and lower true positive rate (.36 versus .5) than models based on movement and vocal biomarkers (Fig. 4(a) and (b)). Clinical thresholds on parent-reported CBCL Internalizing Score for detecting anxiety exhibited superior performance than the top performing ChAMP model in all metrics.

Our current and past studies have paralleled previous work demonstrating varied sensitivity of the CBCL, ranging from .0–.38 [44] to as high as .44 to .86 [45]. In our past work, biomarker-based models yielded a minimum 60% improvement in sensitivity over CBCL-based concordance to consensus report. In this work, we observed higher performance of CBCL-based concordance compared to previous work (.36–.91 vs. 00–.42 [23]) and thus sensitivities observed in ChAMP-based models did not yield such significant improvements relative to CBCL (.5–.74 vs .36–.91). The CBCL psychometrics observed here are similar to, and for some diagnostic groups higher than, those in larger studies [46], [47]. Interestingly, anxiety in this sample of parents was reported with sensitivities higher than previously reported in any of our teams’ past work (.91). Notably, specificity of CBCL-based concordance in this study was similar to prior work (.88–1 [23] vs. .74–.90). ChAMP-based models we engineered performed with nearly parallel specificity to the observed CBCL-concordance in this study (.65–.92 vs. .74–.90).

In exploratory post-hoc analyses, we compared model false negatives (FNs) to true positives (TPs) on demographics (sex, age, race, income) and symptom characteristics (ADHD, depression and anxiety symptom counts and diagnostic subtypes), to investigate reasons the models may be missing some children with diagnoses. For ADHD, TPs were significantly younger than FPs (t(17) = −2.13, p = .049). For Depression, TPs may exhibit more ADHD symptoms than FPs (t(5.03) = 2.10, p = .089), but did not reach statistical significance. For anxiety, there appeared to be no differences between TPs and FPs. Perhaps anxiety should be analyzed as subtypes (i.e., specific anxiety diagnoses, or fear vs. distress) in future analyses to improve model specificity. There were no significant differences between TNs and FNs for any diagnosis.

For CBCL internalizing reports, TPs exhibited more anxiety symptoms according to the KSADS interview than FP (t(38) = 2.09, p = .043) and there were differences by group in terms of caregiver education. Caregivers with higher education were more likely to fall in the FP than the TP group (t(31) = −2.14, p = .040). For CBCL externalizing reports on KSADS ADHD diagnoses (which is not a direct comparison of ADHD), FPs were more likely to be boys than TPs (x(1,20) = 3.78, p = .052) and there was a significant difference by group in terms of household income. Caregivers with higher income were more likely to be in the FP than the TP group (t(13) = −2.21, p = .045). Overall, analyses of subtypes may suggest that caregiver report may be influenced by caregiver demographics, and that physiological/behavioral models still need more analyses to understand better what is driving FPs.

We survey model explainability using Shapley Additive Explanations (SHAP)-based feature contributions [43]. All models selected features originating from more than one of the three behavioral tasks, rather than favoring a task-specific model. Interestingly, the ADHD model was the only of the top models to favor 20 instead of 10 features, and the top 10 features (Fig. 4(d)) nearly all originated from the Bubbles task. The top performing Depression model included both of the two demographic features (Age and Sex), as well as features from both the Approach and Bubbles tasks. The Anxiety top performing model included both demographic features, as well as features across all three tasks.

D. Child Diagnosis and Symptom Severity

Considering these top 10 to 20 features (depending on model), we performed Canonical Correlation Analysis (CCA) to evaluate the relationship between these predictors and symptom counts from the K-SADS-PL. We found that all predictor combinations yielded moderate to strong correlations (r = .51–.83) to diagnostic severity (defined by symptom count). This finding supports the hypothesis that theory-driven biomarkers may complement current screening by not only assisting in diagnostic classification, but also by potentially screening for estimated severity of an underlying disorder.

While these results are promising and show the potential of the ChAMP tool to collect valuable biomarkers for classifying childhood mental health status, significant future work is needed to engineer additional theory-driven features, integrate other physiological signals, and continue to explore modeling approaches to optimize performance. These results imply that each task in the ChAMP battery provides important, independent, and potentially clinically relevant, information for characterizing emotional and behavioral health that could complement parent report. Future work should also be done to evaluate how combinations of parent-report with wearable-sensor and smartphone-based signals may improve clinical capability by providing a complementary data stream for identifying and intervening on early childhood mental health concerns in young children.

IV. Discussion

Through the introduction of the ChAMP system, we provide an accessible, digital tool to collect physiological and behavioral data during brief, structured mood induction tasks and an open-source platform to engineer objective features related to childhood emotional and behavioral health. We provide promising results suggesting that even some of the simplest movement and vocal features from these tasks could serve as biomarkers for childhood mental health.

A. Relationship Within and Across Tasks

Our feature analyses demonstrate several key findings for future use. First, it is important to use a variety of multimodal features to model the complexity of child behavior during this battery. This implies that machine learning techniques which allow for modeling these complex relationships will likely be the best analytical strategy for these data [1], [19], [20]. Herein, we used only high-level features. However, based on our previous work, it is likely that that more granular biomechanical features may provide improved model performance [23]. Thus, we aim to explore a greater variety and complexity of features (e.g., turning behavior, postural sway [48]) in future work. Second, engineering features from temporal segments within each task is crucial as results revealed that specific task segments are differentially important in classifying children with and without a disorder (Fig. 4(d)(f)). While behaviorally coded features from mood induction tasks are typically summed across one (or multiple) tasks [3], [10], our task protocols and measurements were developed and adapted to evoke and quantify differential magnitudes of emotional stimuli across time. Thus, including all three mood induction tasks as one behavior battery is important as features from each appear to contribute uniquely to disorder classification. These findings are important in the context of our past work, as we have never before combined results across mood-induction tasks.

B. Association With Diagnoses and Symptom Counts

In our prior work in a separate sample, children with a diagnosis turned away more and more quickly from an unknown threat in the Approach task compared to children without a diagnosis [18]. We observed this phase showed importance in this new sample as well, especially for the classification of Depression and Anxiety. These heightened features (herein, median and sum of movement magnitude, and previously, more tilt/turn of the child’s body) could represent biomarkers of elevated attentiveness and responsiveness to potential threat, consistent with work demonstrating children with internalizing disorders attended more to potentially threatening faces than those without [33]. Second, we found higher dominant pitch tone after an unexpected startle in the Speech task in children with an Anxiety diagnosis, consistent with our previous work in a separate sample [20]. These heightened biomarkers represent elevated reactivity to acute threat, consistent with work demonstrating adults with fear-related anxiety disorders startle more (higher magnitude eye blink) than those without [6]. Third, we found that the Bubbles task was dominantly representative across top features in the diagnostic prediction models. This supports our previous work demonstrating that the Bubbles task biomarkers were more related to broad emotional dysregulation than specific diagnostic symptoms [21].

We have extended our past work by implementing a variety of supervised classification models which further demonstrate the importance of including multi-modal features across temporal phases, and tasks as multiple modalities, phases, and tasks were included in each optimal diagnostic model. These models perform on-par with, or improve upon, several of our previous models presented in foundational efforts and incorporate several mood induction tasks in a structured battery. We also demonstrate, for the first time, that linear combinations of top-performing ChAMP features during these tasks are strongly associated with symptom counts of child anxiety, depression, and, most notably, ADHD. Future work should consider additional wearable sensors, features, subjects, and modeling approaches (e.g., deep learning or transfer learning).

A strength of this study is the inclusion of a wide range of internalizing disorders, while not excluding externalizing disorders. This allowed us to advance modelling efforts to specialize classification to specific diagnoses under the internalizing and externalizing umbrella, which we previously had not evaluated. This may make our small sample more clinically representative of community populations. Overall, findings show promise that specific segments of each of the three mood induction tasks may contribute to future models for detecting childhood anxiety and depression and for assessing severity of underlying diagnosis.

It is also important to note potential limitations relating to bias. We have found that our models less clearly portray biases related to demographic background in parent-report surveys on child mental health status, however we cannot affirm that the models built are without bias. Notably, our sample may not represent the general population in terms of demographics or impairments. Our sample was less racially and ethnically diverse, and more affluent than that US population. Families of children with impairment appeared more likely than those without to self-select into the study. During recruitment, community waitlists for diagnostic evaluations were long (>9 months) and receiving a diagnostic assessment for their child within two weeks through participation in the study was reported as a significant motivator for families. It is also possible that the emotional impact of COVID anxiety/isolation may have increased child impairment as seen in adolescent samples [49]. We aim in future work to further bolster our training samples with a wider diversity of children and families.

C. ChAMP as an Open-Access Digital Tool

In the past decade, we have seen a rise in validated mobile applications and platforms that aim to increase awareness, monitoring, and intervention of mental illness (e.g., PanicMechanic [50], [51], LAMP [25], FOCUS [52]). There is an increase in commercial and research platforms that incorporate aspects of the digital phenotyping lifecycle (i.e., study protocol portal, data collection tool, mechanism for storing and accessing data, data analysis tools), but very few feature all of the necessary components and data to replicate and validate findings [15]. ChAMP provides accessibility and flexibility in these components, and for the first time, with a focus on early childhood. We hope with this platform that other researchers may extend this work to explore intrasubject variability and repeatability of findings across additional participant groups.

By offering an app and platform, we aim to aid researchers in advancing the field of digital psychiatry [53]. We aim to increase explainability for this toolkit, and accessibility for approved clinical researchers utilizing the toolkit, by providing a website with additional documentation [27]. Future improvements to ChAMP will increase the customizability of our platform to enhance research and future clinical utility [15], [53]. Our next steps toward a customizable tool would allow investigation into additional conditions impacting children across a wider range of ages. Potential modifications could include additional mood induction tasks, project-specific enhancements to app protocols, and companion wearable devices for enabling additional sensing modalities. For instance, autism is an area where movement and speech data has often been used for assessment of diagnostic status and severity [54], [55] as well as ADHD [56], [57]. With more replication and validation across samples by clinical researchers, precision and model performance can be further enhanced. This is important, as a future iteration of a similar objective system must have high accuracy so as to not overburden the system or family with false positive reports or under-identify children that may be struggling un-noticed. Moreover, this system, once further refined, could have potential to ultimately improve assessment by serving as a complementary clinical endpoint to better inform intervention.

Acknowledgment

This work would not be possible without the wonderful participating families and support of the undergraduate research assistants conducting lab visits. All figures were created using Biorender.com.

This work was supported in part by the US NSF under Grant 2046440 and in part by the US NIH under Grant MH123031.

This work involved human subjects or animals in its research. Approval of all ethical and experimental procedures and protocols was granted by the University of Vermont IRB under Application No. Study00001218, and performed in line with the declaration of Helsinki.

Contributor Information

Bryn C. Loftness, University of Vermont’s Complex Systems Center and M-Sense Research Group, Burlington, VT 05405 USA.

Julia Halvorson-Phelan, University of Vermont Medical Center Department of Psychiatry, Burlington, VT 05405 USA.

Aisling O’Leary, University of Vermont M-Sense Research Group, Burlington, VT 05405 USA.

Carter Bradshaw, University of Vermont Medical Center Department of Psychiatry, Burlington, VT 05405 USA.

Shania Prytherch, University of Vermont Medical Center Department of Psychiatry, Burlington, VT 05405 USA.

Isabel Berman, University of Vermont Medical Center Department of Psychiatry, Burlington, VT 05405 USA.

John Torous, Digital Psychiatry Division for Beth Israel Deaconess Medical Center, Harvard Medical School, Boston, MA 02115 USA.

William E. Copeland, University of Vermont Medical Center Department of Psychiatry, Burlington, VT 05405 USA

Nick Cheney, University of Vermont Complex Systems Center, Burlington, VT 05405 USA.

Ryan S. McGinnis, University of Vermont M-Sense Research Group, Burlington, VT 05405 USA

Ellen W. McGinnis, University of Vermont Medical Center Department of Psychiatry, Burlington, VT 05405 USA

References

  • [1].McGinnis EW, Copeland W, Shanahan L, and Egger HL, “Parental perception of mental health needs in young children,” Child Adolesc. Ment. Health, vol. 27, no. 4, pp. 328–334, Nov. 2022, doi: 10.1111/camh.12515. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Hurwitz S and Yirmiya N, “Autism diagnostic observation schedule (ADOS) and its uses in research and practice,” in Comprehensive Guide to Autism Berlin/Heidelberg, Germany: Springer Sci. Bus. Media, 2014, pp. 345–353. [Google Scholar]
  • [3].Gagne JR, Van Hulle CA, Aksan N, Essex MJ, and Goldsmith HH, “Deriving childhood temperament measures from emotion-eliciting behavioral episodes: Scale construction and initial validation,” Psychol. Assessment, vol. 23, no. 2, pp. 337–353, Jun. 2011, doi: 10.1037/a0021746. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [4].Lord C et al. , “The autism diagnostic observation schedule-generic: A standard measure of social and communication deficits associated with the spectrum of autism,” J. Autism Devlop. Disord, vol. 30, no. 3, pp. 205–223, Jun. 2000. [PubMed] [Google Scholar]
  • [5].Sousa AC, Ferrinho SN, and Travassos B, “The use of wearable technologies in the assessment of physical activity in preschool- and school-age youth: Systematic review,” Int. J. Environ. Res. Public Health, vol. 20, no. 4, Jan. 2023, Art. no. 3402, doi: 10.3390/ijerph20043402. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [6].McTeague LM and Lang PJ, “The anxiety spectrum and the reflex physiology of defense: From circumscribed fear to broad distress,” Depression Anxiety, vol. 29, no. 4, pp. 264–281, Apr. 2012, doi: 10.1002/da.21891. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [7].Durbin CE, Hayden EP, Klein DN, and Olino TM, “Stability of laboratory-assessed temperamental emotionality traits from ages 3 to 7,” Emotion, vol. 7, no. 2, pp. 388–399, May 2007, doi: 10.1037/1528-3542.7.2.388. [DOI] [PubMed] [Google Scholar]
  • [8].Kring AM and Sloan DM, “The facial expression coding system (FACES): Development, validation, and utility,” Psychol. Assessment, vol. 19, no. 2, pp. 210–224, 2007. [DOI] [PubMed] [Google Scholar]
  • [9].Goldsmith HH and Rothbart MK, “Prelocomotor and locomotor laboratory temperament assessment battery, lab-tab; version 3.0. technical manual,” Dept. Psychol., Univ. Wisconsin, Madison, WI, USA, 1996. [Google Scholar]
  • [10].Moser JS, Durbin CE, Patrick CJ, and Schmidt NB, “Combining neural and behavioral indicators in the assessment of internalizing psychopathology in children and adolescents,” J. Clin. Child Adolesc. Psychol.: Official J. Soc. Clin. Child Adolesc. Psychol., Amer. Psychol. Assoc, vol. 44, no. 2, pp. 329–340, Jan. 2014, doi: 10.1080/15374416.2013.865191. [DOI] [PubMed] [Google Scholar]
  • [11].Kochanska G, Murray K, and Coy KC, “Inhibitory control as a contributor to conscience in childhood: From toddler to early school age,” Child Devlop, vol. 68, no. 2, pp. 263–277, Apr. 1997. [PubMed] [Google Scholar]
  • [12].Callender KA, Olson SL, Kerr DCR, and Sameroff AJ, “Assessment of cheating behavior in young school-age children: Distinguishing normative behaviors from risk markers of externalizing psychopathology,” J. Clin. Child Adolesc. Psychol, vol. 39, no. 6, pp. 776–788, Nov. 2010, doi: 10.1080/15374416.2010.517165. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [13].Wakschlag LS et al. , “Defining the ‘disruptive’ in preschool behavior: What diagnostic observation can teach us,” Clin. Child Fam. Psychol. Rev, vol. 8, no. 3, pp. 183–201, Sep. 2005, doi: 10.1007/s10567-005-6664-5. [DOI] [PubMed] [Google Scholar]
  • [14].Yáñez-Téllez G et al. , “Cognitive and executive functions in ADHD,” Actas Españolas Psiquiatría, vol. 40, no. 6, pp. 293–298, Nov. 2012. [PubMed] [Google Scholar]
  • [15].Torous J, Kiang MV, Lorme J, and Onnela J-P, “New tools for new research in psychiatry: A scalable and customizable platform to empower data driven smartphone research,” J. Med. Internet Res. Ment. Health, vol. 3, no. 2, May 2016, Art. no. e5165, doi: 10.2196/mental.5165. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [16].Perochon S et al. , “A scalable computational approach to assessing response to name in toddlers with autism,” J. Child Psychol. Psychiatry, vol. 62, no. 9, pp. 1120–1131, 2021, doi: 10.1111/jcpp.13381. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [17].Loftness BC, Halvorson-Phelan J, O’Leary A, Cheney N, McGinnis EW, and McGinnis RS, “UVM KID study: Identifying multi-modal features and optimizing wearable instrumentation to detect child anxiety,” in Proc. IEEE 44th Annu. Int. Conf. Eng. Med. Biol. Soc, 2022, pp. 1141–1144, doi: 10.1109/EMBC48229.2022.9871090. [DOI] [PubMed] [Google Scholar]
  • [18].McGinnis EW et al. , “Movements indicate threat response phases in children at risk for anxiety,” IEEE J. Biomed. Health Inform, vol. 21, no. 5, pp. 1460–1465, Sep. 2017, doi: 10.1109/JBHI.2016.2603159. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [19].McGinnis EW et al. , “Wearable sensors detect childhood internalizing disorders during mood induction task,” PLoS One, vol. 13, no. 4, 2018, Art. no. e0195598, doi: 10.1371/journal.pone.0195598. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [20].McGinnis EW et al. , “Giving voice to vulnerable children: Machine learning analysis of speech detects anxiety and depression in early childhood,” IEEE J. Biomed. Health Inform, vol. 23, no. 6, pp. 2294–2301, Nov. 2019, doi: 10.1109/JBHI.2019.2913590. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [21].McGinnis EW et al. , “Digital phenotype for childhood internalizing disorders: Less positive play and promise for a brief assessment battery,” IEEE J. Biomed. Health Inform, vol. 25, no. 8, pp. 3176–3184, Aug. 2021, doi: 10.1109/JBHI.2021.3053846. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [22].Loftness BC et al. , “Toward digital phenotypes of early childhood mental health via unsupervised and supervised machine learning,” in Proc. IEEE Annu. Int. Conf. Eng. Med. Biol. Soc, 2023, pp. 1–4, doi: 10.1101/2023.02.24.23286417. [DOI] [PubMed] [Google Scholar]
  • [23].McGinnis RS et al. , “Rapid detection of internalizing diagnosis in young children enabled by wearable sensors and machine learning,” PLoS One, vol. 14, no. 1, Jan. 2019, Art. no. e0210267, doi: 10.1371/journal.pone.0210267. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [24].Lagan S et al. , “Mental health app evaluation: Updating the American psychiatric association’s framework through a stakeholder-engaged workshop,” Psychiatr. Serv, vol. 72, no. 9, pp. 1095–1098, Sep. 2021, doi: 10.1176/appi.ps.202000663. [DOI] [PubMed] [Google Scholar]
  • [25].Torous J et al. , “Creating a digital health smartphone app and digital phenotyping platform for mental health and diverse healthcare needs: An interdisciplinary and collaborative approach,” J. Technol. Behav. Sci, vol. 4, no. 2, pp. 73–85, Jun. 2019, doi: 10.1007/s41347-019-00095-w. [DOI] [Google Scholar]
  • [26].Wisniewski H, Henson P, and Torous J, “Using a smartphone app to identify clinically relevant behavior trends via symptom report, cognition scores, and exercise levels: A case series,” Front. Psychiatry, vol. 10, 2019, Art. no. 652, doi: 10.3389/fpsyt.2019.00652. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [27].“ChAMP homepage — ChAMP, A scalable mental health screening app using mood induction tasks,” Sep. 2022. Accessed: Jun. 1, 2023. [Online]. Available: https://brynchristineloftness.github.io/CHAMP/intro.html
  • [28].Lopez-Duran NL, Hajal NJ, Olson SL, Felt BT, and Vazquez DM, “Individual differences in cortisol responses to fear and frustration during middle childhood,” J. Exp. Child Psychol, vol. 103, no. 3, pp. 285–295, Jul. 2009, doi: 10.1016/j.jecp.2009.03.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [29].Kendall PC, Flannery-Schroeder E, Panichelli-Mindel SM, Southam-Gerow M, Henin A, and Warman M, “Therapy for youths with anxiety disorders: A second randomized clinical trial,” J. Consulting Clin. Psychol, vol. 65, no. 3, pp. 366–380, Jun. 1997. [DOI] [PubMed] [Google Scholar]
  • [30].Bae YJ et al. , “The hyporeactivity of salivary cortisol at stress test (TSST-C) in children with internalizing or externalizing disorders is contrastively associated with α-amylase,” J. Psychiatr. Res, vol. 71, pp. 78–88, Dec. 2015, doi: 10.1016/j.jpsychires.2015.09.013. [DOI] [PubMed] [Google Scholar]
  • [31].“Research domain criteria (RDoC),” National Institute of Mental Health (NIMH). Accessed: Jan. 11, 2024. [Online]. Available: https://www.nimh.nih.gov/research/research-funded-by-nimh/rdoc
  • [32].Baji I et al. , “[Symptoms of depression in children and adolescents in relation to psychiatric comorbidities],” Magyar Pszichiátriai Társaság Tudományos Folyóirata, vol. 27, no. 2, pp. 115–126, 2012. [PubMed] [Google Scholar]
  • [33].Salum GA et al. , “Threat bias in attention orienting: Evidence of specificity in a large community-based study,” Psychol. Med, vol. 43, no. 4, pp. 733–745, 2013. [DOI] [PubMed] [Google Scholar]
  • [34].Gross JJ, “Emotion regulation: Affective, cognitive, and social consequences,” Psychophysiology, vol. 39, no. 3, pp. 281–291, May 2002, doi: 10.1017/S0048577201393198. [DOI] [PubMed] [Google Scholar]
  • [35].O’Doherty JP, “Reward representations and reward-related learning in the human brain: Insights from neuroimaging,” Curr. Opin. Neurobiol, vol. 14, no. 6, pp. 769–776, Dec. 2004, doi: 10.1016/j.conb.2004.10.016. [DOI] [PubMed] [Google Scholar]
  • [36].LeDoux JE, Moscarello J, Sears R, and Campese V, “The birth, death and resurrection of avoidance: A reconceptualization of a troubled paradigm,” Mol. Psychiatry, vol. 22, no. 1, pp. 24–36, Oct. 2016, doi: 10.1038/mp.2016.166. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [37].Sege CT, Taylor DL, Lopez JW, Fleischmann H, White EJ, and McTeague LM, “Coping in the clinic: Effects of clinically elevated anxiety on dynamic neurophysiological mechanisms of escape/avoidance preparation,” Biol. Psychiatry: Cogn. Neurosci. Neuroimaging, vol. 8, no. 7, pp. 712–719, Jul. 2023, doi: 10.1016/j.bpsc.2022.07.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [38].Wiseman J, “wiseman/py-webrtcvad,” 2022. Accessed: Dec. 16, 2022. [Online]. Available: https://github.com/wiseman/py-webrtcvad
  • [39].“librosa.feature.chroma_stft — Librosa 0.10.0.dev0 documentation,” Accessed: Jan. 12, 2024. [Online]. Available: https://librosa.org/doc/main/generated/librosa.feature.chroma_stft.html
  • [40].Cuthbert BN and Insel TR, “Toward the future of psychiatric diagnosis: The seven pillars of RDoC,” BioMed Central Med, vol. 11, 2013, Art. no. 126, doi: 10.1186/1741-7015-11-126. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [41].Birmaher B et al. , “Schedule for affective disorders and schizophrenia for school-age children (K-SADS-PL) for the assessment of preschool children- a preliminary psychometric study,” J. Psychiatr. Res, vol. 43, no. 7, pp. 680–686, Apr. 2009, doi: 10.1016/j.jpsychires.2008.10.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [42].Kaufman J, Birmaher B, Brent DA, Ryan ND, and Rao U, “K-SADS-PL,” J. Amer. Acad. Child Adolesc. Psychiatry, vol. 39, no. 10, Oct. 2000, Art. no. 1208, doi: 10.1097/00004583-200010000-00002. [DOI] [PubMed] [Google Scholar]
  • [43].Lundberg SM and Lee S-I, “A unified approach to interpreting model predictions,” in Proc. Adv. Neural Inf. Process. Syst, 2017, pp. 4–5. [Online]. Available: https://papers.nips.cc/paper_files/paper/2017/hash/8a20a8621978632d76c43dfd28b67767-Abstract.html [Google Scholar]
  • [44].Rishel CW, Greeno C, Marcus SC, Shear MK, and Anderson C, “Use of the child behavior checklist as a diagnostic screening tool in community mental health,” Res. Social Work Pract, vol. 15, no. 3, pp. 195–203, May 2005, doi: 10.1177/1049731504270382. [DOI] [Google Scholar]
  • [45].Aschenbrand SG, Angelosante AG, and Kendall PC, “Discriminant validity and clinical utility of the CBCL with anxiety-disordered youth,” J. Clin. Child Adolesc. Psychol, vol. 34, pp. 735–746, 2005. [DOI] [PubMed] [Google Scholar]
  • [46].de la Osa N, Granero R, Trepat E, Domenech JM, and Ezpeleta L, “The discriminative capacity of CBCL/1 1/2–5-DSM5 scales to identify disruptive and internalizing disorders in preschool children,” Eur. Child Adolesc. Psychiatry, vol. 25, pp. 17–23, 2016, doi: 10.1007/s00787-015-0694-4. [DOI] [PubMed] [Google Scholar]
  • [47].Pine DS et al. , “Attention bias to threat in maltreated children: Implications for vulnerability to stress-related psychopathology,” Amer. J. Psychiatry, vol. 162, pp. 291–296, 2005, doi: 10.1176/appi.ajp.162.2.291. [DOI] [PubMed] [Google Scholar]
  • [48].Meyer BM et al. , “How much data is enough? A reliable methodology to examine long-term wearable data acquisition in gait and postural sway,” Sensors, vol. 22, no. 18, Jan. 2022, Art. no. 6982, doi: 10.3390/s22186982. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [49].Wang S et al. , “Depression and anxiety among children and adolescents pre and post COVID-19: A comparative meta-analysis,” Front. Psychiatry, vol. 13, Aug. 2022, Art. no. 917552, doi: 10.3389/fpsyt.2022.917552. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [50].McGinnis RS, McGinnis EW, Petrillo C, Ferri J, Scism J, and Price M, “Validation of smartphone based heart rate tracking for remote treatment of panic attacks,” IEEE J. Biomed. Health Inform, vol. 25, no. 3, pp. 656–662, Mar. 2021, doi: 10.1109/JBHI.2020.3001573. [DOI] [PubMed] [Google Scholar]
  • [51].McGinnis E, O’Leary A, Gurchiek R, Copeland W, and McGinnis R, “PanicMechanicTM: A digital therapeutic delivering biofeedback for panic attacks (preprint),” J. Med. Internet Res. Formative Res, vol. 6, 2021, Art. no. e32982, doi: 10.2196/32982. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [52].Ben-Zeev D, Kaiser SM, Brenner CJ, Begale M, Duffecy J, and Mohr DC, “Development and usability testing of FOCUS: A smartphone system for self-management of schizophrenia,” Psychiatr. Rehabil. J, vol. 36, no. 4, pp. 289–296, 2013, doi: 10.1037/prj0000019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [53].McGinnis RS and McGinnis EW, “Advancing digital medicine with wearables in the wild,” Sensors, vol. 22, no. 12, 2022, Art. no. 4576, doi: 10.3390/s22124576. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [54].Mohanta A, Mukherjee P, and Mirtal VK, “Acoustic features characterization of autism speech for automated detection and classification,” in Proc. Nat. Conf. Commun, 2020, pp. 1–6, doi: 10.1109/NCC48643.2020.9056025. [DOI] [Google Scholar]
  • [55].Siddiqui UA et al. , “Wearable-sensors-based platform for gesture recognition of autism spectrum disorder children using machine learning algorithms,” Sensors, vol. 21, no. 10, 2021, Art. no. 3319, doi: 10.3390/s21103319. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [56].Jiang X et al. , “WeDA: Designing and evaluating a scale-driven wearable diagnostic assessment system for children with ADHD,” in Proc. CHI Conf. Hum. Factors Comput. Syst, 2020, pp. 1–12, doi: 10.1145/3313831.3376374. [DOI] [Google Scholar]
  • [57].Luo J et al. , “A wearable diagnostic assessment system vs. SNAP-IV for the auxiliary diagnosis of ADHD: A diagnostic test,” BioMed Central Psychiatry, vol. 22, no. 1, 2022, Art. no. 415, doi: 10.1186/s12888-022-04038-3. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES