Abstract
Each year, a significant portion of the 40 million individuals in the United States who seek care in emergency departments (EDs) following traumatic experiences develop adverse posttraumatic neuropsychiatric sequelae (APNS). This highlights the widespread impact of trauma and the critical need for effective interventions to address the health outcomes of these events. Despite significant research efforts, advancements in understanding the neurobiological mechanisms of APNS have been hindered, primarily due to reliance on subjective self-reports, which are susceptible to recall biases and careless responses. To overcome this limitation, we investigate the use of objective, longitudinal mobile device data to identify consistent APNS states and examine their dynamic transitions over time. To address the complexity of longitudinal mobile data, we developed a continuous-time hidden Markov factor model and applied it to mobile device data from the Advancing Understanding of Recovery after Trauma (AURORA) study. Findings from this study highlight the great potential of leveraging mobile device data for continuous health monitoring and for guiding personalized treatment approaches through mobile health initiatives. A Python implementation of the proposed method is available at https://anonymous.4open.science/r/CTHMFM.
Index Terms—: Continuous-time hidden Markov model, Mental health, Multivariate longitudinal data
I. Introduction
Adverse posttraumatic neuropsychiatric sequelae (APNS) (e.g., pain, depression, and PTSD) are frequently observed in civilians and military veterans who have experienced traumatic events. These APNS increase the risk of chronic illnesses, including cancer and heart disease, and substantially contribute to drug abuse, suicide, and disability. Moreover, APNS impose enduring psychosocial and financial burdens not only on individuals with the disorder but also on their families, communities, and society as a whole. However, little progress has been made in advancing APNS research over the past few decades due to unique challenges. First, APNS have been evaluated through subjective self-reported measures, which lack objective reliability. Second, the heterogeneity among patients, as recognized in traditional classification and diagnoses, complicates the study of APNS. Lastly, these APNS disorders are often studied and treated independently, despite their frequent co-occurrence [1]. These obstacles hinder the identification of objective markers, the advancement in understanding the neurobiological mechanisms of APNS, and the development of effective preventative/treatment strategies.
Identifying homogeneous states and exploring the dynamic prognosis of APNS in the immediate aftermath of trauma exposure holds promise for enhancing our understanding of APNS and identifying effective intervention options and appropriate timing at the individual level. Regrettably, due to the lack of appropriate data and effective statistical method, no large-scale studies have been conducted to investigate the onset, dynamic transitions (such as recovery and relapse), and associated risk factors of APNS. To help address the challenges, the National Institutes of Mental Health, joined by the US Army Medical Research and Material Command, developed the Advancing Understanding of RecOvery afteR traumA (AURORA) study [1]. This study gathered biobehavioral data from a large cohort of trauma survivors (n = 2,997) across the United States over a year, including self-reported surveys, web-based neurocognitive tests, digital phenotyping data (i.e., from wrist wearables and smartphones), psychophysical tests, neuroimaging assessments, and genomics data. In contrast to previous studies relying on self-report surveys or neuroimages [2], our work aims to address the challenges of APNS by utilizing the objective digital phenotyping data that tracks individuals’ behaviors, moods, and health statuses in real-time, real-life environments. Specifically, we developed a Hidden Markov Factor Model (HMFM) to analyze the mobile device data, allowing us to simultaneously identify homogeneous subtypes, investigate subtype-specific structures, and model individual progression with associated risk factors.
Hidden Markov Models (HMMs) have been widely used in various fields [3]. However, mobile device data presents two unique challenges that standard HMMs cannot handle, including the interdependent variables with unknown interrelationship structures and unevenly spaced measurements.
Mobile device sensor data, such as accelerometer and photoplethysmography (PPG) from smartwatches, involves highly intensive time series data. Instead of being directly used, these data are typically pre-processed to generate technical summaries that represent various characteristics of each time series variable, which are often highly correlated. As the number of features increases, the parameters in the covariance matrix expand exponentially, making a fully free covariance matrix impractical. Consequently, when implementing HMMs, features are usually assumed to be independent given the latent state membership, despite their correlation, potentially introducing estimation bias.
To appropriately model the association between features, factor analysis models (FMs) [4] provide an efficient and parsimonious approach and have been incorporated into HMMs in various ways. For example, the factor analyzed hidden Markov model [5] combines an FM with a discrete-time HMM (DTHMM), which assumes evenly spaced measurements. It has been extensively used in a variety of real-world applications, including speech recognition [6], environmental protection [7], and seizure detection [8]. Similarly, [9] introduced the regime-switching factor model to handle high-dimensional financial market data. However, they all assume homogeneous transition probability matrices, limiting their ability to account for the heterogeneity of transition probabilities over time and among different subjects and explore risk factors of state transition. To simultaneously capture the interrelationships among observed features and account for the variability of transition probabilities, a joint framework incorporating HMM, FM, and a feature-based transition model was recently proposed [10], [11]. However, it is not directly applicable to mobile device data. First, it employs a confirmatory factor model (CFM) with pre-specified structures for the factor loading matrices, which are often unknown a priori. Therefore, an exploratory factor model (EFM) is needed to explore the interrelationships among all observed features. Second, their framework assumes ordered states, which is inappropriate for our use case.
Another challenge posed by mobile device data is the irregular spacing of measurements. For example, activity and heart rate variability (HRV) data were collected only when the participants wore the watches, resulting in non-uniformly spaced observations and significant variation in sampling schedules between individuals. While the aforementioned methods are all based on DTHMM, neglecting the impact of time gaps between consecutive observations on transition rates, continuous-time discrete-state HMM (CTHMM) was developed to handle irregularly spaced measurements [12]. CTHMM and its extensions that incorporate covariates to characterize transition rates are widely used in medical research [13], [14]. However, none of them address the interrelationships among features.
In this paper, to simultaneously address the two challenges and examine heterogeneous transition patterns, we propose to use Continuous-Time HMFM (CT-HMFM), integrating CTHMM, EFM, and a feature-based transition model. Our contributions are as follows: First, we examine the utility of data collected in an open environment from consumer-grade mobile devices for mental health research. This contrasts with most existing studies on data collected in controlled lab environments. Second, we propose CT-HMFM to address the unique challenges introduced by mobile device data and depict the non-homogeneous state transition processes of multiple individuals. Simulation studies using synthetic data demonstrate exceptional parameter estimation and model selection performance. Finally, we analyze HRV and activity data from the AURORA study, followed by interpretations and discussions of biological findings that highlight the immense potential of mobile health data and our proposed method for mental health research.
II. AURORA Dataset
In the AURORA study [1], data were collected from multiple sources, with our analysis focusing on the accelerometer and PPG data collected by the Verily smartwatches. Given the known associations with APNS [15], [16], the accelerometer data were used to quantify physical activity features, while the PPG data were pre-processed to extract heart rate variability (HRV) metrics (for details on the raw data and pre-processing steps, refer to [17]).
Briefly, activity features are extracted during a 24-hour window to evaluate daily activity patterns. After converting accelerometer data to activity counts [18], meanAcc is the average activity counts. Amplitude, a cosinor rhythmometry feature, is computed to capture circadian rhythms [19]. Using the Cole-Kripke algorithm [20], accelerometry epochs are classified into wake or sleep states, and the SWCK quantifies transition rates between wake and sleep. Additionally, the average activity during the five least active hours (L5) is calculated from raw accelerometer data [21], representing nighttime activity.
HRV features were derived from PPG data by first calculating and denoising beat-to-beat (BB) interval [22] time series to obtain normal-to-normal (NN) intervals, which were then analyzed using a 5-minute sliding window. Selected HRV features for this study include average heart rate (NNmean), skewness (NNskew), and standard deviation (SDNN) of NN intervals, indicating heart rate variability and rapid changes. Lfhf is the low-frequency to high-frequency power ratio, with extremes suggesting parasympathetic or sympathetic dominance [22]. DC serves as a mortality risk indicator in cardiac conditions [23]. The lower the DC index, the higher the mortality risk. Additionally, SD1SD2 and ApEn are used to quantify the unpredictability and regularity of successive heartbeats (R-R interval). To match the activity data, daily statistical summaries of each HRV feature are used, including mean, minimum, maximum, interquartile range, and variance.
In this study, we focused on survivors of motor vehicle collision trauma. Since the data collection depends on the participants’ wearing of the devices, missing data is a common issue. To assure data quality, we include only records that have complete activity data and a positive wake percentage. For HRV data, ideally, an individual can have 2,880 records per day. We retain only those days where at least 30% of these records are available to ensure that our daily summary statistics are representative. The final dataset consists of daily summaries of activity and HRV features from 258 patients, with each providing at least 50 days of records.
III. CT-HMFM
Motivated by the structures of the processed AURORA datasets, we focus on data consisting of repeated measurements of features collected over time points for each individual among subjects. Let denote the observed values of the features for subject at time , and let denote the matrix containing all measurements for individual . To analyze these longitudinal data, we adopt a HMM framework, which facilitates the inference of latent states and their dynamic transitions over time. Specifically, we assume that the latent state of individual at time , denoted by , takes values from a finite discrete set . At , the initial state is assumed to follow a multinomial distribution with probabilities , such that . Transition probabilities between theses states at time point are captured by a transition probability matrix , of which the -th entry is , and . The primary goal is to estimate the and , thereby delineating latent Markov processes underlying the observed data. To meet our specific objectives of analyzing the mobile health data, we extend the traditional HMM framework with two additional components. The first, detailed in Section III-A, is a state-specific measurement model using EFM to uncover interrelationships among variables. The second, outlined in Section III-B, is a transition model (TM) designed to capture heterogeneous transition patterns.
A. State-Specific Measurement Model
The first component is an FM, aimed at identifying the interrelationship structures between observed response variables and the underlying latent variables. For individual at time , given that , the FM assumes that:
| (1) |
where is a vector of state-specific expected mean response, is a state-specific factor loading matrix, is a -dimensional vector of latent scores assumed to be independent of and following a standard multivariate normal distribution, and is a diagonal covariance matrix for the error term with positive nonconstant diagonal entries. Alternatively, (1) can also be expressed as
| (2) |
It is crucial to emphasize that, unlike CFMs that require pre-specified structures of the factor loading matrices, our approach imposes no assumptions on . Consequently, the structure of is entirely data-driven, making the first component (1) an EFM.
B. Transition Model
To account for the effects of varying time intervals between observations, a continuous-time Markov process is used. Instead of directly relying on the transition probability matrix , the continuous-time Markov process is defined by a transition intensity matrix [24], which is the limit of as the time interval approaches zero. Specifically, let denote the number of pre-specified time units between and observation, the transition intensity from state to state for subject at time is given by:
| (3) |
and . The corresponding transition probability can then be calculated as the matrix exponential of . Since computing the matrix exponential directly can be computationally intensive, in practice, we approximate the using for a sufficiently large [25].
To investigate the impact of covariates on transition rates, the is modeled with a log-linear model [26], such that
| (4) |
where is a vector of covariates, and is a vector of state-specific fixed effects coefficients, quantifying the effect of covariates on the transition intensity from state to state .
IV. Expectation-Maximization Algorithm
Let , and . Under the standard assumptions of HMMs–(1) given the current state , the observations are conditionally independent, and (2) given and the subject’s contextual features, the subsequent state is independent of any historical information–the joint probability distribution of the observations and all latent variables, , is constructed as follows:
| (5) |
Assuming the independence of , and across individuals, the complete likelihood function for the entire sample is obtained as the product of (5) over . To estimate , the expectation-maximization (EM) algorithm is used, as both and are unobserved. Specifically, the EM algorithm iteratively applies the expectation and maximization steps to find a local maximum of the marginal likelihood, as detailed below.
A. Expectation Step (E-step)
The E-step computes the expected value of the logarithm of , given the observations and the current parameter estimates . Directly computing this expectation, , is challenging due to the computational intensity in calculating the involved conditional state probabilities, including
| (6) |
and
| (7) |
To efficiently compute these probabilities, we utilize a scaled version of the forward-backward algorithm (FBA) [27].
Specifically, we first calculate the forward probability . Let be the probability density of given , and the conditional probability of observation given all past observations. Omitting the dependence on for brevity, the forward probabilities are computed recursively as:
| (8) |
| (9) |
where . Similarly, the backward probability , defined as , is also obtained recursively as:
| (10) |
| (11) |
Using the forward and backward probabilities, the conditional state probabilities are derived as:
| (12) |
| (13) |
With these, can be expressed explicitly. While detailed expressions are provided in Appendix I-A, the resulted expression can be summarized as:
| (14) |
where depends on the initial state distribution, depends on the probability transition matrix, and is a function of parameters , and .
B. Maximization (M-step)
Within each M-step, since , and do not share parameters, we maximize each of them separately. The estimator of , and are directly derived by setting and (see Appendix I-B for details). For , a one-step Fisher scoring (FS) [28] algorithm is implemented to optimize the parameters.
Let be the ordered vector of all transition model parameters, such that , with representing the entry of . Recalling the first derivative of the matrix exponential [14] and using Theorem 1 in [29], the derivation of with respect to is given by:
| (15) |
where . Let denote the entry of the first derivative of with respect to . Having the first derivative of with respect to each component of calculated accordingly, the FS can then be directly implemented to update without requiring the second derivative of the matrix exponential. Specifically, let be the score function, defined as
| (16) |
Let denote the negative Fisher information matrix, whose entry is given by:
| (17) |
With the score function and the Fisher information matrix, the parameters can be updated as
| (18) |
To ensure numerical stability, we control the learning rate in practice by updating
| (19) |
Note that the algorithm requires specification of , which are typically unknown in practice. In this study, we use information criteria to determine , the efficacy of which is evaluated in Section V-C.
V. Simulation Study
This section assesses the proposed methods through simulation studies using synthetic data resembling the AURORA dataset. We generate data with , and . Each individual’s number of observations, , is uniformly sampled from [50, 100]. From this, we randomly select time points from to create sequences of . Each individual’s initial state is drawn from a multinomial distribution with probabilities . Latent state trajectories are then generated based on transition probabilities , given individuals’ features. Observation vectors for an individual in state at time are then sampled from a normal distribution with mean and covariance , where .
In the following, subsection V-A assesses model reliability by comparing empirical parameter estimates against true values; subsection V-B compares the effectiveness of our method against baseline methods; and subsection V-C explores the performance of information criteria in model selection.
A. Simulation 1
To validate the estimation procedure, we implement the EM algorithm with true and . Parameter initialization involves first fitting Gaussian Mixture Models to estimate groups, followed by EFM for each group. Guided by the insights from a pilot study, we set the maximum number of iterations for each replication at 100. The reliability and precision of the proposed methods are then evaluated from two perspectives: i) the accuracy of each individual parameter estimate and ii) the misclassification rate , which quantifies the proportion of estimated states that diverge from the actual states.
The accuracy of parameters , and is assessed by calculating the average absolute difference (AAD) between the estimates and their true values, defined as is an individual entry in matrix and is the total number of free parameters. The mean of AADs (standard errors in the parentheses) aggregated over 100 random seeds are presented in Table I. These mean AADs for all parameter matrices are sufficiently close to zero with small standard errors, indicating effective parameter recovery. In Table II, we report the mean bias (standard error) for each parameter in the transition model, where the biases are all close to zero. Moreover, we present the mean (standard error) of in Table I. On average, misclassification rates are only .24% (0.0005), highlighting the exceptional accuracy of the proposed EM algorithm in estimating latent states.
TABLE I.
The Mean (standard error) AADs of , and .
| .026 (.013) | .015 (.002) | .014 (.001) | .011 (.002) | .0024 (.0005) |
TABLE II.
Bias (standard error) of the parameter estimates for Each transition model parameter .
| .014(.142) | .017(.150) | −.080(.219) | |
| −.021(.126) | .010(.125) | .016(.182) | |
| −.018(.147) | −.004(.139) | .014(.214) | |
| −.003(.112) | .004(.097) | .005(.200) | |
| .003(.112) | .002(.107) | −.006(.192) | |
| .006(.131) | −.032(.091) | .028(.232) |
Intuitively, factors such as N, , sizes of and , variance , differences in and between states, and frequency of state transitions, can affect the performance of parameter estimation. Additional simulations in Appendix II reveals that estimation accuracy for , and , as well as classification accuracy, improve when (i) common variances decrease, (ii) differences in and between states increase, (iii) decreases, or (iv) sample size or panel length increases. Increasing or using a that induces infrequent transitions slightly affects most parameters but enhances the precision of transition probability estimates, thereby reducing misclassification rates. Estimation of improves solely with increases in N or state-to-state differences in .
B. Simulation 2
This section compares the performance of the proposed methods and the baseline approaches in correctly identifying latent states. Three benchmark methods are under our consideration: i) TM+independent HMM, which assumes independence among observed features given the states; ii) CFM+TM+HMM, which addresses interrelationships but inaccurately pre-specifies the latent factor structure by setting certain loading matrix entries to zero; and iii) EFM+HMM, which assumes a homogeneous transition probabilities matrix for all subjects. We first repeat the data generation process of Simulation 1. Then, we consider three additional scenarios by adjusting the state-to-state differences in to be closer (: medium diff), increasing the similarity of the at different states (: medium diff), and increasing the significance of the covariance matrix , respectively.
As depicted in Figure 1, our proposed methods (HMFM) consistently outperform the benchmark methodologies in all settings. Regardless of sample size, our methods consistently achieve the lowest misclassification rate, nearly approximating zero, thereby emphasizing the importance of each component in our proposed models. Specifically, the comparison with TM+independent HMM shows the importance of accounting for the interrelationship between observed features; the comparison with CFM+TM+HMM reveals the risk of incorrectly specifying the interrelationship structure; and the comparison with EFM+HMM demonstrates the inadequacy of assuming homogeneous transition probabilities.
Fig. 1.

of various methods. The error bars represent the 95% CI. The first column shows the results under the settings we used in simulation 1. The last three columns summarize the results under different settings by varying the true value of , and , respectively.
C. Simulation 3
Information criteria such as the Akaike information criteria (AIC) and the Bayesian information criteria (BIC) have been widely used in model selection [10], [30]. Within this simulation study, we investigate whether the AIC or BIC is reliable for determining J and K simultaneously. We repeat the data generation process of Simulation 1, but implement the proposed methods with a different set of for each replicate when fitting the generated data. Let and . We consider all possible combinations of and , yielding a total of nine fitted candidate models for each replicate. For 100 replications, both AIC and BIC consistently recommend the model with accurate and . Therefore, we believe that the sample size and the number of observations per individual in the processed AURORA data will yield reliable information criteria-based model selection results and, consequently, reliable parameter estimation.
VI. Analysis of the AURORA Data
Given the irregular data collection from mobile devices, we applied the CT-HMFM to smartwatch data from the AURORA study. Considering and , we evaluate 54 candidate models. For each candidate model, the EM algorithm is implemented with multiple random seeds, and the seed yielding the highest estimated likelihood is selected. Finally, model comparison using AIC and BIC led to the selection of a model with and . Subsequent subsections will discuss the interpretation of parameter estimates and biological insights from three perspectives: i) the interpretation of the three estimated states, ii) symptom co-occurrence patterns, and iii) the influence of demographic factors on transition probabilities.
A. Interpretation of Hidden States
To investigate the biological differences between states, we first focus on the selected features. Figure 2 depicts scaled sample means for each feature across states, with a 99% confidence interval (CI). Further pairwise Tukey tests indicate significant differences between states for nearly all features, except for amplitude, SWCK, L5, and NNskew.q3 between state 1 and state 2. Specifically, features related to average heart rate (NNmean), heart rate variability (SDNN), and heart deceleration capacity (dc) vary significantly across the three latent states, decreasing sequentially from state 1 to state 3. According to previous research, lower heart rate variability and deceleration capacity are associated with a higher mortality rate [23], [31], suggesting that states 1 through 3 represent decreasing levels of health, with state 1 being the healthiest. Additionally, activity levels are similar and higher in states 1 and 2 compared to state 3, indicating better overall health for participants in the first two states. Among features related to heart rate unpredictability (lfhf, ApEn, and SD1SD2), state 3 demonstrates significantly higher values for SD1SD2-related features but lower values for lfhf-related features compared to states 1 and 2, suggesting a different interpretation of the estimated states than our previous interpretation. However, it is important to note that the relationship between these features and the psychological or physiological state is neither straightforward nor unique [32].
Fig. 2.

Relative sample mean for features in each estimated state, with 99% CI error bars that are too small to distinguish.
To confirm the validity of the three states, we further compare their differences regarding self-reported symptoms from a flash survey. Based on the RDoC framework, ten latent constructs associated with APNS were developed using survey items selected by domain experts: Pain, Loss, Sleep Discontinuity, Nightmare, Somatic Symptoms, Mental Fatigue, Avoidance, Re-experience, and Anxious. Retaining only observations for each individual whose estimated states are known on the same day they submitted survey responses, we summarized the flash survey data with means and 95% CIs in Figure 3. Overall, state 1 exhibits the lowest severity level for all ten symptoms, while state 3 has the highest severity level. Tukey tests reveal no significant differences between states 1 and 2 in hyperarousal, re-experience, anxiety, and somatic symptoms, but both are significantly different from state 3 in these constructs. While the differences in nightmare and sleep discontinuity between states 3 and 2 are not significant, they are statistically more severe than in state 1. For mental fatigue and depression, only the difference between state 1 and state 3 is statistically significant.
Fig. 3.

Sample mean for each symptom in each estimated state. The error bars represent the 95% CI. is the least severity, is the greatest.
In summary, both the flash survey data and the AURORA data (HRV, Activity) support our interpretation of the three latent states. State 1 is the healthiest, while state 3 indicates having the most severe APNS symptoms.
B. Co-occurring Pattern of Symptoms
When studying the co-occurring of symptoms within each hidden state, we limit our attention to observations collected during the first week. For each state, the correlations between all ten symptoms are calculated. In state 1 (relative health state), there is a high degree of correlation (.782) between hyperarousal and anxiety, suggesting that patients in state 1 experiencing severe hyperarousal symptoms are also likely to suffer from severe anxiety. State 2 shows no highly correlated symptoms. In state 3 (the state with more severe disorders), symptoms such as depression, hyperarousal, anxiety, and re-experience are more likely to co-occur with pairwise correlation ranging from .716 to .91.
C. Transition Probability
This section investigates the heterogeneity of 1-day transition probabilities among subjects, focusing on transitions with a time interval . We estimated the transition probabilities for males and females within the sample age range, as depicted in Figure 4. Lines with circles illustrate the probability of remaining in the same state, lines with stars indicate transitions to a more severe state, and lines with ‘x’ reflect the chance of improvement in psychological conditions.
Fig. 4.

Estimated transition probability. Fix . (a, b) indicates a transition from state a to state b.
Overall, both males and females have a tendency to remain in their current state, with infrequent state transitions, aligning with most literature. For males, the probability of staying at states 3 and 2 increases with age, while it decreases for state 1. Moreover, while the likelihood of psychological deterioration increases with age, the chance of improvement decreases. Specifically, while the probability of transitioning from the most severe state (state 3) to the healthiest state (state 1) approaches zero as age increases, the likelihood of the reverse transition increases as age decreases, with direct transitions between state 1 and state 3 being particularly rare. The female group exhibits a similar trend to the male group, but with a higher likelihood of remaining in the most severe state (state 3) compared to males.
In summary, our analysis of the AURORA data suggests that older patients are more likely to transition to more severe psychological states. Moreover, achieving psychological improvement becomes increasingly challenging as one ages.
D. State Specific Associations Between Activity and HRV Features
The model provides a deeper understanding of the latent states, not only in terms of their means but also through the associations between activity and HRV features. These associations offer new insights into the underlying physiology of APNS symptoms. The state-specific associations are captured in the structure of the factor loading matrix, which reveals both shared patterns and distinct differences across states, reflecting heterogeneous relationships. For example, the analysis shows a consistent positive correlation between greater daytime activity and better sleep quality across all states. Additionally, heart rate variability and irregularity (measured by SDNN and ApEn) are correlated across all states. However, their association with heart deceleration capacity (dc-related features) is notably stronger in states 2 and 3, highlighting state-specific physiological patterns.
VII. Conclusions and Discussion
This paper explores the unique challenges and potential of using intensive mobile device data to provide objective measures and study the dynamic transitions of APNS symptoms after trauma exposure. The study addresses challenges such as interdependent variables with unknown relationships, heterogeneous transition probabilities, and irregular measurement intervals. The findings suggest that mobile device data from consumer-grade devices, combined with the proposed analytical methods, hold immense promise for assessing, monitoring, and understanding the dynamic progression of APNS symptoms post-trauma.
Traditionally, APNS assessments rely heavily on self-reported measures, which are inherently subjective and prone to recall biases or careless responses. By contrast, mobile device data collected through consumer-grade wearables like smartwatches and fitness trackers offer an objective and continuous perspective on individuals’ physical and mental health. While previous studies have largely relied on data collected in controlled environments, such as sleep labs or clinical settings, data obtained in natural, real-world environments introduces additional noise and presents unique analytical challenges. Extracting meaningful health-related insights from these data remains an open question.
This study demonstrates the immense potential of mobile device data as a secondary objective measure for a broad spectrum of APNS symptoms. These real-time data not only allow clinicians to monitor patients continuously in their natural environments but also provide more timely insights compared to periodic self-reports or clinic visits. By mapping how individuals transition between health states over time and identifying high-risk groups and recovery factors, clinicians gain a better understanding of the natural progression of trauma recovery. This knowledge enables the design of personalized, preventive interventions tailored to individual needs.
Furthermore, the proposed model is adaptable to data from other mobile devices, such as smartphones. Given the widespread ownership and daily use of mobile devices, this approach offers a scalable solution for continuous health monitoring. It also holds significant potential for personalized treatment through mobile health initiatives, particularly in under-resourced areas where access to traditional clinic-based care is limited.
Acknowledgments
The research is partially supported by NSF under Grant DMS-1555244 and 2113637, NIMH under Grant U01MH110925, the US Army Medical Research and Material Command, The One Mind Foundation, and The Mayday Fund.
Appendix I
Technical Details
A. Supplement for E-step
Denote and . Each of the three parts has an explicit form:
| (20) |
| (21) |
| (22) |
B. Supplement for M-step
By setting the first derivative of to zero, the parameters related to the initial state distribution are estimated as:
| (23) |
Similarly, the parameters used to characterize the conditional distribution of given are estimated by setting the first derivative of equal to 0, with
| (24) |
Meanwhile, we got the updated estimation of , equals
| (25) |
Appendix II
Additional Simulation Results
This section presents additional simulation results that investigate the impact of various factors on estimation performance. Using the settings for simulation 1 in Section V-A as the baseline, we conducted eight additional sets of simulations, each varying one component while maintaining rest components as the baseline setup. These components include: i) sample size (N), ii) number of measurements for each individual , iii) J, iv) K, v) size of common variance , vi) state-to-state difference in , vii) state-to-state difference in , and viii) transition frequency. For common variance , we evaluated scenarios with , , and (baseline). For and , we adjusted state-to-state differences in two additional settings. For the test evaluating the effect of transition frequency, a frequent transition is defined as the probability of remaining in the same state being less than 0.70. The in the baseline setting corresponds to infrequent transition. For tests evaluating the effects of J, K, N, and , the baseline setups are modified as indicated in Table III.
TABLE III.
The Mean (standard error) AADs of , and , and of estimations under different CT settings.
| ADD | ||||||
|---|---|---|---|---|---|---|
| .026(.013) | .015(.002) | .014(.001) | .011(.002) | .120(.027) | .0024(.0005) | |
| .026(.013) | .012(.002) | .011(.001) | .005(.001) | .119(.027) | .0003(.0001) | |
| .026(.013) | .010(.003) | .007(.001) | .001(.000) | .119(.027) | .0000(.0000) | |
| : large diff | .026(.013) | .015(.002) | .014(.001) | .011(.002) | .120(.027) | .002(.0005) |
| : medium diff | .027(.013) | .015(.001) | .015(.001) | .011(.002) | .124(.028) | .007(.001) |
| : minor diff | .034(.017) | .017(.003) | .478(.057) | .011(.002) | .161(.040) | .085(.004) |
| : large diff | .026(.013) | .015(.002) | .014(.001) | .011(.002) | .120(.027) | .002(.0005) |
| : medium diff | .026(.013) | .014(.002) | .014(.001) | .011(.002) | .122(.029) | .006(.001) |
| : minor diff | .026(.013) | .014(.002) | .014(.001) | .011(.002) | .122(.028) | .004(.001) |
| : infreq transit | .026(.013) | .015(.002) | .014(.001) | .011(.002) | .120(.027) | .002(.0005) |
| : freq transit | .027(.013) | .015(.002) | .014(.001) | .011(.002) | .137(.043) | .007(.001) |
| .027(.021) | .012(.002) | .012(.001) | .011(.002) | .087(.026) | .0015(.0003) | |
| .026(.013) | .015(.002) | .014(.001) | .011(.002) | .120(.027) | .0024(.0005) | |
| .024(.011) | .017(.002) | .016(.001) | .011(.001) | .150(.028) | .0032(.0005) | |
| .026(.013) | .015(.002) | .014(.001) | .010(.002) | .124(.027) | .0069(.0008) | |
| .026(.013) | .015(.002) | .014(.001) | .011(.002) | .120(.027) | .0024(.0005) | |
| .026(.013) | .015(.002) | .015(.001) | .012(.002) | .119(.027) | .0005(.0002) | |
| .052(.029) | .030(.004) | .028(.002) | .022(.003) | .254(.058) | .0027(.0008) | |
| .042(.019) | .021(.003) | .020(.001) | .015(.002) | .175(.038) | .0025(.0006) | |
| .018(.010) | .009(.001) | .009(.001) | .007(.001) | .076(.017) | .0024(.0003) | |
| .029(.014) | .029(.004) | .027(.002) | .021(.004) | .251(.061) | .0028(.0009) | |
| .027(.014) | .020(.003) | .019(.001) | .015(.003) | .180(.043) | .0025(.0005) | |
| .028(.015) | .012(.001) | .011(.001) | .008(.001) | .098(.021) | .0023(.0003) |
Contributor Information
Lin Ge, Department of Statistics, North Carolina State University, Raleigh, NC 27695 USA..
Xinming An, Department of Anesthesiology, The University of North Carolina at Chapel Hill, Chapel Hill, NC 27514 USA..
Donglin Zeng, Department of Biostatistics, The University of North Carolina at Chapel Hill, Chapel Hill, NC 27599 USA..
Samuel McLean, Department of Psychiatry, The University of North Carolina at Chapel Hill, Chapel Hill, NC 27514 USA..
Ronald Kessler, Department of Health Care Policy, Harvard Medical School, Boston, MA 02115 USA..
Rui Song, Department of Statistics, North Carolina State University, Raleigh, NC 27695 USA..
References
- [1].McLean SA, Ressler K, Koenen KC, Neylan T, Germine L, Jovanovic T, Clifford GD, Zeng D, An X, Linnstaedt S et al. , “The aurora study: a longitudinal, multimodal library of brain biology and function after traumatic stress exposure,” Molecular psychiatry, vol. 25, no. 2, pp. 283–296, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [2].Marquand AF, Wolfers T, Mennes M, Buitelaar J, and Beckmann CF, “Beyond lumping and splitting: a review of computational approaches for stratifying psychiatric disorders,” Biological psychiatry: cognitive neuroscience and neuroimaging, vol. 1, no. 5, pp. 433–447, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [3].Mor B, Garhwal S, and Kumar A, “A systematic review of hidden markov models and their applications,” Archives of computational methods in engineering, vol. 28, no. 3, pp. 1429–1448, 2021. [Google Scholar]
- [4].Kim J-O and Mueller CW, Factor analysis: Statistical methods and practical issues. sage, 1978, vol. 14. [Google Scholar]
- [5].Rosti AI and Gales MJ, “Factor analysed hidden markov models,” in 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 1. IEEE, 2002, pp. I–949. [Google Scholar]
- [6].Rosti AI and Gales M, “Factor analysed hidden markov models for speech recognition,” Computer Speech & Language, vol. 18, no. 2, pp. 181–200, 2004. [Google Scholar]
- [7].Maruotti A, Bulla J, Lagona F, Picone M, and Martella F, “Dynamic mixtures of factor analyzers to characterize multivariate air pollutant exposures,” The Annals of Applied Statistics, vol. 11, no. 3, pp. 1617–1648, 2017. [Google Scholar]
- [8].Madadi M, “A hidden markov factor analysis framework for seizure detection in epilepsy patients,” 2019. [Google Scholar]
- [9].Liu X and Chen R, “Regime-Switching Factor Models for High-Dimensional Time Series,” Statistica Sinica, vol. 26, no. 4, pp. 1427–1451, 2016. [Online]. Available: https://www.jstor.org/stable/44114341 [Google Scholar]
- [10].Song X, Xia Y, and Zhu H, “Hidden Markov latent variable models with multivariate longitudinal data.” Biometrics, vol. 73, no. 1, pp. 313–323, 2017. [Online]. Available: 10.1111/biom.12536 [DOI] [PMC free article] [PubMed] [Google Scholar]
- [11].Zhou X, Kang K, Kwok T, and Song X, “Joint hidden markov model for longitudinal and time-to-event data with latent variables,” Multivariate Behavioral Research, vol. 57, no. 2–3, pp. 441–457, 2022. [DOI] [PubMed] [Google Scholar]
- [12].Cox DR and Miller HD, The theory of stochastic processes. Routledge, 2017. [Google Scholar]
- [13].Amoros R, King R, Toyoda H, Kumada T, Johnson PJ, and Bird TG, “A continuous-time hidden markov model for cancer surveillance using serum biomarkers with application to hepatocellular carcinoma,” Metron, vol. 77, no. 2, pp. 67–86, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [14].Zhou J, Song X, and Sun L, “Continuous time hidden markov model for longitudinal data,” Journal of Multivariate Analysis, vol. 179, p. 104646, 2020. [Google Scholar]
- [15].Hartmann R, Schmidt FM, Sander C, and Hegerl U, “Heart rate variability as indicator of clinical state in depression,” Frontiers in psychiatry, vol. 9, p. 735, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [16].Jung W, Jang K-I, and Lee S-H, “Heart and brain interaction of psychiatric illness: a review focused on heart rate variability, cognitive function, and quantitative electroencephalography,” Clinical Psychopharmacology and Neuroscience, vol. 17, no. 4, p. 459, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [17].Cakmak AS, Alday EAP, Da Poian G, Rad AB, Metzler TJ, Neylan TC, House SL, Beaudoin FL, An X, Stevens JS et al. , “Classification and prediction of post-trauma outcomes related to ptsd using circadian rhythm changes measured via wrist-worn research watch in a large longitudinal cohort,” IEEE journal of biomedical and health informatics, vol. 25, no. 8, pp. 2866–2876, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [18].Borazio M, Berlin E, Kücükyildiz N, Scholl P, and Van Laerhoven K, “Towards benchmarked sleep detection with wrist-worn sensing units,” in 2014 IEEE International Conference on Healthcare Informatics. IEEE, 2014, pp. 125–134. [Google Scholar]
- [19].Cornelissen G, “Cosinor-based rhythmometry,” Theoretical Biology and Medical Modelling, vol. 11, no. 1, pp. 1–24, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [20].Cole RJ, Kripke DF, Gruen W, Mullaney DJ, and Gillin JC, “Automatic sleep/wake identification from wrist activity,” Sleep, vol. 15, no. 5, pp. 461–469, 1992. [DOI] [PubMed] [Google Scholar]
- [21].Van Someren EJ, Swaab DF, Colenda CC, Cohen W, McCall WV, and Rosenquist PB, “Bright light therapy: improved sensitivity to its effects on rest-activity rhythms in alzheimer patients by application of nonparametric methods,” Chronobiology international, vol. 16, no. 4, pp. 505–518, 1999. [DOI] [PubMed] [Google Scholar]
- [22].Shaffer F and Ginsberg JP, “An overview of heart rate variability metrics and norms,” Frontiers in public health, vol. 5, p. 258, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [23].Kantelhardt JW, Bauer A, Schumann AY, Barthel P, Schneider R, Malik M, and Schmidt G, “Phase-rectified signal averaging for the detection of quasi-periodicities and the prediction of cardiovascular risk,” Chaos: An Interdisciplinary Journal of Nonlinear Science, vol. 17, no. 1, p. 015112, 2007. [DOI] [PubMed] [Google Scholar]
- [24].Albert A, “Estimating the infinitesimal generator of a continuous time, finite state markov process,” The Annals of Mathematical Statistics, pp. 727–753, 1962. [Google Scholar]
- [25].Ross SM, Kelly JJ, Sullivan RJ, Perry WJ, Mercer D, Davis RM, Washburn TD, Sager EV, Boyce JB, and Bristow VL, Stochastic processes. Wiley; New York, 1996, vol. 2. [Google Scholar]
- [26].Habtemichael TG, Goshu AT, and Buta GB, “Missclassification of hiv disease stages with continuous time hidden markov models,” Journal of Advances in Medicine and Medical Research, pp. 1–15, 2018. [Google Scholar]
- [27].Rabiner LR, “A tutorial on hidden markov models and selected applications in speech recognition,” Proceedings of the IEEE, vol. 77, no. 2, pp. 257–286, 1989. [Google Scholar]
- [28].Kalbfleisch J and Lawless JF, “The analysis of panel data under a markov assumption,” Journal of the american statistical association, vol. 80, no. 392, pp. 863–871, 1985. [Google Scholar]
- [29].Van Loan C, “Computing integrals involving the matrix exponential,” IEEE transactions on automatic control, vol. 23, no. 3, pp. 395–404, 1978. [Google Scholar]
- [30].Preacher KJ, Zhang G, Kim C, and Mels G, “Choosing the optimal number of factors in exploratory factor analysis: A model selection perspective,” Multivariate Behavioral Research, vol. 48, no. 1, pp. 28–56, 2013. [DOI] [PubMed] [Google Scholar]
- [31].Kleiger RE, Stein PK, and Bigger JT Jr, “Heart rate variability: measurement and clinical utility,” Annals of Noninvasive Electrocardiology, vol. 10, no. 1, pp. 88–101, 2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [32].von Rosenberg W, Chanwimalueang T, Adjei T, Jaffer U, Goverdovsky V, and Mandic DP, “Resolving ambiguities in the 1f/hf ratio: Lf-hf scatter plots for the categorization of mental and physical stress from hrv,” Frontiers in physiology, vol. 8, p. 360, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
