Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2026 Feb 10.
Published in final edited form as: IEEE J Biomed Health Inform. 2025 Feb 10;29(2):741–749. doi: 10.1109/JBHI.2024.3514678

Continuous-Time Hidden Markov Factor Model for Mobile Health Data: Application to Adverse Posttraumatic Neuropsychiatric Sequelae

Lin Ge 1, Xinming An 2,*, Donglin Zeng 3, Samuel McLean 4, Ronald Kessler 5, Rui Song 6,*
PMCID: PMC11970992  NIHMSID: NIHMS2056384  PMID: 40030431

Abstract

Each year, a significant portion of the 40 million individuals in the United States who seek care in emergency departments (EDs) following traumatic experiences develop adverse posttraumatic neuropsychiatric sequelae (APNS). This highlights the widespread impact of trauma and the critical need for effective interventions to address the health outcomes of these events. Despite significant research efforts, advancements in understanding the neurobiological mechanisms of APNS have been hindered, primarily due to reliance on subjective self-reports, which are susceptible to recall biases and careless responses. To overcome this limitation, we investigate the use of objective, longitudinal mobile device data to identify consistent APNS states and examine their dynamic transitions over time. To address the complexity of longitudinal mobile data, we developed a continuous-time hidden Markov factor model and applied it to mobile device data from the Advancing Understanding of Recovery after Trauma (AURORA) study. Findings from this study highlight the great potential of leveraging mobile device data for continuous health monitoring and for guiding personalized treatment approaches through mobile health initiatives. A Python implementation of the proposed method is available at https://anonymous.4open.science/r/CTHMFM.

Index Terms—: Continuous-time hidden Markov model, Mental health, Multivariate longitudinal data

I. Introduction

Adverse posttraumatic neuropsychiatric sequelae (APNS) (e.g., pain, depression, and PTSD) are frequently observed in civilians and military veterans who have experienced traumatic events. These APNS increase the risk of chronic illnesses, including cancer and heart disease, and substantially contribute to drug abuse, suicide, and disability. Moreover, APNS impose enduring psychosocial and financial burdens not only on individuals with the disorder but also on their families, communities, and society as a whole. However, little progress has been made in advancing APNS research over the past few decades due to unique challenges. First, APNS have been evaluated through subjective self-reported measures, which lack objective reliability. Second, the heterogeneity among patients, as recognized in traditional classification and diagnoses, complicates the study of APNS. Lastly, these APNS disorders are often studied and treated independently, despite their frequent co-occurrence [1]. These obstacles hinder the identification of objective markers, the advancement in understanding the neurobiological mechanisms of APNS, and the development of effective preventative/treatment strategies.

Identifying homogeneous states and exploring the dynamic prognosis of APNS in the immediate aftermath of trauma exposure holds promise for enhancing our understanding of APNS and identifying effective intervention options and appropriate timing at the individual level. Regrettably, due to the lack of appropriate data and effective statistical method, no large-scale studies have been conducted to investigate the onset, dynamic transitions (such as recovery and relapse), and associated risk factors of APNS. To help address the challenges, the National Institutes of Mental Health, joined by the US Army Medical Research and Material Command, developed the Advancing Understanding of RecOvery afteR traumA (AURORA) study [1]. This study gathered biobehavioral data from a large cohort of trauma survivors (n = 2,997) across the United States over a year, including self-reported surveys, web-based neurocognitive tests, digital phenotyping data (i.e., from wrist wearables and smartphones), psychophysical tests, neuroimaging assessments, and genomics data. In contrast to previous studies relying on self-report surveys or neuroimages [2], our work aims to address the challenges of APNS by utilizing the objective digital phenotyping data that tracks individuals’ behaviors, moods, and health statuses in real-time, real-life environments. Specifically, we developed a Hidden Markov Factor Model (HMFM) to analyze the mobile device data, allowing us to simultaneously identify homogeneous subtypes, investigate subtype-specific structures, and model individual progression with associated risk factors.

Hidden Markov Models (HMMs) have been widely used in various fields [3]. However, mobile device data presents two unique challenges that standard HMMs cannot handle, including the interdependent variables with unknown interrelationship structures and unevenly spaced measurements.

Mobile device sensor data, such as accelerometer and photoplethysmography (PPG) from smartwatches, involves highly intensive time series data. Instead of being directly used, these data are typically pre-processed to generate technical summaries that represent various characteristics of each time series variable, which are often highly correlated. As the number of features increases, the parameters in the covariance matrix expand exponentially, making a fully free covariance matrix impractical. Consequently, when implementing HMMs, features are usually assumed to be independent given the latent state membership, despite their correlation, potentially introducing estimation bias.

To appropriately model the association between features, factor analysis models (FMs) [4] provide an efficient and parsimonious approach and have been incorporated into HMMs in various ways. For example, the factor analyzed hidden Markov model [5] combines an FM with a discrete-time HMM (DTHMM), which assumes evenly spaced measurements. It has been extensively used in a variety of real-world applications, including speech recognition [6], environmental protection [7], and seizure detection [8]. Similarly, [9] introduced the regime-switching factor model to handle high-dimensional financial market data. However, they all assume homogeneous transition probability matrices, limiting their ability to account for the heterogeneity of transition probabilities over time and among different subjects and explore risk factors of state transition. To simultaneously capture the interrelationships among observed features and account for the variability of transition probabilities, a joint framework incorporating HMM, FM, and a feature-based transition model was recently proposed [10], [11]. However, it is not directly applicable to mobile device data. First, it employs a confirmatory factor model (CFM) with pre-specified structures for the factor loading matrices, which are often unknown a priori. Therefore, an exploratory factor model (EFM) is needed to explore the interrelationships among all observed features. Second, their framework assumes ordered states, which is inappropriate for our use case.

Another challenge posed by mobile device data is the irregular spacing of measurements. For example, activity and heart rate variability (HRV) data were collected only when the participants wore the watches, resulting in non-uniformly spaced observations and significant variation in sampling schedules between individuals. While the aforementioned methods are all based on DTHMM, neglecting the impact of time gaps between consecutive observations on transition rates, continuous-time discrete-state HMM (CTHMM) was developed to handle irregularly spaced measurements [12]. CTHMM and its extensions that incorporate covariates to characterize transition rates are widely used in medical research [13], [14]. However, none of them address the interrelationships among features.

In this paper, to simultaneously address the two challenges and examine heterogeneous transition patterns, we propose to use Continuous-Time HMFM (CT-HMFM), integrating CTHMM, EFM, and a feature-based transition model. Our contributions are as follows: First, we examine the utility of data collected in an open environment from consumer-grade mobile devices for mental health research. This contrasts with most existing studies on data collected in controlled lab environments. Second, we propose CT-HMFM to address the unique challenges introduced by mobile device data and depict the non-homogeneous state transition processes of multiple individuals. Simulation studies using synthetic data demonstrate exceptional parameter estimation and model selection performance. Finally, we analyze HRV and activity data from the AURORA study, followed by interpretations and discussions of biological findings that highlight the immense potential of mobile health data and our proposed method for mental health research.

II. AURORA Dataset

In the AURORA study [1], data were collected from multiple sources, with our analysis focusing on the accelerometer and PPG data collected by the Verily smartwatches. Given the known associations with APNS [15], [16], the accelerometer data were used to quantify physical activity features, while the PPG data were pre-processed to extract heart rate variability (HRV) metrics (for details on the raw data and pre-processing steps, refer to [17]).

Briefly, activity features are extracted during a 24-hour window to evaluate daily activity patterns. After converting accelerometer data to activity counts [18], meanAcc is the average activity counts. Amplitude, a cosinor rhythmometry feature, is computed to capture circadian rhythms [19]. Using the Cole-Kripke algorithm [20], accelerometry epochs are classified into wake or sleep states, and the SWCK quantifies transition rates between wake and sleep. Additionally, the average activity during the five least active hours (L5) is calculated from raw accelerometer data [21], representing nighttime activity.

HRV features were derived from PPG data by first calculating and denoising beat-to-beat (BB) interval [22] time series to obtain normal-to-normal (NN) intervals, which were then analyzed using a 5-minute sliding window. Selected HRV features for this study include average heart rate (NNmean), skewness (NNskew), and standard deviation (SDNN) of NN intervals, indicating heart rate variability and rapid changes. Lfhf is the low-frequency to high-frequency power ratio, with extremes suggesting parasympathetic or sympathetic dominance [22]. DC serves as a mortality risk indicator in cardiac conditions [23]. The lower the DC index, the higher the mortality risk. Additionally, SD1SD2 and ApEn are used to quantify the unpredictability and regularity of successive heartbeats (R-R interval). To match the activity data, daily statistical summaries of each HRV feature are used, including mean, minimum, maximum, interquartile range, and variance.

In this study, we focused on survivors of motor vehicle collision trauma. Since the data collection depends on the participants’ wearing of the devices, missing data is a common issue. To assure data quality, we include only records that have complete activity data and a positive wake percentage. For HRV data, ideally, an individual can have 2,880 records per day. We retain only those days where at least 30% of these records are available to ensure that our daily summary statistics are representative. The final dataset consists of daily summaries of activity and HRV features from 258 patients, with each providing at least 50 days of records.

III. CT-HMFM

Motivated by the structures of the processed AURORA datasets, we focus on data consisting of repeated measurements of p features collected over Ti time points for each individual i among N subjects. Let yit denote the observed values of the p features for subject i at time t, and let Yi=yi1,,yiTi denote the p×Ti matrix containing all measurements for individual i. To analyze these longitudinal data, we adopt a HMM framework, which facilitates the inference of latent states and their dynamic transitions over time. Specifically, we assume that the latent state of individual i at time t, denoted by wit, takes values from a finite discrete set {1,,J}. At t=1, the initial state is assumed to follow a multinomial distribution with probabilities π=π1,,πJ, such that i=1Jπi=1. Transition probabilities between theses states at time point t2 are captured by a J×J transition probability matrix Pit, of which the (k,j)-th entry is Pit,kj=Pwit=jwi,t-1=k, and Pit,kk=1-j:jkPit,kj. The primary goal is to estimate the π and Pit, thereby delineating latent Markov processes underlying the observed data. To meet our specific objectives of analyzing the mobile health data, we extend the traditional HMM framework with two additional components. The first, detailed in Section III-A, is a state-specific measurement model using EFM to uncover interrelationships among variables. The second, outlined in Section III-B, is a transition model (TM) designed to capture heterogeneous transition patterns.

A. State-Specific Measurement Model

The first component is an FM, aimed at identifying the interrelationship structures between observed response variables and the underlying latent variables. For individual i at time t, given that wit=j, the FM assumes that:

yitwit=j=μj+Λjzit+eit,zit~i.i.d.𝒩0,IK,eit~i.i.d.𝒩(0,Ψ),ziteit, (1)

where μj is a p×1 vector of state-specific expected mean response, Λj is a p×K state-specific factor loading matrix, zit is a K-dimensional vector of latent scores assumed to be independent of wit and following a standard multivariate normal distribution, and Ψ is a p×p diagonal covariance matrix for the error term eit with positive nonconstant diagonal entries. Alternatively, (1) can also be expressed as

yitwit=j~i.i.d.𝒩μj,ΛjΛj+Ψ. (2)

It is crucial to emphasize that, unlike CFMs that require pre-specified structures of the factor loading matrices, our approach imposes no assumptions on Λj. Consequently, the structure of Λj is entirely data-driven, making the first component (1) an EFM.

B. Transition Model

To account for the effects of varying time intervals between observations, a continuous-time Markov process is used. Instead of directly relying on the transition probability matrix P, the continuous-time Markov process is defined by a transition intensity matrix Q [24], which is the limit of P as the time interval approaches zero. Specifically, let δit denote the number of pre-specified time units between tth and (t-1)th observation, the transition intensity from state j to state k for subject i at time t is given by:

qjk=limδit0Pwit=kwi,t-1=jδit>0,jk, (3)

and qjj=-kjqjk. The corresponding transition probability Pδit can then be calculated as the matrix exponential of δit*Q. Since computing the matrix exponential directly can be computationally intensive, in practice, we approximate the exp(Q) using (I+Q/a)a for a sufficiently large a [25].

To investigate the impact of covariates on transition rates, the Q is modeled with a log-linear model [26], such that

logqjkxit=xitBjk, (4)

where xit is a d×1 vector of covariates, and Bjk is a d×1 vector of state-specific fixed effects coefficients, quantifying the effect of covariates on the transition intensity from state j to state k.

IV. Expectation-Maximization Algorithm

Let λ=μjj=1J,Λjj=1J,Ψ,Bkjk,j=1J,π,Wi=wi1,,wiTi, and Zi=zi1,,zit. Under the standard assumptions of HMMs–(1) given the current state wit, the observations yit are conditionally independent, and (2) given wit and the subject’s contextual features, the subsequent state wi,t+1 is independent of any historical information–the joint probability distribution of the observations and all latent variables, Lci(λ), is constructed as follows:

Pwi1t=2TiPwitwi,t-1,xitt=1TiPyitwit,zitPzit. (5)

Assuming the independence of Yi,Wi, and Zi across individuals, the complete likelihood function Lc for the entire sample is obtained as the product of (5) over i. To estimate λ, the expectation-maximization (EM) algorithm is used, as both Wi and Zi are unobserved. Specifically, the EM algorithm iteratively applies the expectation and maximization steps to find a local maximum of the marginal likelihood, as detailed below.

A. Expectation Step (E-step)

The E-step computes the expected value of the logarithm of Lc, given the observations and the current parameter estimates λv. Directly computing this expectation, EλvlogLc(λ)Y,X, is challenging due to the computational intensity in calculating the involved conditional state probabilities, including

ϵikjv(t)Pwi,t=j,wi,t-1=kYi,λv (6)

and

γijvtPwit=jYi,λv. (7)

To efficiently compute these probabilities, we utilize a scaled version of the forward-backward algorithm (FBA) [27].

Specifically, we first calculate the forward probability αij(t)=Pwit=jyi1,,yit. Let Pjyit be the probability density of yit given wit=j, and ci(t) the conditional probability of observation yit given all past observations. Omitting the dependence on δit for brevity, the forward probabilities are computed recursively as:

αij1=πjPjyi1j=1JπjPjyi1=πjPjyi1ci1; (8)
αij(t)=Pjyitk=1Jαik(t-1)Pitkjci(t), (9)

where ci(t)=j=1JPjyitk=1Jαik(t-1)Pitkj. Similarly, the backward probability βij(t), defined as Pyi,t+1,,yi,Tiwit=j,λci(t+1), is also obtained recursively as:

βijTi=1, (10)
βij(t)=k=1JPi,t+1,jkPkyi,t+1βik(t+1)ci(t+1). (11)

Using the forward and backward probabilities, the conditional state probabilities are derived as:

γijvt=αijtβijt, (12)
ϵikjv(t)=αikv(t-1)PitkjPjyitβijv(t)ci(t). (13)

With these, EλvlogLc(λ)Y,X can be expressed explicitly. While detailed expressions are provided in Appendix I-A, the resulted expression can be summarized as:

constant+hπ+hBkjk,j=1J-12hΨ,Λjj=1J,μjj=1J, (14)

where h(π) depends on the initial state distribution, hBkjk,j=1J depends on the probability transition matrix, and hΨ,Λjj=1J,μjj=1J is a function of parameters Ψ,Λj, and μj.

B. Maximization (M-step)

Within each M-step, since hΨ,Λjj=1J,μjj=1J,h(π), and hBkjk,j=1J do not share parameters, we maximize each of them separately. The estimator of π,Λj,μj, and Ψ are directly derived by setting h(π)=0 and hΨ,Λjj=1J,μjj=1J=0 (see Appendix I-B for details). For Bkjk,j=1J, a one-step Fisher scoring (FS) [28] algorithm is implemented to optimize the parameters.

Let θ be the ordered vector of all transition model parameters, such that θ=vecBkj,kj, with θu representing the uth entry of θ. Recalling the first derivative of the matrix exponential [14] and using Theorem 1 in [29], the derivation of expAθu with respect to θu is given by:

θuexpAθu=expAθuA˜θu0Aθu0:J,J:2J, (15)

where A˜θu=A˜ijθu=Aijθuθu. Let Pkjδitθu denote the (k,j) entry of the first derivative of Pδit with respect to θu. Having the first derivative of Pδit=expδit*Q with respect to each component of θ calculated accordingly, the FS can then be directly implemented to update θ without requiring the second derivative of the matrix exponential. Specifically, let S* be the score function, defined as

Su*θ=i=1Nt=2Tij=1Jk=1JϵikjvtPkjδitPkjδitθu. (16)

Let M* denote the negative Fisher information matrix, whose (u,v) entry is given by:

Muv*θ=i=1Nt=2Tij=1Jk=1Jγikvt-1PkjδitPkjδitθuPkjδitθv. (17)

With the score function and the Fisher information matrix, the parameters θ can be updated as

θv+1=θv+M*θv-1S*θv. (18)

To ensure numerical stability, we control the learning rate in practice by updating

θv+1=θv+M*θv+S*θvTS*θv-1S*θv. (19)

Note that the algorithm requires specification of (K,J), which are typically unknown in practice. In this study, we use information criteria to determine (K,J), the efficacy of which is evaluated in Section V-C.

V. Simulation Study

This section assesses the proposed methods through simulation studies using synthetic data resembling the AURORA dataset. We generate data with N=200,p=23,d=3,J=3, and K=3. Each individual’s number of observations, Ti, is uniformly sampled from [50, 100]. From this, we randomly select Ti time points from {1,,100} to create sequences of δit. Each individual’s initial state is drawn from a multinomial distribution with probabilities π=13,13,13. Latent state trajectories are then generated based on transition probabilities Pitδit, given individuals’ features. Observation vectors yit for an individual i in state j at time t are then sampled from a normal distribution with mean μj and covariance ΛjΛj+Ψ, where Ψ=I.

In the following, subsection V-A assesses model reliability by comparing empirical parameter estimates against true values; subsection V-B compares the effectiveness of our method against baseline methods; and subsection V-C explores the performance of information criteria in model selection.

A. Simulation 1

To validate the estimation procedure, we implement the EM algorithm with true J and K. Parameter initialization involves first fitting Gaussian Mixture Models to estimate groups, followed by EFM for each group. Guided by the insights from a pilot study, we set the maximum number of iterations for each replication at 100. The reliability and precision of the proposed methods are then evaluated from two perspectives: i) the accuracy of each individual parameter estimate and ii) the misclassification rate Cmis, which quantifies the proportion of estimated states that diverge from the actual states.

The accuracy of parameters π,μ,Λ, and Ψ is assessed by calculating the average absolute difference (AAD) between the estimates and their true values, defined as AAD(o)=i=1roiˆ-oir,oi is an individual entry in matrix o and r is the total number of free parameters. The mean of AADs (standard errors in the parentheses) aggregated over 100 random seeds are presented in Table I. These mean AADs for all parameter matrices are sufficiently close to zero with small standard errors, indicating effective parameter recovery. In Table II, we report the mean bias (standard error) for each parameter in the transition model, where the biases are all close to zero. Moreover, we present the mean (standard error) of Cmis in Table I. On average, misclassification rates are only .24% (0.0005), highlighting the exceptional accuracy of the proposed EM algorithm in estimating latent states.

TABLE I.

The Mean (standard error) AADs of π,μ,Λ,Ψ, and Cmis.

π μ Λ Ψ Cmis
.026 (.013) .015 (.002) .014 (.001) .011 (.002) .0024 (.0005)

TABLE II.

Bias (standard error) of the parameter estimates for Each transition model parameter Bjkl.

Bjk Bjk0 Bjk1 Bjk2
B12 .014(.142) .017(.150) −.080(.219)
B13 −.021(.126) .010(.125) .016(.182)
B21 −.018(.147) −.004(.139) .014(.214)
B23 −.003(.112) .004(.097) .005(.200)
B31 .003(.112) .002(.107) −.006(.192)
B32 .006(.131) −.032(.091) .028(.232)

Intuitively, factors such as N, Ti, sizes of J and K, variance Ψ, differences in μj and Λj between states, and frequency of state transitions, can affect the performance of parameter estimation. Additional simulations in Appendix II reveals that estimation accuracy for μ,Λ,Ψ, and B, as well as classification accuracy, improve when (i) common variances decrease, (ii) differences in μj and Λj between states increase, (iii) J decreases, or (iv) sample size (N) or panel length Ti increases. Increasing K or using a B that induces infrequent transitions slightly affects most parameters but enhances the precision of transition probability estimates, thereby reducing misclassification rates. Estimation of π improves solely with increases in N or state-to-state differences in μj.

B. Simulation 2

This section compares the performance of the proposed methods and the baseline approaches in correctly identifying latent states. Three benchmark methods are under our consideration: i) TM+independent HMM, which assumes independence among observed features given the states; ii) CFM+TM+HMM, which addresses interrelationships but inaccurately pre-specifies the latent factor structure by setting certain loading matrix entries to zero; and iii) EFM+HMM, which assumes a homogeneous transition probabilities matrix for all subjects. We first repeat the data generation process of Simulation 1. Then, we consider three additional scenarios by adjusting the state-to-state differences in μj to be closer (μ: medium diff), increasing the similarity of the Λj at different states (Λ: medium diff), and increasing the significance of the covariance matrix Ψ(Ψ=2×I), respectively.

As depicted in Figure 1, our proposed methods (HMFM) consistently outperform the benchmark methodologies in all settings. Regardless of sample size, our methods consistently achieve the lowest misclassification rate, nearly approximating zero, thereby emphasizing the importance of each component in our proposed models. Specifically, the comparison with TM+independent HMM shows the importance of accounting for the interrelationship between observed features; the comparison with CFM+TM+HMM reveals the risk of incorrectly specifying the interrelationship structure; and the comparison with EFM+HMM demonstrates the inadequacy of assuming homogeneous transition probabilities.

Fig. 1.

Fig. 1.

Cmis of various methods. The error bars represent the 95% CI. The first column shows the results under the settings we used in simulation 1. The last three columns summarize the results under different settings by varying the true value of μ,Λ, and Ψ, respectively.

C. Simulation 3

Information criteria such as the Akaike information criteria (AIC) and the Bayesian information criteria (BIC) have been widely used in model selection [10], [30]. Within this simulation study, we investigate whether the AIC or BIC is reliable for determining J and K simultaneously. We repeat the data generation process of Simulation 1, but implement the proposed methods with a different set of (J,K) for each replicate when fitting the generated data. Let J={2,3,4} and K={2,3,4}. We consider all possible combinations of J and K, yielding a total of nine fitted candidate models for each replicate. For 100 replications, both AIC and BIC consistently recommend the model with accurate J and K. Therefore, we believe that the sample size and the number of observations per individual in the processed AURORA data will yield reliable information criteria-based model selection results and, consequently, reliable parameter estimation.

VI. Analysis of the AURORA Data

Given the irregular data collection from mobile devices, we applied the CT-HMFM to smartwatch data from the AURORA study. Considering J=1,2,,6 and K=1,2,,9, we evaluate 54 candidate models. For each candidate model, the EM algorithm is implemented with multiple random seeds, and the seed yielding the highest estimated likelihood is selected. Finally, model comparison using AIC and BIC led to the selection of a model with J=3 and K=8. Subsequent subsections will discuss the interpretation of parameter estimates and biological insights from three perspectives: i) the interpretation of the three estimated states, ii) symptom co-occurrence patterns, and iii) the influence of demographic factors on transition probabilities.

A. Interpretation of Hidden States

To investigate the biological differences between states, we first focus on the selected features. Figure 2 depicts scaled sample means for each feature across states, with a 99% confidence interval (CI). Further pairwise Tukey tests indicate significant differences between states for nearly all features, except for amplitude, SWCK, L5, and NNskew.q3 between state 1 and state 2. Specifically, features related to average heart rate (NNmean), heart rate variability (SDNN), and heart deceleration capacity (dc) vary significantly across the three latent states, decreasing sequentially from state 1 to state 3. According to previous research, lower heart rate variability and deceleration capacity are associated with a higher mortality rate [23], [31], suggesting that states 1 through 3 represent decreasing levels of health, with state 1 being the healthiest. Additionally, activity levels are similar and higher in states 1 and 2 compared to state 3, indicating better overall health for participants in the first two states. Among features related to heart rate unpredictability (lfhf, ApEn, and SD1SD2), state 3 demonstrates significantly higher values for SD1SD2-related features but lower values for lfhf-related features compared to states 1 and 2, suggesting a different interpretation of the estimated states than our previous interpretation. However, it is important to note that the relationship between these features and the psychological or physiological state is neither straightforward nor unique [32].

Fig. 2.

Fig. 2.

Relative sample mean for features in each estimated state, with 99% CI error bars that are too small to distinguish.

To confirm the validity of the three states, we further compare their differences regarding self-reported symptoms from a flash survey. Based on the RDoC framework, ten latent constructs associated with APNS were developed using survey items selected by domain experts: Pain, Loss, Sleep Discontinuity, Nightmare, Somatic Symptoms, Mental Fatigue, Avoidance, Re-experience, and Anxious. Retaining only observations for each individual whose estimated states are known on the same day they submitted survey responses, we summarized the flash survey data with means and 95% CIs in Figure 3. Overall, state 1 exhibits the lowest severity level for all ten symptoms, while state 3 has the highest severity level. Tukey tests reveal no significant differences between states 1 and 2 in hyperarousal, re-experience, anxiety, and somatic symptoms, but both are significantly different from state 3 in these constructs. While the differences in nightmare and sleep discontinuity between states 3 and 2 are not significant, they are statistically more severe than in state 1. For mental fatigue and depression, only the difference between state 1 and state 3 is statistically significant.

Fig. 3.

Fig. 3.

Sample mean for each symptom in each estimated state. The error bars represent the 95% CI. 0 is the least severity, 1 is the greatest.

In summary, both the flash survey data and the AURORA data (HRV, Activity) support our interpretation of the three latent states. State 1 is the healthiest, while state 3 indicates having the most severe APNS symptoms.

B. Co-occurring Pattern of Symptoms

When studying the co-occurring of symptoms within each hidden state, we limit our attention to observations collected during the first week. For each state, the correlations between all ten symptoms are calculated. In state 1 (relative health state), there is a high degree of correlation (.782) between hyperarousal and anxiety, suggesting that patients in state 1 experiencing severe hyperarousal symptoms are also likely to suffer from severe anxiety. State 2 shows no highly correlated symptoms. In state 3 (the state with more severe disorders), symptoms such as depression, hyperarousal, anxiety, and re-experience are more likely to co-occur with pairwise correlation ranging from .716 to .91.

C. Transition Probability

This section investigates the heterogeneity of 1-day transition probabilities among subjects, focusing on transitions with a time interval δit=1. We estimated the transition probabilities for males and females within the sample age range, as depicted in Figure 4. Lines with circles illustrate the probability of remaining in the same state, lines with stars indicate transitions to a more severe state, and lines with ‘x’ reflect the chance of improvement in psychological conditions.

Fig. 4.

Fig. 4.

Estimated transition probability. Fix δit=1. (a, b) indicates a transition from state a to state b.

Overall, both males and females have a tendency to remain in their current state, with infrequent state transitions, aligning with most literature. For males, the probability of staying at states 3 and 2 increases with age, while it decreases for state 1. Moreover, while the likelihood of psychological deterioration increases with age, the chance of improvement decreases. Specifically, while the probability of transitioning from the most severe state (state 3) to the healthiest state (state 1) approaches zero as age increases, the likelihood of the reverse transition increases as age decreases, with direct transitions between state 1 and state 3 being particularly rare. The female group exhibits a similar trend to the male group, but with a higher likelihood of remaining in the most severe state (state 3) compared to males.

In summary, our analysis of the AURORA data suggests that older patients are more likely to transition to more severe psychological states. Moreover, achieving psychological improvement becomes increasingly challenging as one ages.

D. State Specific Associations Between Activity and HRV Features

The model provides a deeper understanding of the latent states, not only in terms of their means but also through the associations between activity and HRV features. These associations offer new insights into the underlying physiology of APNS symptoms. The state-specific associations are captured in the structure of the factor loading matrix, which reveals both shared patterns and distinct differences across states, reflecting heterogeneous relationships. For example, the analysis shows a consistent positive correlation between greater daytime activity and better sleep quality across all states. Additionally, heart rate variability and irregularity (measured by SDNN and ApEn) are correlated across all states. However, their association with heart deceleration capacity (dc-related features) is notably stronger in states 2 and 3, highlighting state-specific physiological patterns.

VII. Conclusions and Discussion

This paper explores the unique challenges and potential of using intensive mobile device data to provide objective measures and study the dynamic transitions of APNS symptoms after trauma exposure. The study addresses challenges such as interdependent variables with unknown relationships, heterogeneous transition probabilities, and irregular measurement intervals. The findings suggest that mobile device data from consumer-grade devices, combined with the proposed analytical methods, hold immense promise for assessing, monitoring, and understanding the dynamic progression of APNS symptoms post-trauma.

Traditionally, APNS assessments rely heavily on self-reported measures, which are inherently subjective and prone to recall biases or careless responses. By contrast, mobile device data collected through consumer-grade wearables like smartwatches and fitness trackers offer an objective and continuous perspective on individuals’ physical and mental health. While previous studies have largely relied on data collected in controlled environments, such as sleep labs or clinical settings, data obtained in natural, real-world environments introduces additional noise and presents unique analytical challenges. Extracting meaningful health-related insights from these data remains an open question.

This study demonstrates the immense potential of mobile device data as a secondary objective measure for a broad spectrum of APNS symptoms. These real-time data not only allow clinicians to monitor patients continuously in their natural environments but also provide more timely insights compared to periodic self-reports or clinic visits. By mapping how individuals transition between health states over time and identifying high-risk groups and recovery factors, clinicians gain a better understanding of the natural progression of trauma recovery. This knowledge enables the design of personalized, preventive interventions tailored to individual needs.

Furthermore, the proposed model is adaptable to data from other mobile devices, such as smartphones. Given the widespread ownership and daily use of mobile devices, this approach offers a scalable solution for continuous health monitoring. It also holds significant potential for personalized treatment through mobile health initiatives, particularly in under-resourced areas where access to traditional clinic-based care is limited.

Acknowledgments

The research is partially supported by NSF under Grant DMS-1555244 and 2113637, NIMH under Grant U01MH110925, the US Army Medical Research and Material Command, The One Mind Foundation, and The Mayday Fund.

Appendix I

Technical Details

A. Supplement for E-step

Denote Λj~=Λj,μjp×(K+1) and zit~=zitT,1T(K+1). Each of the three parts has an explicit form:

hπ=i=1Nj=1Jγijv1logπj, (20)
hBkjk,j=1J=i=1Nt=2Tij,k=1JϵikjvtlogPitkj, (21)
h(Ψ,{Λj}j=1J,{μj}j=1J)=i=1Nt=1Tij=1Jγijv(t)log|Ψ|+γijv(t)yitΨ1yit2γijv(t)yitΨ1Λj˜Eλv(zit˜yit,wit)+γijv(t)tr(Λj˜Ψ1Λj˜Eλv(zit˜zit˜yit,wit)). (22)

B. Supplement for M-step

By setting the first derivative of h(π) to zero, the parameters related to the initial state distribution are estimated as:

πjnew=i=1Nγijv(1)i=1Nk=1Jγikv(1). (23)

Similarly, the parameters used to characterize the conditional distribution of yit given wit are estimated by setting the first derivative of hΨ,Λjj=1J,μjj=1J equal to 0, with

Λ˜jnew=i=1Nt=1TiγijvtyitEλvzit˜yit,witi=1Nt=1TiγijvtEλvzit˜zit˜yit,wit. (24)

Meanwhile, we got the updated estimation of Ψ,Ψnew, equals

diagi=1Nt=1Tij=1Jγijv(t)yit-Λ~jnewEλvzit~yit/i=1NTi. (25)

Appendix II

Additional Simulation Results

This section presents additional simulation results that investigate the impact of various factors on estimation performance. Using the settings for simulation 1 in Section V-A as the baseline, we conducted eight additional sets of simulations, each varying one component while maintaining rest components as the baseline setup. These components include: i) sample size (N), ii) number of measurements for each individual Ti, iii) J, iv) K, v) size of common variance Ψ, vi) state-to-state difference in μj, vii) state-to-state difference in Λj, and viii) transition frequency. For common variance Ψ, we evaluated scenarios with .1I, .5I, and 1I (baseline). For μj and Λj, we adjusted state-to-state differences in two additional settings. For the test evaluating the effect of transition frequency, a frequent transition is defined as the probability of remaining in the same state being less than 0.70. The B in the baseline setting corresponds to infrequent transition. For tests evaluating the effects of J, K, N, and Ti, the baseline setups are modified as indicated in Table III.

TABLE III.

The Mean (standard error) AADs of π,μ,Λ,Ψ, and B, and Cmis of estimations under different CT settings.

ADD π μ Λ Ψ B Cmis
Ψ=1*I .026(.013) .015(.002) .014(.001) .011(.002) .120(.027) .0024(.0005)
Ψ=.5*I .026(.013) .012(.002) .011(.001) .005(.001) .119(.027) .0003(.0001)
Ψ=.1*I .026(.013) .010(.003) .007(.001) .001(.000) .119(.027) .0000(.0000)
μ: large diff .026(.013) .015(.002) .014(.001) .011(.002) .120(.027) .002(.0005)
μ: medium diff .027(.013) .015(.001) .015(.001) .011(.002) .124(.028) .007(.001)
μ: minor diff .034(.017) .017(.003) .478(.057) .011(.002) .161(.040) .085(.004)
Λ: large diff .026(.013) .015(.002) .014(.001) .011(.002) .120(.027) .002(.0005)
Λ: medium diff .026(.013) .014(.002) .014(.001) .011(.002) .122(.029) .006(.001)
Λ: minor diff .026(.013) .014(.002) .014(.001) .011(.002) .122(.028) .004(.001)
B: infreq transit .026(.013) .015(.002) .014(.001) .011(.002) .120(.027) .002(.0005)
B: freq transit .027(.013) .015(.002) .014(.001) .011(.002) .137(.043) .007(.001)
J=2 .027(.021) .012(.002) .012(.001) .011(.002) .087(.026) .0015(.0003)
J=3 .026(.013) .015(.002) .014(.001) .011(.002) .120(.027) .0024(.0005)
J=4 .024(.011) .017(.002) .016(.001) .011(.001) .150(.028) .0032(.0005)
K=2 .026(.013) .015(.002) .014(.001) .010(.002) .124(.027) .0069(.0008)
K=3 .026(.013) .015(.002) .014(.001) .011(.002) .120(.027) .0024(.0005)
K=5 .026(.013) .015(.002) .015(.001) .012(.002) .119(.027) .0005(.0002)
N=50 .052(.029) .030(.004) .028(.002) .022(.003) .254(.058) .0027(.0008)
N=100 .042(.019) .021(.003) .020(.001) .015(.002) .175(.038) .0025(.0006)
N=500 .018(.010) .009(.001) .009(.001) .007(.001) .076(.017) .0024(.0003)
10Ti30 .029(.014) .029(.004) .027(.002) .021(.004) .251(.061) .0028(.0009)
30Ti50 .027(.014) .020(.003) .019(.001) .015(.003) .180(.043) .0025(.0005)
100Ti150 .028(.015) .012(.001) .011(.001) .008(.001) .098(.021) .0023(.0003)

Contributor Information

Lin Ge, Department of Statistics, North Carolina State University, Raleigh, NC 27695 USA..

Xinming An, Department of Anesthesiology, The University of North Carolina at Chapel Hill, Chapel Hill, NC 27514 USA..

Donglin Zeng, Department of Biostatistics, The University of North Carolina at Chapel Hill, Chapel Hill, NC 27599 USA..

Samuel McLean, Department of Psychiatry, The University of North Carolina at Chapel Hill, Chapel Hill, NC 27514 USA..

Ronald Kessler, Department of Health Care Policy, Harvard Medical School, Boston, MA 02115 USA..

Rui Song, Department of Statistics, North Carolina State University, Raleigh, NC 27695 USA..

References

  • [1].McLean SA, Ressler K, Koenen KC, Neylan T, Germine L, Jovanovic T, Clifford GD, Zeng D, An X, Linnstaedt S et al. , “The aurora study: a longitudinal, multimodal library of brain biology and function after traumatic stress exposure,” Molecular psychiatry, vol. 25, no. 2, pp. 283–296, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Marquand AF, Wolfers T, Mennes M, Buitelaar J, and Beckmann CF, “Beyond lumping and splitting: a review of computational approaches for stratifying psychiatric disorders,” Biological psychiatry: cognitive neuroscience and neuroimaging, vol. 1, no. 5, pp. 433–447, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Mor B, Garhwal S, and Kumar A, “A systematic review of hidden markov models and their applications,” Archives of computational methods in engineering, vol. 28, no. 3, pp. 1429–1448, 2021. [Google Scholar]
  • [4].Kim J-O and Mueller CW, Factor analysis: Statistical methods and practical issues. sage, 1978, vol. 14. [Google Scholar]
  • [5].Rosti AI and Gales MJ, “Factor analysed hidden markov models,” in 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 1. IEEE, 2002, pp. I–949. [Google Scholar]
  • [6].Rosti AI and Gales M, “Factor analysed hidden markov models for speech recognition,” Computer Speech & Language, vol. 18, no. 2, pp. 181–200, 2004. [Google Scholar]
  • [7].Maruotti A, Bulla J, Lagona F, Picone M, and Martella F, “Dynamic mixtures of factor analyzers to characterize multivariate air pollutant exposures,” The Annals of Applied Statistics, vol. 11, no. 3, pp. 1617–1648, 2017. [Google Scholar]
  • [8].Madadi M, “A hidden markov factor analysis framework for seizure detection in epilepsy patients,” 2019. [Google Scholar]
  • [9].Liu X and Chen R, “Regime-Switching Factor Models for High-Dimensional Time Series,” Statistica Sinica, vol. 26, no. 4, pp. 1427–1451, 2016. [Online]. Available: https://www.jstor.org/stable/44114341 [Google Scholar]
  • [10].Song X, Xia Y, and Zhu H, “Hidden Markov latent variable models with multivariate longitudinal data.” Biometrics, vol. 73, no. 1, pp. 313–323, 2017. [Online]. Available: 10.1111/biom.12536 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [11].Zhou X, Kang K, Kwok T, and Song X, “Joint hidden markov model for longitudinal and time-to-event data with latent variables,” Multivariate Behavioral Research, vol. 57, no. 2–3, pp. 441–457, 2022. [DOI] [PubMed] [Google Scholar]
  • [12].Cox DR and Miller HD, The theory of stochastic processes. Routledge, 2017. [Google Scholar]
  • [13].Amoros R, King R, Toyoda H, Kumada T, Johnson PJ, and Bird TG, “A continuous-time hidden markov model for cancer surveillance using serum biomarkers with application to hepatocellular carcinoma,” Metron, vol. 77, no. 2, pp. 67–86, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [14].Zhou J, Song X, and Sun L, “Continuous time hidden markov model for longitudinal data,” Journal of Multivariate Analysis, vol. 179, p. 104646, 2020. [Google Scholar]
  • [15].Hartmann R, Schmidt FM, Sander C, and Hegerl U, “Heart rate variability as indicator of clinical state in depression,” Frontiers in psychiatry, vol. 9, p. 735, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [16].Jung W, Jang K-I, and Lee S-H, “Heart and brain interaction of psychiatric illness: a review focused on heart rate variability, cognitive function, and quantitative electroencephalography,” Clinical Psychopharmacology and Neuroscience, vol. 17, no. 4, p. 459, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [17].Cakmak AS, Alday EAP, Da Poian G, Rad AB, Metzler TJ, Neylan TC, House SL, Beaudoin FL, An X, Stevens JS et al. , “Classification and prediction of post-trauma outcomes related to ptsd using circadian rhythm changes measured via wrist-worn research watch in a large longitudinal cohort,” IEEE journal of biomedical and health informatics, vol. 25, no. 8, pp. 2866–2876, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [18].Borazio M, Berlin E, Kücükyildiz N, Scholl P, and Van Laerhoven K, “Towards benchmarked sleep detection with wrist-worn sensing units,” in 2014 IEEE International Conference on Healthcare Informatics. IEEE, 2014, pp. 125–134. [Google Scholar]
  • [19].Cornelissen G, “Cosinor-based rhythmometry,” Theoretical Biology and Medical Modelling, vol. 11, no. 1, pp. 1–24, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [20].Cole RJ, Kripke DF, Gruen W, Mullaney DJ, and Gillin JC, “Automatic sleep/wake identification from wrist activity,” Sleep, vol. 15, no. 5, pp. 461–469, 1992. [DOI] [PubMed] [Google Scholar]
  • [21].Van Someren EJ, Swaab DF, Colenda CC, Cohen W, McCall WV, and Rosenquist PB, “Bright light therapy: improved sensitivity to its effects on rest-activity rhythms in alzheimer patients by application of nonparametric methods,” Chronobiology international, vol. 16, no. 4, pp. 505–518, 1999. [DOI] [PubMed] [Google Scholar]
  • [22].Shaffer F and Ginsberg JP, “An overview of heart rate variability metrics and norms,” Frontiers in public health, vol. 5, p. 258, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [23].Kantelhardt JW, Bauer A, Schumann AY, Barthel P, Schneider R, Malik M, and Schmidt G, “Phase-rectified signal averaging for the detection of quasi-periodicities and the prediction of cardiovascular risk,” Chaos: An Interdisciplinary Journal of Nonlinear Science, vol. 17, no. 1, p. 015112, 2007. [DOI] [PubMed] [Google Scholar]
  • [24].Albert A, “Estimating the infinitesimal generator of a continuous time, finite state markov process,” The Annals of Mathematical Statistics, pp. 727–753, 1962. [Google Scholar]
  • [25].Ross SM, Kelly JJ, Sullivan RJ, Perry WJ, Mercer D, Davis RM, Washburn TD, Sager EV, Boyce JB, and Bristow VL, Stochastic processes. Wiley; New York, 1996, vol. 2. [Google Scholar]
  • [26].Habtemichael TG, Goshu AT, and Buta GB, “Missclassification of hiv disease stages with continuous time hidden markov models,” Journal of Advances in Medicine and Medical Research, pp. 1–15, 2018. [Google Scholar]
  • [27].Rabiner LR, “A tutorial on hidden markov models and selected applications in speech recognition,” Proceedings of the IEEE, vol. 77, no. 2, pp. 257–286, 1989. [Google Scholar]
  • [28].Kalbfleisch J and Lawless JF, “The analysis of panel data under a markov assumption,” Journal of the american statistical association, vol. 80, no. 392, pp. 863–871, 1985. [Google Scholar]
  • [29].Van Loan C, “Computing integrals involving the matrix exponential,” IEEE transactions on automatic control, vol. 23, no. 3, pp. 395–404, 1978. [Google Scholar]
  • [30].Preacher KJ, Zhang G, Kim C, and Mels G, “Choosing the optimal number of factors in exploratory factor analysis: A model selection perspective,” Multivariate Behavioral Research, vol. 48, no. 1, pp. 28–56, 2013. [DOI] [PubMed] [Google Scholar]
  • [31].Kleiger RE, Stein PK, and Bigger JT Jr, “Heart rate variability: measurement and clinical utility,” Annals of Noninvasive Electrocardiology, vol. 10, no. 1, pp. 88–101, 2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [32].von Rosenberg W, Chanwimalueang T, Adjei T, Jaffer U, Goverdovsky V, and Mandic DP, “Resolving ambiguities in the 1f/hf ratio: Lf-hf scatter plots for the categorization of mental and physical stress from hrv,” Frontiers in physiology, vol. 8, p. 360, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES