Abstract
The problem of analyzing temporally ordered sequences of observations generated by molecular, physiological or psychological processes to make predictions about the outcome of these processes arises in many domains of clinical informatics. In this paper, we focus on predicting the outcome of patient-provider communication sequences in the context of the clinical dialog. Specifically, we consider prediction of the motivational interview success (i.e. eliciting a particular type of patient behavioral response) based on an observed sequence of coded patient-provider communication exchanges as a sequence classification problem. We propose two solutions to this problem, one that is based on Recurrent Neural Networks (RNNs) and another that is based on Markov Chain (MC) and Hidden Markov Model (HMM), and compare the accuracy of these solutions using communication sequences annotated with behavior codes from the real-life motivational interviews. Our experiments indicate that the deep learning-based approach is significantly more accurate than the approach based on probabilistic models in predicting the success of motivational interviews (0.8677 versus 0.7038 and 0.6067 F1-score by RNN, MC and HMM, respectively, when using undersampling to correct for class imbalance, and 0.8381 versus 0.7775 and 0.7520 F1-score by RNN, MC and HMM, respectively, when using over-sampling). These results indicate that the proposed method can be used for real-time monitoring of progression of clinical interviews and more efficient identification of effective provider communication strategies, which in turn can significantly decrease the effort required to develop behavioral interventions and increase their effectiveness.
Introduction
Temporally ordered sequences of discrete or continuous observations generated by molecular, psychological or psychological process(es) arise in many different areas of biology and medicine (e.g., DNA base-pairs, protein sequences, ECG measurements, laboratory results, diagnostic codes, utterances in the clinical dialog). Classification (or categorization) is a type of analysis of those sequences that has a broad range of important practical applications, from protein function1 or structure2 prediction to detecting individuals with a heart disease3. Taking into account both the entire set of observations in a sequence, as well as the temporal order and potential dependencies between observations, makes sequence classification a more challenging task than a classification of independent observations. Predicting the outcome of those sequences (e.g. physiological or behavioral response) can also be viewed as a sequence classification problem.
In general, sequence classification methods fall into one of three major classes: feature-based, distance-based and model-based. Feature-based methods transform a sequence into a feature vector and apply a standard supervised machine learning method, such as Support Vector Machine4 or Decision Tree5 to arrive at classification decision. The methods in this class have had limited success since traditional feature representation methods cannot easily account for the order of and dependencies between observations in a sequence.
Distance-based methods classify a sequence by finding the most similar sequences with known classes based on a distance metric. The most commonly used distance metric is Euclidean distance with Dynamic Time Wrapping6. However, distance metrics are primarily designed for time series data, in which the observations are discretized by timestamps. The third type of sequence classification methods first creates a probabilistic model, such as the Markov Chain (MC) or Hidden Markov Model7 (HMM), for sequences in each class based on the training data and then, classifies new sequences by applying the created models. While MCs and HMMs can capture first- and second-order dependencies between adjacent observations in a sequence, learning higher-order dependencies with these models requires prohibitively large amounts of data. By encoding sequences into low-dimensional representations, Recurrent Neural Networks (RNNs) are able to capture both short- and long-term dependencies and were shown to be effective at modeling different types of sequential data8. Long Short-Term Memory (LSTM)9 is a variant of RNNs, which successfully addressed the vanishing gradient problem10 of traditional RNN. LSTM demonstrated excellent performance in different domains, from speech11 and handwriting recognition12 to health informatics13, 14. Specifically, LSTM was used as part of a multi-label classification method to recognize patterns in multivariate time series of clinical measurements, such as body temperature, heart rate and blood pressure13. LSTM was also effectively used for predicting the diagnosis and medication codes, given a sequence of codes from the previous patient visits14. A further simplification and improvement of LSTM model, called the Gated Recurrent Unit (GRU)15, was later proposed. LSTM and GRU demonstrated markedly better performance among all other RNN variants for a variety of tasks in different domains.
In this paper, we address the problem of predicting the outcome of coded patient-provider communication (PPC) sequences in the context of the clinical dialog. Specifically, we focus on predicting the success (i.e. eliciting a particular type of patient behavioral response) of motivational interviews with obese adolescents and their caregivers based on an observed sequence of coded PPC exchanges during those interviews. Childhood obesity is a serious public health concern in the United States. Recent estimates indicate that approximately one-third (31.8%) of U.S. children 2-19 years of age are overweight and 16.9% are obese16. Adolescents, who are obese, are likely to be obese in adulthood and have a greater risk of heart disease, type 2 diabetes, stroke, cancer, and osteoarthritis17. One approach to effective obesity intervention is Motivational Interviewing (MI), an evidence-based counseling technique to increase intrinsic motivation and self-efficacy for health-related behavior change. The goal of MI is to encourage patients to explore their own desires, ability, reasons, need for and commitment to the targeted behavior change. These statements collectively referred to as “change talk” (CHT), consistently predict the actual behavior change18 that can be sustained for as long as 34 months19 after an interview. However, the ability of providers to consistently elicit this type of patient communication requires knowledge of effective communication strategies for a variety of patients, which can only be obtained through analysis of a large number of annotated interviews. Since manual examination and analysis of MI interview transcripts is a very time-consuming process, designing effective MI interventions and tailoring them to particular populations can take years. Therefore, there is a need for informatics-based methods to facilitate the development of effective behavioral interventions, in general, and theoretically-grounded computational models to explore the mechanisms of MI’s efficacy, in particular.
Our goal is to compare the accuracy of probabilistic models, such as MC and HMM, and deep learning methods, such as LSTM and GRU, for the task of predicting the success of clinical interviews (i.e. eliciting a particular type of patient behavioral response, such as CHT) at any point during a clinical interview based on a sequence of coded previous PPC exchanges in the same interview. This study is a continuation of our previous work20, 21, in which we explored several machine learning methods for automatic annotation of clinical interview fragments with a large number of patient and provider behavior codes from a specialized codebook22. While there have been some previous qualitative studies of patient-provider dialog in a clinical setting23, no previous work explored the applicability of state-of-the-art methods for sequence modeling to the analysis of PPC exchanges, in general, and predicting the desired patient behavioral response in the context of motivational interviews, in particular.
Methods
Data collection
The experimental dataset for this work was constructed from the transcripts of 129 motivational interviews, which consist of a total of 50,239 segmented and annotated utterances. Each transcript corresponds to an MI interview session, which typically involves a counselor, an adolescent and a caregiver. The utterances were annotated based on the MYSCOPE codebook22, in which the behavior codes are grouped into the patient (adolescent and caregiver) codes and the counselor codes. Annotated utterances were divided into successful and unsuccessful communication sequences. Successful communication sequences are the ones, which resulted in positive change talk (CHT+) or commitment language (CML+) statements by an adolescent or a caregiver, while unsuccessful sequences are the ones, which resulted in negative change talk (CHT-) or commitment language (CML-), or the ones, in which no change talk or commitment language statements were made.
A fragment of an adolescent session transcript is presented in Table 1. In this example, SS → OQO → HUPO → OQTBN → CHT + is a successful sequence, in which a counselor starts with an open-ended question and ultimately is able to elicit a positive change talk statement. As follows from this example, similar utterances, such as “Yeah” and “Yes”, can be assigned different behavior codes (CHT+ and HUPW), depending on the context.
Table 1:
Fragment of the annotated transcript of a dialogue between a counselor and an adolescent. MYSCOPE codes assigned to the utterances and their meaning are shown in the first two columns.
| Code | Behavior | Speaker | Utterance |
|---|---|---|---|
| SS | Structure Session | Counselor | Okay. Can I meet with Xxxx alone for a few minutes? |
| OQO | Open-ended question, other | Counselor | So, Xxxx, how you doing? |
| HUPO | High uptake, other | Adolescent | Fine |
| OQTBN | Open-ended question, target behavior neutral | Counselor | That’s good. So, tell me how do you feel about your weight? |
| CHT+ | Change talk positive | Adolescent | It’s not the best. |
| CQECHT+ | Closed question, elicit change talk positive | Counselor | It’s not the best? |
| CHT+ | Change talk positive | Adolescent | Yeah |
| CQTBN | Closed question, target behavior neutral | Counselor | Okay, so have you tried to lose weight before? |
| HUPW | High uptake, weight | Adolescent | Yes |
The resulting experimental dataset was highly imbalanced. Out of 5143 observed sequences, 4225 or 82.15% were positive and only 918 or 17.85% were negative. No major differences were observed in the average length of successful (9.79 utterances) and unsuccessful (9.65 utterances) sequences.
Since severely imbalanced datasets often distort the true performance of a classification method relative to a simple “majority vote” baseline (e.g. simply classifying every communication sequence as successful would result in 82.15% accuracy on our dataset), it is important to properly address the class imbalance. We evaluated the performance of probabilistic and deep learning methods using both under-sampling and over-sampling for balancing the number of samples in different classes. Synthetic Minority Over Sampling Technique (SMOTE)24 is a widely used oversampling method for imbalanced datasets, in which new synthetic examples are generated for minority classes. Specifically, we generated synthetic examples at the borderline between the majority and minority classes25. On the other hand, the under-sampling method reduces the number of samples in majority class by replacing the clusters of samples identified by the k-means clustering algorithm with the cluster centroids.
Sequence classification methods
In general, a sequence can be viewed as a temporally ordered set of observations. In this study, an observation corresponds to a behavior code, which has a symbolic representation, such as LUP+ (low uptake, positive), OQECHT+ (open-ended question, elicit change talk positive), etc. Given a sequence of behavior codes Si = {c1, c2,…, cn} representing PPC exchanges during some part of a motivational interview, the task of predicting interview success can be considered as sequence classification. Given a set of class labels L = {l1, l2,…, lm} (in our case, the labels are “successful” and “unsuccessful” motivational interview), a sequence classifier C learns a function Si → li, li ∈ L that maps a sequence Si into a class label li ∈ L.
Our proposed baseline prediction method consists of two steps. In the first step, we model successful and unsuccessful patient-provider interactions using first and second-order Markov Chain and Hidden Markov Model, which are popular probabilistic models for discrete observation sequences with finite vocabulary. In the second step, we classify each test sequence based on the maximum likelihood of generating that sequence from each model. Although HMM was originally developed for speech recognition7, it is one of the most widely used methods for sequence modeling26, 27. However, the latest advances in deep learning suggest that RNNs may provide better results than conventional machine learning methods for the task of sequence classification. To verify this hypothesis, we employed two state-of-the-art variants of RNN in our experiments: Long Short-Term Memory (LSTM)9 and Gated Recurrent Unit (GRU)15.
Markov Chain (MC) is a probabilistic model that conditions each observation in a sequence only on preceding observation and not on any other past observation. First, we estimated two Markov models M and M, summarizing counselor strategies and patient responses, in the cases of successful (M) and unsuccessful (M) motivational interviews. A Markov model M can be represented as a weighted directed graph G = (V, E, p), in which:
V = {CML+, CHT+, CHT–, AMB–, LUP+, LUP–, HUPW, OQO, CQTBN, CQECHT+,…} is a set of vertices, consisting of adolescent, caregiver and counselor MI behavior codes;
E ⊆ V × V is a set of edges corresponding to possible transitions from one MI behavior code to the other in a sequence;
- pM: E → [0…1] is a function that assigns probability p(ci|cj) to an edge between the MI behavior codes ci and cj based on the maximum likelihood estimation:
(1)
where nci,cj and nci are the number of times a transition between the MI behavior codes ci and cj and the number of times the code ci have been observed in the training data, respectively. Given a Markov model M (such that S ⊆ V), the probability that a sequence of MI behavior codes S = {C1,…,CN} has been generated from a Markov model M is:
| (2) |
In the second step, we quantify the likelihood of success of a given motivational interview at a certain time point given a sequence of MI behavior codes S observed prior to that point as:
| (3) |
If p(S → successful) > 0, a communication sequence is predicted to be successful (i.e. result in positive change talk or commitment language). Otherwise, it is predicted to be unsuccessful.
The above model is also referred as first-order MC, since it only considers immediately preceding behavior code, when computing the state transition probabilities. In our experiment, we also considered second-order Markov model, which conditions each observation on the preceding two observations.
Hidden Markov Model (HMM) is another probabilistic model used for modeling processes varying in time. HMMs are widely used for sequence analysis because of their ability to identify hidden states, corresponding to clusters of observations. Mathematically, HMM can be defined as λ = (A, B, π), where:
A is an N × N state transition probability distribution matrix A = {aij}
B is an N × M matrix B = {bj(k)} with observation symbol probability distribution for each state
π is the initial state distribution vector π = {πi}
Hence, N is a number of hidden states in the model and M is a number of distinct observations per hidden state, i.e. the discrete vocabulary size. The key difference between HMM and MC is that HMM requires specifying the number of hidden states as a model parameter. HMM deduces a sequence of hidden states that best explains the observations along with the state transition probabilities and the distributions of observations (emission probabilities) per each hidden state. The Baum-Welch algorithm7 is used to estimate the parameters of HMMs for successful and unsuccessful interviews using the corresponding training set, while the Viterbi algorithm7 is used to determine the most likely sequence of hidden states for a given sequence of observations. After assignment of hidden states, the log-likelihood of success for an interview can be estimated using Eq. 3 as well.
Behavior code embeddings. Representation of behavior codes was inspired by the recent success of word embeddings28– 30. Embedding is a representation of an object in low-dimensional space using a real-valued vector. In our study, embeddings of behavior codes were obtained as a by-product of training LSTM and GRU after feeding one-hot vectors as a representation of behavior codes as input to these RNNs. Behavior code embeddings have the property of representing similar codes with the vectors that are close to each other in low-dimensional space. Figure 1 illustrates the MYSCOPE code embeddings visualized in 2-dimensional space by t-SNE31. It can be seen that positive behavior codes such as OQECHT+, OQECML+, AF, AFL, SUP, RCML+S, CQECML+, etc. formed a cluster in the left part of Figure 1. The nearest neighbors of CQECML+ are highlighted by different color intensity (i.e. OQECML+ being more purple indicates that it is more similar to CQECML+). The right part of the figure demonstrates another cluster formed with negative behavior codes including CQECML-, AMB-, RCHT-C, OQECHT-, GINFO-, RBAC, LUP-, RCHT-S, RPTBC, RAMBC, AMB-, RCML-S, etc. It is interesting that the behaviors intended to elicit CHT+/CML+ group together, whereas the ones intended to elicit CHT-/CML-also group together and are located in the opposite regions of semantic space.
Figure 1:

2-D representation of behavior code embeddings.
Recurrent Neural Networks (RNN) are a class of neural networks that have an internal memory, which makes them particularly suitable for processing sequences of observations. The ability of RNNs to capture long-term dependencies and remember past observations for predicting future observations is their main advantage over MCs and HMMs. These features are very useful in the analysis of motivational interviews, in which any behavior observed at a particular point in the interview may be indicative of other behaviors that are observed later. In order to mitigate the vanishing gradient problem of earlier versions of RNN10, Hochreiter et al.9 proposed Long Short Term Memory networks (LSTM). There are several variants of LSTM model, among which the most notable one is the Gated Recurrent Unit32 (GRU). GRUs are simpler than LSTMs and have been shown to be effective for a variety of Natural Language Processing tasks32. GRU is formally defined as follows:
| (4) |
| (5) |
| (6) |
| (7) |
In Eq. 4-7, σ corresponds to sigmoid function and ⊙ designates an element-wise product. The update gate zt and reset gate rt at time step t are computed by the Eq. (4) and (5), where Wz, Wr, Wh, Uz, Ur, Uh are the weight matrices and bz, bh and br are bias vectors. The activation ht of the GRU at time t is a linear combination of the previous activation ht–1 and the candidate activation , which is represented by Eq. (7) and (6).
The RNN architecture employed for sequence classification is shown in Figure 2. As can be seen from Figure 2, softmax is used at each time step to predict the class of a sequence observed so far. Since the sequence label is predicted at each observation, the proposed architecture is referred to as Recurrent Neural Network with Target Replication (TR). It is trained by minimizing the following hybrid loss function:
| (8) |
Figure 2:

Proposed RNN model with target replication (TR).
As follows from Eq. 8, the total loss is a convex combination of the final loss L(ȳ(T), y(T)) and the average loss over all observations in a sequence, where T is the total number of observations, ȳ(t) is the output at step t, and α ∊ [0,1] is a hyperparameter controlling the relative importance of each loss type. We experimentally determined that the best performance is achieved when α = 0.5. Our model also contains several other hyperparameters, such as the number of embedding dimensions, the number of hidden units, learning rate, batch size, etc., which were optimized on the validation set. We implemented our models in Tensorflow with Adam optimizer as well as early stopping based on the validation loss and observed that our model converges after 100 epochs.
Evaluation metrics
Performance of probabilistic and deep learning methods* was evaluated in terms of precision, recall, and F-measure using 10 folds cross-validation and weighted macro-averaging of these metrics over the folds. However, LSTM and GRU were trained on 80% of the data and validated on 10%, with the remaining 10% of the data used for testing.
Results
All sequence classification methods were evaluated in the case of both under and over-sampling. Predictive performance summary of all methods is summarized in Table 2.
Table 2:
Performance of MC, HMM, LSTM and GRU with and without target replication (TR) for predicting the success of patient-provider communication sequences when under- and over-sampling were used to balance the dataset. The highest value for each performance metric is highlighted in bold.
| Method | Under-sampling | Over-sampling | ||||
|---|---|---|---|---|---|---|
| Precision | Recall | F1-Score | Precision | Recall | F1-Score | |
| Markov Chain 1st Order | 0.7060 | 0.7044 | 0.7038 | 0.7932 | 0.7799 | 0.7775 |
| Markov Chain 2nd Order | 0.6395 | 0.6385 | 0.6379 | 0.7111 | 0.7029 | 0.7000 |
| Hidden Markov Model | 0.6244 | 0.6143 | 0.6067 | 0.7775 | 0.7567 | 0.7520 |
| LSTM | 0.8672 | 0.8626 | 0.8622 | 0.8411 | 0.8372 | 0.8368 |
| LSTM-TR | 0.8733 | 0.8681 | 0.8677 | 0.8424 | 0.8385 | 0.8381 |
| GRU | 0.8674 | 0.8648 | 0.8646 | 0.8379 | 0.8342 | 0.8337 |
| GRU-TR | 0.8705 | 0.8676 | 0.8673 | 0.8412 | 0.8377 | 0.8373 |
Predictive performance in the case of under-sampling
We used a small learning rate of 0.00005 and the batch size of 8 along with early stopping strategy for training deep learning models on the dataset balanced with under-sampling. Five major conclusions can be drawn from the results in Table 2. First, recurrent neural networks outperform probabilistic models and achieve 16.39%-26.1% higher F1-score. Second, LSTM with target replication has the best performance over all other RNN methods, and achieved F1-score 0.8677 with precision 0.8733 and recall 0.8681. Third, target replication strategy improves the performance of GRU and LSTM, with conventional GRU showing better performance than traditional LSTM. Fourth, among probabilistic models, the MC based method generally outperforms HMM across all metrics for under-sampled sequences. Fifth, second-order MC has lower precision, recall, and F-measure than first-order MC. In particular, precision, recall and F-measure decrease by 9.42%, 9.36% and 9.36%, when going from first to second-order MC model.
Predictive performance in the case of over-sampling
Similar to the under-sampling scenario, early stopping strategy was also employed for training deep learning models on the dataset balanced with over-sampling. However, in this case, RNN models were trained with the learning rate of 0.00010 and the batch size of 55. Experimental results indicate that HMM had better performance than second-order MC, achieving 9.34%, 7.65%, and 7.43% higher precision, recall, and F-measure, while HMM still had 1.98%, 2.97%, and 3.28% lower precision, recall, and F-measure than first-order MC. Also similar to the under-sampling scenario, target replication improves the performance of RNN models and LSTM with target replication has the highest F1-score among all models. However, the predictive performance of LSTM and RNN decreases when over-sampling is used, while the performance of probabilistic models increases.
Most likely communication sequences
Table 3 provides examples of typical patient-provider communication sequences that frequently appear in successful and unsuccessful motivational interviews. We observed that in successful motivational interviews information is frequently provided using patient-centered communication (GINFO+) and structure session (SS) utterances, in which the counselor either explains the therapeutic agenda or attempts to transition to a new topic or session content. Sometimes, counselors also acknowledge the clients’ communication or an off topic comment (SO). We also observed that affirmations (AF) and open-ended questions (OQECML+) have a strong effect on eliciting positive change talk or commitment language, which is consistent with MI theory. It can also be seen that providing advice using non-patient centered strategies (ADV-) leads to negative ambivalence (AMB-), which results in the interview heading in therapeutically wrong direction. Questions posed to elicit negative change talk or commitment language lead to CHT-, CML- or AMB-, which is consistent with the manual analysis by clinicians.
Table 3:
Most likely communication sequences in successful and unsuccessful motivational interviews.
| Type | Most likely communication sequences |
|---|---|
| successful | GINFO+: General information, positive → LUP+: Low uptake, positive → OQTBN: Open-ended question, target behavior neutral |
| successful | SS: Structure session → GINFO+: General information, positive → CQECHT+: Closed-ended question, elicit change talk positive |
| successful | SO: Statement, other → LUP+: Low uptake, positive → AF: Affirm → HUPW: High uptake, weight → OQECML+: Open-ended question, elicit commitment language positive. |
| unsuccessful | ADV+: Advise, positive → AMB-: Ambivalence negative → OQECHT-: Open-ended question, elicit change talk negative |
| unsuccessful | CQECHT+: Open-ended question, elicit change talk positive → RCHT-S: Reflect, change talk negative → OQECHT-: Open-ended question, elicit change talk negative |
| unsuccessful | SUP: Support → AF: Affirm → CQTBN: Closed-ended question, target behavior neutral → OQECHT-: Open-ended question, elicit change talk negative → AMB-: Ambivalence negative |
Discussion
By analyzing the experimental results of different communication sequence outcome prediction methods proposed in this paper, we arrived at the following conclusions. First, the overall predictive performance of RNN based methods is substantially higher than that of probabilistic models. In particular, the RNN-based methods achieve near-human accuracy for predicting the success of motivational interviews. This indicates that RNN is able to capture the structure of discourse in motivational interviews by preserving long-term dependencies among the behavior codes, which reflect the overall progression of the interviews. This provides evidence that RNNs are able to successfully replicate human cognitive processes to integrate previous information when making decisions. In addition to that, embeddings allow to reduce the dimensionality of codes in PPC sequences and consequently improve both precision and recall of prediction.
Second, using target replication to compute the loss at each time step results in better performance for all configurations of the proposed RNN-based methods. This indicates that the average of the losses over all steps emphasizes the dependencies between the pairs of patient and provider codes, which results in more accurate estimates of the model parameters. Better estimates of parameters in RNN models of motivational interviews are propagated to the next step based on the relative importance of intermediate output, where they are aggregated into predictions for the entire sequence. This allows to achieve an improvement in prediction accuracy.
Third, using first-order Markov model results in better prediction accuracy compared to higher-order Markov models, which we attribute to the fact that the number of states in higher-order Markov models may grow exponentially with their order. As a result, accurate estimation of transition probabilities requires much larger training data. Using smaller datasets, which is the case when under-sampling is employed, will result in a sparsity problem, when many transitions are either not observed in the training set at all or observed only a few times, leading to missing or potentially inaccurate probability estimates. Obtaining large training sets cannot be easily accomplished in many domains, including motivational interviewing. In this study, we found out that using first-order Markov models is a reasonable trade-off between efficiency and accuracy.
Fourth, similar to traditional Markov model, HMM achieves a dramatic improvement in the prediction accuracy when larger training set is used. This indicates that sufficient training data is required to find the optimal settings of hyperparameters, such as the number of hidden states, initial state distribution, transition probabilities, and emission probabilities.
Fifth, the proposed method can be used to identify the most effective communication strategies for eliciting a particular type of behavioral response. Awareness of these strategies by researchers can significantly decrease the time and effort required to develop effective interventions to address many public health conditions, such as childhood obesity, and tailor these interventions to particular patient cohorts. Awareness of these strategies by the counselors can lead to a greater success rate of motivational interviews.
Conclusion
In this paper, we compared the accuracy of Recurrent Neural Networks with Markov Chain and Hidden Markov Model for the task of predicting the success of motivational interviews. We found out that individual PPC exchanges are highly indicative of the overall progression and future trajectory of clinical interviews and can be used to predict their overall success. Our proposed methods can facilitate motivational interviewing researchers in establishing causal relationships between different communication strategies and the desired behavioral outcomes during the interviews without resource-intensive manual qualitative analysis of interview transcripts, which can significantly decrease the time and effort required to develop behavioral interventions. Our proposed methods can also help to identify the most likely sequences in successful and unsuccessful motivational interviews, which can directly inform clinical practice and increase the effectiveness of behavioral interventions. Our experimental results also indicate that the proposed methods can be used for real-time monitoring of the progression of clinical interviews. This work also has broad implications for public health research by providing a theoretically-grounded computational approach to qualitative data analysis.
Acknowledgments
This study was supported by a grant from the National Institutes of Health, NIDDK R21DK108071, Carcone and Kotov, MPIs. We would like to thank the student assistants in the Department of Family Medicine and Public Health Sciences at Wayne State University School of Medicine for their manual annotation of experimental dataset with MYSCOPE codes.
Footnotes
Authors provided an equal contribution.
the code for all models is publicly available at https://github.com/teanalab/myscope-sequential-analysis
References
- [1].Yakhnenko O, Silvescu A, Honavar V. Discriminatively trained markov model for sequence classification. In: Proceedings of the 5th IEEE International Conference on Data Mining; IEEE. 2005. [Google Scholar]
- [2].Deshpande M, Karypis G. Evaluation of techniques for classifying biological sequences. In: Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining; Springer. 2002. pp. 417–431. [Google Scholar]
- [3].Wei L, Keogh E. Semi-supervised time series classification. In: Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; ACM. 2006. pp. 748–753. [Google Scholar]
- [4].Leslie C, Kuang R. Fast string kernels using inexact matching for protein sequences. Journal of Machine Learning Research. 2004;5(Nov):1435–1455. [Google Scholar]
- [5].Chuzhanova NA, Jones AJ, Margetts S. Feature selection for genetic sequence classification. Bioinformatics. 1998;14(2):139–143. doi: 10.1093/bioinformatics/14.2.139. [DOI] [PubMed] [Google Scholar]
- [6].Keogh EJ, Pazzani MJ. Scaling up dynamic time warping for datamining applications. In: Proceedings of the 6th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; ACM. 2000. pp. 285–289. [Google Scholar]
- [7].Rabiner LR. A tutorial on hidden Markov models and selected applications in speech recognition. Proceedings of the IEEE. 1989;77(2):257–286. [Google Scholar]
- [8].Lipton ZC, Berkowitz J, Elkan C. A critical review of recurrent neural networks for sequence learning. arXiv preprint arXiv:150600019. 2015 [Google Scholar]
- [9].Hochreiter S, Schmidhuber J. Long short-term memory. Neural Computation. 1997;9(8):1735–1780. doi: 10.1162/neco.1997.9.8.1735. [DOI] [PubMed] [Google Scholar]
- [10].Bengio Y, Frasconi P, Simard P. The problem of learning long-term dependencies in recurrent networks. In: Proceedings of the 1993 IEEE International Conference on Neural Networks; IEEE. 1993. pp. 1183–1188. [Google Scholar]
- [11].Graves A, Mohamed AR, Hinton G. Speech recognition with deep recurrent neural networks. In: Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing; IEEE. 2013. pp. 6645–6649. [Google Scholar]
- [12].Nion T, Menasri F, Louradour J, Sibade C, Retornaz T, Metaireau PY, et al. Handwritten information extraction from historical census documents. In: Proceedings of the 12th International Conference on Document Analysis and Recognition. 2013:822–826. [Google Scholar]
- [13].Lipton ZC, Kale DC, Elkan C, Wetzell R. Learning to diagnose with LSTM recurrent neural networks. arXiv preprint arXiv:151103677. 2015 [Google Scholar]
- [14].Choi E, Bahadori MT, Schuetz A, Stewart WF, Sun J. Doctor ai: Predicting clinical events via recurrent neural networks. In: Proceedings of Machine Learning for Healthcare Conference. 2016:301–318. [PMC free article] [PubMed] [Google Scholar]
- [15].Chung J, Gulcehre C, Cho K, Bengio Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:14123555. 2014 [Google Scholar]
- [16].Ogden CL, Carroll MD, Kit BK, Flegal KM. Prevalence of obesity and trends in body mass index among US children and adolescents, 1999-2010. Journal of American Medical Association. 2012;307(5):483–490. doi: 10.1001/jama.2012.40. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [17].General US. Washington, DC: HHS; 2010. Surgeon Generals vision for a healthy and fit nation. [Google Scholar]
- [18].Apodaca TR, Longabaugh R. Mechanisms of change in motivational interviewing: a review and preliminary evaluation of the evidence. Addiction. 2009;104(5):705–715. doi: 10.1111/j.1360-0443.2009.02527.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [19].Walker D, Stephens R, Rowland J, Roffman R. The influence of client behavior during motivational interviewing on marijuana treatment outcome. Addictive Behaviors. 2011;36(6):669–673. doi: 10.1016/j.addbeh.2011.01.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [20].Kotov A, Hasan M, Carcone A, Dong M, Naar-King S, BroganHartlieb K. Interpretable probabilistic latent variable models for automatic annotation of clinical text. In: Proceedings of the 2015 Annual Symposium of the American Medical Informatics Association. 2015:785–794. [PMC free article] [PubMed] [Google Scholar]
- [21].Hasan M, Kotov A, Carcone AI, Dong M, Naar S, Hartlieb KB. A study of the effectiveness of machine learning methods for classification of clinical interview fragments into a large number of categories. Journal of Biomedical Informatics. 2016;62:21–31. doi: 10.1016/j.jbi.2016.05.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [22].Carcone AI, Naar-King S, Brogan K, Albrecht T, Barton E, Foster T, et al. Provider communication behaviors that predict motivation to change in black adolescents with obesity. Journal of Developmental and Behavioral Pediatrics. 2013;34(8):599. doi: 10.1097/DBP.0b013e3182a67daf. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [23].Eide H, Quera V, Graugaard P, Finset A. Physician-patient dialogue surrounding patients expression of concern applying sequence analysis to RIAS. Social Science & Medicine. 2004;59(1):145–155. doi: 10.1016/j.socscimed.2003.10.011. [DOI] [PubMed] [Google Scholar]
- [24].Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP. SMOTE synthetic minority over-sampling technique. Journal of Artificial Intelligence Research. 2002;16:321–357. [Google Scholar]
- [25].Nguyen HM, Cooper EW, Kamei K. Borderline over-sampling for imbalanced data classification. International Journal of Knowledge Engineering and Soft Data Paradigms. 2011;3(1):4–21. [Google Scholar]
- [26].Mutsam N, Pernkopf F. Maximum margin hidden Markov models for sequence classification. Pattern Recognition Letters. 2016;77:14–20. [Google Scholar]
- [27].Won KJ, Prugel-Bennett A, Krogh A. Training HMM structure with genetic algorithm for biological sequence analysis. Bioinformatics. 2004;20(18):3613–3619. doi: 10.1093/bioinformatics/bth454. [DOI] [PubMed] [Google Scholar]
- [28].Bengio Y, Ducharme R, Vincent P, Jauvin C. A neural probabilistic language model. Journal of machine learning research. 2003;3(Feb):1137–1155. [Google Scholar]
- [29].Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems. 2013:3111–3119. [Google Scholar]
- [30].Pennington J, Socher R, Manning C. Glove Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. 2014:1532–1543. [Google Scholar]
- [31].Maaten Lvd, Hinton G. Visualizing data using t-SNE. Journal of Machine Learning Research. 2008;9(Nov):2579–2605. [Google Scholar]
- [32].Cho K, Van Merrienboer B, Bahdanau D, Bengio Y. On the properties of neural machine translation Encoderdecoder approaches. arXiv preprint arXiv14091259. 2014 [Google Scholar]
