Skip to main content
Chronic Stress logoLink to Chronic Stress
. 2018 Jan 10;2:2470547017747553. doi: 10.1177/2470547017747553

Data Science in the Research Domain Criteria Era: Relevance of Machine Learning to the Study of Stress Pathology, Recovery, and Resilience

Isaac R Galatzer-Levy 1,2,, Kelly V Ruggles 3, Zhe Chen 1,4
PMCID: PMC5841258  NIHMSID: NIHMS922085  PMID: 29527592

Abstract

Diverse environmental and biological systems interact to influence individual differences in response to environmental stress. Understanding the nature of these complex relationships can enhance the development of methods to (1) identify risk, (2) classify individuals as healthy or ill, (3) understand mechanisms of change, and (4) develop effective treatments. The Research Domain Criteria initiative provides a theoretical framework to understand health and illness as the product of multiple interrelated systems but does not provide a framework to characterize or statistically evaluate such complex relationships. Characterizing and statistically evaluating models that integrate multiple levels (e.g. synapses, genes, and environmental factors) as they relate to outcomes that are free from prior diagnostic benchmarks represent a challenge requiring new computational tools that are capable to capture complex relationships and identify clinically relevant populations. In the current review, we will summarize machine learning methods that can achieve these goals.

Keywords: machine learning, stress, Research Domain Criteria, computational psychiatry, stress pathology, data science, resilience

Introduction

The Research Domain Criteria (RDoC) conceptualizes mental health and mental illness as the result of multiple overlapping and interdependent dimensions.1,2 This framework provides significant opportunity for advances in research into stress psychopathology and stress resilience as the etiology of such responses are, by definition, due to interactions between diverse internal and external factors. Empirically, biological systems that relate to stress pathology such as HPA axis regulation,3,4 immune functioning,5,6 the renin–angiotensin system,7 the sympathetic–adrenal–medullary system,8 and circadian rhythms9,10 are known to have multiple overlapping components spanning from genes to neurocircuits.11 Further, these systems affect each other in complex, often multidirectional, ways across the central and peripheral nervous systems both in response to prior and current stress, daily demands, and internal rhythms.1119 Integrating information across these dimensions to make clinical decisions about an individual patient represents a significant challenge that may be necessary to overcome to advance therapeutics.

The RDoC initiative not only encourages a reconceptualization of the factors that impact health and psychopathology but also encourages a rethinking of the primary outcomes under study with explicit direction to move away from diagnostic classification.1,2 Stress can produce temporary or even permanent alterations in cognition,20 memory,21 arousal,22 sleep,9,10 mood,23 motor activity,24 and approach/avoidance behaviors.25 Examining such behaviors as the primary outcome makes sense as psychiatric diagnoses aggregate diverse presentations resulting in diagnoses that can encompass vast clinical presentations making them too heterogeneous to be useful as research tools.26

Characterizing health and pathology and uncovering mechanisms underlying these outcomes without the traditional mile markers of psychiatric diagnosis presents with a significant conceptual and computational challenge. The limited guidance that has been given as part of the RDoC initiative regarding computational methods is that “Most important, this framework needs to integrate many different levels of data to develop a new approach to classification based on pathophysiology and linked more precisely to interventions for a given individual.”2 Machine learning (ML) approaches are designed to achieve these goals.

ML methods can be cast into three general categories: (1) Unsupervised methods, which describe a class of algorithms that find relationships between variables without reference to a specific outcome. Unsupervised learning models provide information about how variables cluster together or relate to each other without an explicit outcome of interest. (2) Supervised models are designed to predict or classify an outcome of interest such as the presence or absence of a mental disorder. (3) Reinforcement learning (RL) examines how actions in one’s environment (such as treatment) alter behavioral states. These methods provide a powerful set of tools to examine mechanisms, predict risk, and develop treatment based on complex sources of information. In this review, we will focus on computational methods and examples from stress pathology research that attempt to achieve these goals. The goal of this review is to provide a broad overview of ML concepts and their relevance to stress pathology research in the RDoC era.

What Is Machine Learning?

ML refers to a large class of algorithms that attempt to learn patterns from data to improve performance and make predictions.27 Such algorithms recursively search for relationships in data by applying a set of logical rules and mathematical tools. Because such algorithms are powerful tools for identifying relationships between variables, ML methods are prone to overfitting or fitting a model that is specific to the data at hand but is not generalizable. For this reason, ML algorithms also integrate safeguards against overfitting.

There are many different algorithms that are designed to achieve the same general goals (i.e. supervised, unsupervised, and RL). No single algorithm works best in all contexts. Often data scientists will compare results from a number of different algorithms or select one based on specific needs. For example, ML approaches vary in their interpretability. In many nonscientific contexts, data analysts may be less concerned with interpretation compared to model building. A stockbroker attempting to predict if the Dow Jones will increase in the next quarter may fit a model to make a decision about the likely course of the market without much interest in the nature of the underlying relationships that lead the Dow to increase or decrease. However, an economist investigating the same question may be much more interested in underlying the factors that lead to the outcome. Methods such as support vector machines (introduced below) are powerful methods for predictive modeling but are known as a “black box” because the nature of the underlying relationships is not accessible. Conversely, methods such as graph models (also introduced below) are highly interpretable but their stability and accuracy for decision-making can be limited. As such, when choosing a modeling approach, data scientists often weigh their goals in terms of the need to interpret and the need to build a stable model.

A general strength of ML methods is their ability to integrate larger sets of variables and capture complex dependencies between variables. ML methods can model dependencies between variables using Boolean logic (AND, OR, NOT), absolute conditionality (IF, THEN, ELSE), and conditional probabilities (probability of X given Y). Such an approach allows models to capture multiple dependent relationships and, as such, have increased relevance to real-world scenarios where multiple factors are in play. In the context of stress pathology such as posttraumatic stress disorder (PTSD), for example, multiple risk factors have been identified but none robustly predict risk alone.29 This may indicate that multiple factors work together and/or risk factors vary between individuals.

Female gender is a case in point as it has consistently been replicated as a risk factor for PTSD but only accounts for a small percentage of variance and is only relevant to some who develop the disorder.30 Recent findings in endocrinology, genetics, and epigenetics help to explain why female gender increases risk as the role of estrogen signaling in HPA axis regulation has come into focus,3134 indicating that risk associated with female gender may be nested in underlying biological functions related to estrogen signaling. Indeed, women have been shown to vary in their risk for PTSD depending on when in their cycle they experience a traumatic event.35 Finally, the different causes of stress-related pathology may not be reducible to biological explanations alone. Early environment has been shown to permanently alter HPA axis functioning.36 Like many biological systems, these dependencies are fundamentally nonlinear,37 creating a need to characterize complex nonlinear relationships. ML methods can be utilized to build models based on such complex environmental and biological dependencies to make predictions about risk in future cases.

Bayesian Estimation

The backbone of traditional statistical theory and associated statistical tests is the goal of null hypothesis testing which tests P(D|H0) meaning the probability of the data given the assumption that the null hypothesis is true or that the assumption is that there are no relationships between the variables in the model.38 Null hypothesis testing is embedded in statistical theory as a safeguard against a priori assumptions about the nature of populations under study or their relationships to covariates.39 However, a consequence of this level of rigor is that researchers cannot use prior research to make estimates.

While this may seem like an esoteric statistical issue, it has real-life consequences for a researcher’s ability to develop methods for mechanism identification, prediction, and individualized treatment.40 In the context of a treatment study, for example, the null hypothesis is that the treatment has no effect greater than chance. This rigorous assumption is useful when examining a novel treatment. But when a treatment has demonstrated a consistent but moderate effect, such as exposure therapy for phobias and PTSD, researchers may turn their attention from the question of if exposure therapy has an effect to research aimed at determining for whom does the treatment have an effect. The latter research question is outside of the realm of null hypothesis testing because such models make assumptions based on previous data that the treatment is effective in some cases and that success is dependent on some factors. To make a decision about the probability of successful treatment given, some individual characteristics require assumptions about future events given past information. Such questions can be mathematically formulated using Bayes theorem.

Bayes’ theorem states that P(A|B)=P(B|A)P(A)P(B). This simple formula provides a method to estimate the posterior probability of one occurrence given another. For example, prior research has demonstrated that the BDNF val66/met polymorphism is associated with recovery in exposure therapy.41 These findings can be translated using Bayes theorem to calculate the probability of recovery given that the patient is a Val66/met carrier as

P(Recovery|Val66/met)=P(Val66/met|Recovery)P(Recovery)P(Vol66/met)

An additional benefit of Bayesian estimation is that it greatly simplifies estimation, allowing for the integration of more variables with fewer subjects.42 To illustrate this, imagine that you sit down to watch TV when you realize you have lost the remote. A null hypothesis test would use frequentist methods such as maximum likelihood estimation43 that make no assumptions about where the remote control might be. Following this school of thought, you would sample, or look any physical space, that the remote could fit in. This rigorous approach would assume an equal probability that the remote control is in the oven or under the bed as it is to be jammed in the sofa cushion. Acting as a Bayesian, you would use prior knowledge to estimate a distribution related to the probability of the remote’s position. You may start in the three places the remote is most often and then radiate out to less probabilistic locations (e.g. the oven). This conception translated to research allows for much less sampling and computational effort to test the same hypotheses. Further, it allows for increased model complexity because researchers can state and test complex dependencies. For example, the distribution of locations of the remote control may change depending on who was last watching TV, and as such, you may make a different estimate that is informed by the probabilistic location given a particular individual.

Returning to exposure therapy, it is unlikely that the BDNF val66/met polymorphism alone will predict treatment success with high enough accuracy to make a treatment decision. However, researchers may improve prediction by integrating other relevant predictors that relate to the probability of treatment success. These predictors may be independent meaning that the probabilistic information they provide is independent of the genetic effect. Predictors can also be dependent in a manner that together provides as more accurate picture of the genetic risk. For example, the BDNF val66/met polymorphism is likely to affect the probability of recovery in exposure therapy because of its effect on BDNF protein synthesis, a growth factor involved in neurogenesis that affects learning and memory. The probabilistic estimate of recovery therefore may be enhanced by estimating the probability given the presence of val66/met, increases synthesis of BDNF, and increases in neurogenesis in relevant brain regions. Bayesian estimation provides a framework to build models based on prior experience (e.g. data) to make predictions about future cases.

Unsupervised Learning

Unsupervised learning refers to a class of algorithms that attempt to draw inferences about the relationship between variables in the absence of an outcome of interest.28,44 For example, a researcher may want to determine physiological channels that cluster together in response to a stressor or regions of the brain that are coactivated to characterize brain circuits. Researchers may also want to define populations based on such clusters45 rather than relying on a priori definitions such as diagnostic status. This is of particular relevance in the RDoC era which does not rely on traditional psychiatric classification methods to define health and illness. Finally, unsupervised methods are also of value for data reduction.46 Data reduction methods allow data scientists to filter down from a very large set of variables. Such an approach is useful when working, for example, with genetic and epigenetic data where the variable count can be in the millions.47

Feature Selection and Feature Extraction

One common use of unsupervised learning method for data reduction is to reduce the dimensionality of a set of variables (or features) by removing redundant or irrelevant variables48 or by combining variables into composite values.49 Commonly in social and biological sciences, researchers are confronted with situations where a large number of variables may be of theoretical interest, but empirically, they are largely overlapping in the information that they provide. For example, cortisol and corticotropin-releasing hormone (CRH) are causally related to each other as CRH stimulates the production of cortisol.50 However, they may correlate to such a high degree that the information they provide is largely overlapping, or redundant. A researcher may want to down-select to reduce the number of variables in the model to guard against overfitting due to the curse of dimensionality whereby models become increasingly accurate in differentiating populations as the number of variables in the model increases.51 Similarly, these two markers (CRH and cortisol) may cluster together while an additional marker, such as glucose, may not make glucose irrelevant as it is unrelated to the larger cluster of variables. Similarly, the researcher may want to down-select irrelevant variables to reduce dimensionality and ultimately reduce the changes of overfitting the model.

Feature selection is distinguished from another commonly used unsupervised method, feature extraction.52 In this context, new, more stable, variables are created by combining variables or extracting the shared variance between variables. Returning to the example of physiological data measured in response to stress, a researcher may want to derive a single variable that represents the relationship between physiological measures. This can reduce the number of variables in a model and can also add stability in measurement. In this instance, researchers may use methods such as principle components analysis (PCA),49 which captures the shared variance between multiple variables which can ultimately be utilized as a variable in future analyses.

We provide a simple, illustrative example whereby a researcher wants to determine crime in his research subject’s neighborhood to use as a proxy measure for stress and danger in the subject’s environment. To achieve this, the researcher downloads crime statistics based on subject’s zip code, yielding multiple crime statistics including petty larceny, murder, rape, misdemeanor sex crimes among many others. This set is too large to analyze on its own, and further, any particular variable may not be very informative. As Figure 1 demonstrates, PCA can reduce dimensionality significantly to extract high variance components. In this case, two components were extracted that approximate violent crimes (i.e. assault, robbery, shootings, rape, and murder) and nonviolent crimes (i.e. misdemeanor sexual assault, loitering, and grand larceny). By reducing dimensionality in this way, researchers can then study a smaller set of variables that relate to broader constructs.

Figure 1.

Figure 1.

Principle components analysis (PCA) of census crime statistics. The figure demonstrates a PCA of census data. Crime statistics demonstrated to primary principle components or sets of shared variance. Component 1, which primarily comprised variance from violent crimes, accounted for approximately 66.9% of the total variance while Component 2, which primarily comprised variance from nonviolent crimes, accounted for 21.5% of the total variance.

Population Clustering

Increasingly, researchers are interested in identifying populations empirically rather than relying on a priori definitions. To achieve this, researchers often attempt to identify individuals who cluster together into clinically relevant populations. By identifying such populations, researchers can then test hypotheses about them. This approach is particularly relevant in the RDoC era where researchers are discouraged from using diagnoses to define populations.

There are many methods to cluster populations. One commonly utilized approach is to identify latent or not directly observable populations by identifying underlying mixture distributions (i.e. mixture modeling).53 For example, Figure 2 demonstrates a bimodal observed distribution with two underlying latent normal distributions. This is refered to as a mixture distribution.45 Returning to the example of measurements of physiological arousal in response to a stressor, these distributions may capture low-arousal and high-arousal individuals. These populations may be of clinical relevence and can now be examined as an outcome in lew of diagnoses.

Figure 2.

Figure 2.

Example of a two-mixture distribution. In this example, two latent (unobserved) distributions that are overlapping (mixture distributions) and that are both Gaussian normal (red and green) are identified underlying an observed nonnormal distribution (grey).

The general prinicple of mixture modeling can be extended to longidinal data to examine change over time. This approach is relevent when researchers hypothesize that populations are differentiated not only by their level of severity but also change. Returing to the example of physiological stress response data, researchers may be interested to know if there are distinct populations based on the ability to habituate to loud tones or to aquire and extinguish associations between conditioned and unconditioned stimuli as these both are models that are hypothesized to underly diverse stress pathologies.

Latent growth mixture modeling (LGMM) is one such approach that is commonly used in stress pathology research.54 This approach utilizes repeated measures to estimate a set of latent variables that indicate general levels on a particular variable (intercept parameter) and change across measurement occations (e.g. slope and quadratic parameters). From these variables, LGMM attempts to identify a second-order latent variable (class) which defines populations based on their similarities in the intercept, slope, and quadratic parameters. Figure 3 provides an example of trajectories derived based on eyeblink startle in response to threat (fear) acquisition and extinction training.55 In this example, by first identifying distinct trajectories in acquisition and extinction learning, researchers were able to determine the relationship between individual’s trajectory during extinction learning and risk genes as well as clinical presentation.

Figure 3.

Figure 3.

Three class latent growth mixture modeling (LGMM) of fear conditioning and extinction learning. Binned observations of eyeblink startle response are examined in response to a blue square paired with an air blast to the larynx (acquisition) and in response to the blue square without the air blast (extinction). LGMM was utilized to test for the number of classes and their parameters of change (e.g. slope and quadratic parameters). Results demonstrate that individuals follow three distinct trajectories of acquisition and extinction learning. By identifying trajectories, researchers can further examine hypotheses about the identified populations. These trajectories were shown to be associated with genetic variance and hyperarousal PTSD symptomatology.

Graphical Models

A limitation of models that include complex dependencies across a large number of variables is that they are hard to interpret. Graphical models provide a framework to represent high-dimensional relationships in two-dimensional space to aid in interpretation and, in some instances, facilitate hypothesis testing.56 While the mathematical basis of such models may vary (most commonly between Bayesian networks and Markov random fields), effecting the number of variables that can be examined together as well as computational time,57 the underlying concepts are very similar. Researchers can derive the structure of multiple interrelated variables by algorithmically testing conditional dependencies between all variables in the model. For example, the set of variables (a, b, c, d) may all demonstrate a univariate relationship with x. By testing the relationship between a while conditioning on b, c, d and doing the same for b, then c, then d, the algorithm can determine which variable is directly connected to x (Figure 4(a)). By running over all variables, the algorithm can identify those variables that relate to x through other variables. By repeating this procedure, a large network of variables representing complex dependencies can be derived.58 While this is one example of how graphical models are derived algorthmically, it captures the general principles. Such models are increasingly utilized in stress pathology research to understand how multiple relevent dimensions relate to each other. A number of recent publications, for example, have examined how symptoms of pathology, including complicated grief,59,60 comorbid depression and obsessive compulsive symptoms,61 and PTSD62 relate causally.

Figure 4.

Figure 4.

Example of a graphical model. (a) The figure demonstrates a toy example whereby x is only directly connected to D. This indicates that x is independent of all other variables given D. Similarly, D is independent of A and B given C. However, A, B, and C may effect X through D. While this is one example of how granical models are derived algorthmically, it captures the general principles. If the example we provided was real data, the researcher may use the graph to derive hypotheses about how mechanisms relate or how best to treat a disorder. (b) The figure demonstrates an example with real data of the interrelationship between PTSD symtpoms among adult survivors of childhood sexual abuse. The thickness of lines represents the strength of the relationship while the color represents positive (green) and negative (red) relationships.

After graphical models are identified, they can be utilized for other purposes beyound simple description. First, graphical models can be used for feature selection as the set of variables that is directly connected to a variable of interest theoretically contains most of the probabalistic information about that variable.63 The set of directly connected variables can then be selected, and all other variables can be treated as redundant or irrelevent. Further, by modeling the structure between variables in a graph, researchers can conduct data experiments where they set the value of a particular variable to determine the downstream effects on other variables of interest.64 For example, a researcher who has derived a graphical model of a gene expression network may want to know if he altered the value of a particular target with a drug, would it alter the downstream expression patterns. The research could derive preliminary evidence by setting the value of that target to determine how it changes variables that are downstream of the target to develop hypotheses about the effect of the drug before collecting experimental data.

As an example, McNally et al.65 utilized Bayesian network models to determine how symptoms of PTSD interrelate among victims of childhood sexual abuse. By deriving a network of relationships (see Figure 4(b); published with permission from the authors), the authors demonstrate that symptoms of PTSD influence each other rather than simply clustering together. The authors demonstrate that specific symptoms play a more centralized role in the development and maintenance of the symptom constellation as a whole. This analysis provides simple descriptive information about how a large set of variables effect each other. Such analyses provide useful information as a clinician may consider interventions that address specific symptoms that are of central importance to alter the network of symptoms overall.

Supervised Learning

Imagine a scenario where a mental health researcher wants to determine what information (genetics and epigenetics, peripheral neuroendocrinology, clinical self-report, etc.) most accurately differentiates cases from control subjects. In many instances, the researcher may have evidence from the literature that these elements are related to the clinical outcome of interest but do not have an a priori hypothesis regarding which variables are important for such classification or how they interact to effect risk. Such a task is increasing in relevance as researchers attempt to build predictive or classification models for mental disorders.

Supervised ML is a class of data modeling methods that is concerned with the development of algorithms that can learn a function from data that optimally predicts a specified outcome.27,28 Just like traditional statistics, supervised models fall into two classes, classification models that attempt to predict a categorical outcome and regression models that attempt to predict a continuous outcome.

The goal of supervised ML methods is to build an accurate classification or regression model that can be used to make decisions about patients in the future (i.e. beyond the data at hand). Supervised models typically attempt to learn a function using the available variables that fits a set of cases where the label (traditionally what is thought of as the dependent variable) is known, referred to as the training set. The process of fitting the model, or testing different parameters and sets of variables, is much more liberal compared to traditional statistical approaches. Subsequently, this function is tested on cases where the label is hidden from the researcher to test the accuracy of the model, known as the testing set. If the function works roughly equivalently in the training and testing sets, then the model is thought to be well fit and the derived function may be trustworthy to make decisions about new cases. If the model fits significantly better in the training set, the model is thought to be overfit meaning that the function that was built was so specifically fits the training set that it has no generalizability and will not be likely to make accurate decisions in future cases. The process of training in a random subset of the data and then testing in another random subset is known as cross-validation.

In this section, we will discuss key benefits and limitations of supervised ML classification methods in the context of mental health research. While there are many algorithms that have been developed for such purposes, we will discuss three methods, Random Forests,66 Support Vector Machines (SVM),67 and Regularized Regression as key examples because of their popularity and accessibility in many software packages.

Generality of Supervised ML Algorithms

Although distinct algorithms utilize different approaches, generally supervised ML algorithms have the same goal. Given a set of training examples N[(x1,y1), … , (xN,yN)] where xi is a vector of variables for the ith case and yi is its class label (i.e. case or control), the learning algorithm’s goal is to identify a function g: X->Y in which X is the input space and Y is the output space. This function (g) is one element of a space of possible functions G, commonly known as the hypothesis space.

Classification Algorithms

Random Forests

Random forests66 are known as an ensemble learning classification method. In this approach, a multitude of decision trees are constructed during the training phase and then outputs the model class across individual decision trees. To better understand, we must define a decision tree. Decision trees are a predictive modeling approach. The term “tree” comes from the use of class labels (such as case and control) as leaves where the branches represent conjunctions of features that lead to the class labels. Classification trees are those where the goal of the analysis is to predict a discrete outcome such as depression caseness (present vs. absent), while regression trees are those where the outcome is a real number such as a depression score. This set of methods is often referred to as Classification and Regression Tree analysis. Decision trees operate by iteratively identifying variables that either account for the most variance in the outcome (in the case of continuous scores) or highest probability of differentiating categories (in the case of categorical outcomes). Figure 5 demonstrates a decision tree predicting PTSD symptoms based on multiple clinical, demographic, and environmental measures. As this example illustrates, decision trees provide useful information about cut scores for risk and risk based on multiple characteristics.

Figure 5.

Figure 5.

Decision tree example. The figure demonstrates an example of a decision tree predicting PTSD scores one month following emergency room (ER) admission as predicted by multiple rating scales in the ER (Subjective units of distress (SUDS) rating; Peritraumatic Dissociative Experiences Questionnaire (PDEQ); Immediate Stress Reactions Checklist (ISRC)), violence and nonviolent crime-based PCA-derived scores using census data, gender, and age. As the figure demonstrates, the average PTSD score (based on the PTSD Checklist 5 (PCL-5)) across the population is 27.16. Those with elevated PDEQ scores (≥32) have elevated PTSD scores (38.2) compared to those with PDEQ scores below 32 (24.03). Among those with low PDEQ scores, women are at reduced risk (20.75) compared to men (25.45). However, women exposed to higher levels of community crime have elevated PTSD scores (24.5) compared to those who are exposed to lower levels (18.5).

A significant limitation is that such methods can lead to overfitting, especially when the trees are “tall” where there are multiple extending branches. One way to prevent this is through the use of ensemble methods by repeatedly resampling from the data to build many trees (a forest) and then “vote,” or identify model branches across trees. A commonly utilized ensemble method is bootstrap aggregation (Bagging)68 in which the algorithm repeatedly selects random samples with replacement from the training set to fit trees and then averages predictions across all trees. This tends to reduce overfitting because each individual tree may be highly sensitive to noise in the training set while the model average across many trees is not, but only under the condition that individual trees are not highly correlated. Random forests extend this method by selecting random subsets of features to grow trees, often referred to as “feature bagging.” The purpose of this addition is to reduce correlations between trees.

Support Vector Machines

SVM classification algorithms attempt to build a classifier in multidimensional space (across many features or variables) that differentiates classes of individuals (e.g. cases vs. controls).69 SVMs achieve this by identifying a linear decision surface (e.g. a line in two-dimensional space) that separates classes with the largest distance (also called largest gap or margin) between objects that are at the borderline. While an infinite number of lines (or decision surfaces) can separate two classes, only one decision surface (support vector) exists that separates classes with the largest gap between borderline objects (see Figure 6). In this example, the support vector consists of two objects, one case and two controls (signified by yellow centers), together defining the line with the largest gap separating the two populations.

Figure 6.

Figure 6.

Linear decision surface with widest margin. SVMs attempt to identify a line with the largest gap that separates out predetermined populations.

There are many instances where there is no way to linearly separate objects belonging to two classes. When no linear decision surface can be identified, SVMs “map” the data into higher dimensional space, termed feature space, where a separating linear decision surface can be identified. This act of mapping to higher dimensional space to identify a linear surface is known as the kernel trick.70 The kernel trick extends SVMs beyond linear classification to nonlinear classification. This framework makes SVMs particularly useful in contexts where a data set has hundreds or thousands of dimensions such as genes or proteins (see Figure 7). SVMs thus construct a hyperplane (or set of hyperplanes) in high-dimensional space that can be used for classification (or regression in the case of Support Vector Regression71 where the score is a real number).

Figure 7.

Figure 7.

Features that are not linearly separable being pulled into high-dimensional feature space. SVMs and other ML methods employ a technique known at the kernel trick whereby a linear decision surface is identifiable in situations where populations are not linearly separable by pulling data into higher dimensional space.

Regularized Regression

Another commonly utilized set of methods for both model fitting and feature selection is regularized regression. In many situations, it is not appropriate to assume that variables will relate to each other in a linear fashion as linear regression does. Regularized regression techniques, such as the least absolute shrinkage and selection operator (LASSO), ridge, and elastic net regression, are useful in such a context because they allow the data analyst to select the preferred level of model complexity from linear to highly nonlinear. Increasing the complexity of the model can lead to overfitting as such models can find odd patterns that are unique only to the data at hand. As such, regularized regression models include a regularization term which imposes a penalty as models increase in complexity, making model fit harder to achieve as the complexity of the model increases. The coefficients of variables that are not relevant to the model are shrunk to decrease their impact on the model (in the case of some models such as LASSO and elastic nets, they are shrunk to 0). This allows analysts to use these models for feature selection as the most relevant variables will be selected into the model and irrelevant ones will be discarded. One of the key benefits of regularized regression is that the models are highly interpretable because it is evident what variables are predictive as well as the degree of predictive accuracy.

Model Building and Validation

Often when building a model, data scientists will integrate multiple techniques to find and validate the best solution. Figure 8 provides a schematic of an approach that integrates multiple techniques: Figure 8(1): Individuals are clustered into one of the three groups (chronic, recovery, and resilient), using LGMM. Figure 8(2): A diverse set of variables of different types such as physiology, labs, and self-report assessments are prepared for modeling. Figure 8(3): The large set of variables is entered into an unsupervised feature selection algorithm (in this case, network models are employed for feature selection). Figure 8(4): A model is built that classifies individuals based on the remaining variables into the three groups (chronic, recovery, and resilient) based on knowledge of who is a member of each group. Next, a random subset of the original data that were not used to build the model is used to test it. Data sources are compiled (Figure 8(a)) and entered into the model that was built during the training step (Figure 8(b)). Figure 8(c): Based on the model, individuals are classified into groups. Figure 8(d): The accuracy of the model in correctly selecting individual’s membership in each group is calculated. In an ideal scenario, this model is then tested on a truly independent data set. This approach has been utilized for the prediction of PTSD following exposure to a potentially traumatic event7275 and is a common approach in other areas of medicine.76

Figure 8.

Figure 8.

Machine learning classification workflow. The figure provides a schematic for a common approach to supervised ML prediction or classification. In this example, (1) we have individuals who are known to be part of one of the three populations (chronic, recovery, and resilient) along with (2) a set of variables of different types such as physiology, labs, and self-report assessments. (3) The large set of variables is entered into an unsupervised feature selection algorithm (in this case, network models are used for feature selection). (4) A model is built that classifies individuals based on the remaining variables into the three groups (chronic, recovery, and resilient) based on knowledge of who is a member of each group. This step is known as the training step, or model building. Next, the model is tested during the testing step. In this case, a random subset of the original data that was not used to build the model is used. (a) Data sources are compiled and (b) entered into the model that was built during the training step. (c) Based on the model, individuals are classified into groups. (d) The accuracy of the model in correctly selecting an individual’s membership in each group is calculated. In an ideal scenario, this model is then tested on a truly independent data set.

Reinforcement Learning

Dopamine (DA) is a neurotransmitter in the brain that initiates adrenalin during the activation of the stress response. DA rules motivational forces and psychomotor speed in the central nervous system. When a person is experiencing stress, the response system will be turned on, which will elevate stress hormones such as cortisol and reduce the level of serotonin and DA. Chronic stress or oversecretion of stress hormones may lead to imbalance of DA levels, and dysfunction of DA system (e.g. ventral tegmental area and nucleus accumbens) can potentially trigger various mental disorders, such as addiction, depression, distress, and anxiety. For instance, while high levels of DA cause drug “highs” or impulsivity (e.g. in addiction) and hyperactivity, low levels of DA may cause sluggishness and hypoactivity.

RL is an area of ML inspired by animal learning, behavioral psychology,77,78 and dynamic programming methods.79 In ML, it is also formulated as a Markov decision process. The RL method is developed to resolve a temporal credit assignment problem, which provides a framework for modeling reward/punishment-driven adaptive behavior80,81 and emotions.82 Specifically, a subject or agent will learn to optimize a strategy to maximize the payoff or future reward through trial and error. The strategy is determined by its own value function V(s). The temporal difference learning is the most common model-free RL algorithm,83 which aims to learn a value function V(s) for the state [s] (the state can be a finite or infinite set) according to the one-step ahead prediction error (PE).

V(st)V(st)+α[rt+1+γV(st+1)-V(st)]

where rt denotes the reward or reinforcement signal at time t (which can be positive or negative or zero), 0 ≤ γ ≤ 1 is a discount factor for the reward, and 0 < α < 1 is a learning rate parameter. A positive PE ≡ [rt+1 + γV(st+1) − V(st)] implies that the reward is greater than expected and therefore increasing the value V(st), whereas a negative PE leads to a decrease in V(st). Imagine an approach/avoidance task, the subject may need to learn a strategy (two actions: approach vs. avoid) depending on the cued stimulus in order to maximize the reward or avoid the punishment. In the simplest case, the value function can be represented as a look-up table, and each state–action pair is associated a value. Once the RL is accomplished, the agent will use the learned value function to guide behavior (i.e. action) in the environment (see Figure 9).

Figure 9.

Figure 9.

Reinforcement learning schematic. Reinforcement learning (RL) can be formulated as a Markov decision process of an agent interacting with the environment in order to maximize the future reward. At each time step t, given the current state st (and current reward rt), the agent needs to learn a strategy (i.e. the “value function”) that selects the optimal decision or action at. The action will have an impact on the environment that induces the next reward signal rt+1 (which can be positive, negative, or zero) and also produces the next state st+1. The RL continues with a trial-and-error process until it learns an optimal or suboptimal strategy.

Stress has played an important role in DA-related pathophysiology. For instance, the relationship between stress and drug abuse can be modeled by dopaminergic/corticosteroid interactions.84 In a pioneering RL application to psychiatric disorders, addiction has been modeled as RL gone awry.85 Specifically, the effect of addictive drug is to produce a positive PE independent of the change in value function, making it impossible for the agent to learn a value function that will cancel out the drug-induced increase in PE. Specifically, the PE is replaced by

PE=max[r(sk)+γV(sk)-V(sl)+D(sk),D(sk)]

where D(sk) indicates a DA surge occurring on entry into state sk. This modified equation will always produce a positive PE signal. Therefore, the values of states leading to a DA surge (D > 0) will approach infinity. While the DA neurons encode the PE, the addicted person’s brain will represent a specific “addicted” state sk (e.g. drug, sex, or casino gambling) with an unusually high value function V(sk).

More generally, different RL rules or rates can be adjusted according to either positive/negative reinforcement or positive/negative punishment to reflect the difference in rule sensitivity.

When confronting with aversive stimuli (stress factors), the agent can learn to inhibit a value function associated with the stress state sk. The PE may be modified as

PE=min[r(sk)+γV(sk)-V(sl)+D(sk),D(sk)]

where D(sk) indicates a DA reduction occurring on entry into state sk. In this case, D(sk) can be a small positive value or even a negative value. In the long term, V(sk) will converge to a negative value function independent of the positive or negative reinforcement.

Deep Learning

Deep learning is the application of multilayer (more than one hidden layer) artificial neural networks to learn complex representations of high-dimensional data patterns, such as images, videos, speech, and language.86 Deep learning may employ various network architectures, such as deep belief networks or recurrent neural networks. Learning algorithm can be supervised, semi-supervised, or unsupervised. Due to the large and deep network architecture, state-of-the-art optimization algorithms have been developed to tune the unknown high-dimensional (∼order of thousands or even tens of thousands) parameters.87 Research in the past decade has witnessed remarkable achievements in Artificial Intelligence(AI) in the era of BIG DATA.Due to powerful ability in representation and pattern discovery, we will expect a potential research application in computational psychiatry, where various heterogeneous sources of data (such as genes, behavior, family and medicine history, and neuroimaging) can be integrated within the ML framework to discover markers of risk and targets for treatment. The potential for deep learning will only be realized in mental health research as appropriate data sources become available. However, as large amounts of information on single individuals become available, the potential for discovery and characterization is enormous. Mental health researchers will soon be able to tap into massive sources of continuously recorded data that captures behavior in real time. Deep learning methods may quickly redefine behavioral constructs such as stress, ways of measuring them, and even the discovery of ways to manipulate such behavior for therapeutic purposes. A limitation of such models is that they are not straightforward to interpret and are very prone to overfitting.

Conclusion

ML-based methods provide a computational framework to conduct research in the RDoC era. These methods, with their ability to integrate multiple overlapping sources of data and define clinically relevant populations, have a great deal to offer stress pathology and stress resilience research. The promise of this nascent field will only be truly realized as sources of data become available that are of the size and scope to truly build and validate such complex models. Because of the power of these tools to find solutions, there is a heightened need for caution, rigor, and an understanding of the underlying principles and limitations of such approaches. We hope that this review has provided information about diverse methods in a manner that encourage researchers interested in stress pathology to begin to think about how they can instantiate computational models that match the complexity of their hypotheses.

Declaration of Conflicting Interests

The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Funding

The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: National Institute of Mental Health (K01MH102415), National Institute of Neurological Disorders and Stroke (R01-NS100065), and US National Science Foundation (IIS-1307645).

References

  • 1.Cuthbert BN, Insel TR. Toward the future of psychiatric diagnosis: the seven pillars of RDoC. BMC Med 2013; 11(1): 126. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Insel T, Cuthbert B, Garvey M, et al. Research domain criteria (RDoC): toward a new classification framework for research on mental disorders. Am J Psychiatr 2010; 167(7): 748–751. [DOI] [PubMed] [Google Scholar]
  • 3.De Kloet C, Vermetten E, Geuze E, Kavelaars AM, Heijnen CJ, Westenberg HG. Assessment of HPA-axis function in posttraumatic stress disorder: pharmacological and non-pharmacological challenge tests, a review. J Psychiatr Res 2006; 40(6): 550–567. [DOI] [PubMed] [Google Scholar]
  • 4.Daskalakis NP, Bagot RC, Parker KJ, Vinkers CH, de Kloet ER. The three-hit concept of vulnerability and resilience: toward understanding adaptation to early-life adversity outcome. Psychoneuroendocrinology 2013; 38(9): 1858–1873. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Segerstrom SC, Miller GE. Psychological stress and the human immune system: a meta-analytic study of 30 years of inquiry. Psychol Bull 2004; 130(4): 601–630. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Jones KA, Thomsen C. The role of the innate immune system in psychiatric disorders. Mol Cell Neurosci 2013; 53: 52–62. [DOI] [PubMed] [Google Scholar]
  • 7.De Kloet ER, Joëls M, Holsboer F. Stress and the brain: from adaptation to disease. Nat Rev Neurosci 2005; 6(6): 463–475. [DOI] [PubMed] [Google Scholar]
  • 8.Nash M, Galatzer-Levy I, Krystal JH, Duman R, Neumeister A. Neurocircuitry and neuroplasticity in PTSD. In: Friedman MJ, Keane TM, Resick PA. (eds). Handbook of PTSD: Science and Practice, 2nd ed New York, NY: Guilford Press, 2014, pp. 251–274. [Google Scholar]
  • 9.LeGates TA, Fernandez DC, Hattar S. Light as a central modulator of circadian rhythms, sleep and affect. Nat Rev Neurosci 2014; 15(7): 443. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Gamble KL, Berry R, Frank SJ, Young ME. Circadian clock control of endocrine factors. Nat Rev Endocrinol 2014; 10(8): 466–475. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.McEwen BS, Bowles NP, Gray JD, et al. Mechanisms of stress in the brain. Nat Neurosci 2015; 18(10): 1353. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Scheiermann C, Kunisaki Y, Frenette PS. Circadian control of the immune system. Nat Rev Immunol 2013; 13(3): 190. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Goldstein DS, McEwen B. Allostasis, homeostats, and the nature of stress. Stress 2002; 5(1): 55–58. [DOI] [PubMed] [Google Scholar]
  • 14.Karlamangla AS, Singer BH, McEwen BS, Rowe JW, Seeman TE. Allostatic load as a predictor of functional decline. MacArthur studies of successful aging. J Clin Epidemiol 2002; 55: 696–710. [DOI] [PubMed] [Google Scholar]
  • 15.McEwen B. Allostatis and allostatic load: implications for neuropsychopharmacology. Neuropsychopharmacology 2000; 22: 108–124. [DOI] [PubMed] [Google Scholar]
  • 16.McEwen BS. Neuroendocrine interactions. In: Bloom FE, Kupfer DJ. (eds). Psychopharmacology: The Fourth Generation of Progress, New York, NY: Raven Press, 1995, pp. 705–718. [Google Scholar]
  • 17.McEwen BS. From molecules to mind. Stress, individual differences, and the social environment. Ann N Y Acad Sci 2001; 935: 42–49. [PubMed] [Google Scholar]
  • 18.McEwen BS. Physiology and neurobiology of stress and adaptation: central role of the brain. Physiol Rev 2007; 87(3): 873–904. [DOI] [PubMed] [Google Scholar]
  • 19.McEwen BS. Sleep deprivation as a neurobiologic and physiologic stressor: allostasis and allostatic load. Metabolism 2006; 55(10 Suppl 2): S20–S23. [DOI] [PubMed] [Google Scholar]
  • 20.Arnsten AF. Stress weakens prefrontal networks: molecular insults to higher cognition. Nat Neurosci 2015; 18(10): 1376–1385. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Huang R-R, Hu W, Yin YY, Wang YC, Li WP, Li WZ. Chronic restraint stress promotes learning and memory impairment due to enhanced neuronal endoplasmic reticulum stress in the frontal cortex and hippocampus in male mice. Int J Mol Med. 2015; 35(2): 553–559. [DOI] [PubMed] [Google Scholar]
  • 22.Sanford LD, Suchecki D, Meerlo P. Stress, arousal, and sleep. In: Meerlo P, Benca RM, Abel T. (eds). Sleep, Neuronal Plasticity and Brain Function, New York, NY: Springer, 2014, pp. 379–410. [Google Scholar]
  • 23.Syed SA, Nemeroff CB. Early life stress, mood, and anxiety disorders [published online ahead of print April 10, 2017]. Chronic Stress. doi:10.1177/2470547017694461. [DOI] [PMC free article] [PubMed]
  • 24.Ensho T, Nakahara K, Suzuki Y, Murakami N, Neuropeptide S. Neuropeptide S increases motor activity and thermogenesis in the rat through sympathetic activation. Neuropeptides 2017; 65: 21–27. [DOI] [PubMed] [Google Scholar]
  • 25.Kaldewaij R, Koch SB, Volman I, Toni I, Roelofs K. On the control of social approach–avoidance behavior: neural and endocrine mechanisms. Curr Top Behav Neurosci 2017; 30: 275–293. [DOI] [PubMed] [Google Scholar]
  • 26.Galatzer-Levy IR, Bryant RA. 636,120 ways to have posttraumatic stress disorder. Perspect Psychol Sci 2013; 8(6): 651–662. [DOI] [PubMed] [Google Scholar]
  • 27.Mohri M, Rostamizadeh A, Talwalkar A. Foundations of Machine Learning, Cambridge, MA: MIT Press, 2012. [Google Scholar]
  • 28.Hastie T, Tibshirani R, Friedman J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction, Berlin, Germany: Springer, 2003. [Google Scholar]
  • 29.Bryant RA. Predictors of post-traumatic stress disorder following burns injury. Burns 1996; 22(2): 89–92. [DOI] [PubMed] [Google Scholar]
  • 30.Brewin CR, Andrews B, Valentine JD. Meta-analysis of risk factors for posttraumatic stress disorder in trauma-exposed adults. J Consult Clin Psychol 2000; 68(5): 748–766. [DOI] [PubMed] [Google Scholar]
  • 31.Ressler K. Mechanisms of estrogen-dependent regulation of ADCYAP1R1, a risk factor for PTSD. Psychoneuroendocrinology 2016; 71: 5. [Google Scholar]
  • 32.Glover EM, Jovanovic T, Norrholm SD. Estrogen and extinction of fear memories: implications for posttraumatic stress disorder treatment. Biol Psychiatr 2015; 78(3): 178–185. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Dias BG, Ressler KJ. PACAP and the PAC1 receptor in post-traumatic stress disorder. Neuropsychopharmacology 2013; 38(1): 245. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Ressler KJ, Mercer KB, Bradley B, et al. Post-traumatic stress disorder is associated with PACAP and the PAC1 receptor. Nature 2011; 470(7335): 492–497. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Pineles SL, Nillni YL, King MW, et al. Extinction retention and the menstrual cycle: different associations for women with posttraumatic stress disorder. J Abnorm Psychol 2016; 125(3): 349. [DOI] [PubMed] [Google Scholar]
  • 36.Sullivan R. Early life trauma with attachment produces later life neurobehavioral deficits but are paradoxically rescued by the odors paired with the early life trauma. In: Neuropsychopharmacology, vol. 39 (Suppl 1). London, England: Nature Publishing Group, 2014, pp. S65. [Google Scholar]
  • 37.Huys QJ, Maia TV, Frank MJ. Computational psychiatry as a bridge from neuroscience to clinical applications. Nat Neurosci 2016; 19(3): 404. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Gelman A, Carlin JB, Stern HS, Dunson DB, Dunson A, Rubin DB. Bayesian Data Analysis 2014; Vol. 2 Boca Raton, FL: CRC Press. [Google Scholar]
  • 39.Stigler SM. The History of Statistics: The Measurement of Uncertainty Before 1900, Cambridge, MA: Harvard University Press, 1986. [Google Scholar]
  • 40.Montague PR, Dolan RJ, Friston KJ, Dayan P. Computational psychiatry. Trends Cognit Sci 2012; 16(1): 72–80. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Felmingham KL, Dobson-Stone C, Schofield PR, Quirk GJ, Bryant RA. The brain-derived neurotrophic factor Val66Met polymorphism predicts response to exposure therapy in posttraumatic stress disorder. Biol Psychiatr 2013; 73(11): 1059–1063. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Bernardo JM, Smith AF. Bayesian Theory, Bristol, England: IOP Publishing, 2001. [Google Scholar]
  • 43.Scholz F. Maximum likelihood estimation. In: Kotz S, Johnson NL, Read CB. (eds). Encyclopedia of Statistical Sciences, Hoboken, NJ: John Wiley and Sons, 1985. [Google Scholar]
  • 44.Barlow HB. Unsupervised learning. Neural Comput 1989; 1(3): 295–311. [Google Scholar]
  • 45.Figueiredo MAT, Jain AK. Unsupervised learning of finite mixture models. IEEE Trans Pattern Anal Mach Intell 2002; 24(3): 381–396. [Google Scholar]
  • 46.Aliferis CF, Tsamardinos I, Statnikov A. HITON: A Novel Markov Blanket Algorithm for Optimal Variable Selection. American Medical Informatics Association, 2003. [PMC free article] [PubMed]
  • 47.Ray B, Henaff M, Ma S, et al. Information content and analysis methods for multi-modal high-throughput biomedical data. Sci Rep 2014; 4: 4411. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Dy JG, Brodley CE. Feature selection for unsupervised learning. J Mach Learn Res 2004; 5: 845–889. [Google Scholar]
  • 49.Dunteman GH. Principal Components Analysis, London, England: SAGE, 1989. [Google Scholar]
  • 50.Yehuda R. Post-traumatic stress disorder. N Engl J Med 2002; 346(2): 108–114. [DOI] [PubMed] [Google Scholar]
  • 51.Keogh E, Mueen A. Curse of dimensionality. In: Sammut C, Webb G. (eds). Encyclopedia of Machine Learning, Berlin, Germany: Springer, 2011, pp. 257–258. [Google Scholar]
  • 52.Guyon I, Elisseeff A. An introduction to feature extraction. In: Guyon I, Nikravesh M, Gunn S, Zadeh LA. (eds). Feature Extraction. Studies in Fuzziness and Soft Computing, Berlin, Germany: Springer, 2006, pp. 1–25. [Google Scholar]
  • 53.Muthen B. Latent variable mixture modeling. In: Marcoulides GA, Schumacker RE. (eds). New Developments and Techniques in Structural Equation Modeling, Mahwah, NJ: Lawrence Erlbaum, 2001, pp. 1–33. [Google Scholar]
  • 54.Muthen B. Latent variable analysis: growth mixture modeling and related techniques for longitudinal data. In: Kaplan D. (ed). Handbook of Quantitative Methodology for the Social Sciences, Newbury Park, CA: SAGE Publications, 2004, pp. 345–368. [Google Scholar]
  • 55.Galatzer-Levy IR, Andero R, Sawamura T, et al. A cross species study of heterogeneity in fear extinction learning in relation to FKBP5 variation and expression: implications for the acute treatment of posttraumatic stress disorder. Neuropharmacology 2017; 116: 188–195. [DOI] [PubMed] [Google Scholar]
  • 56.Koller D, Friedman N. Probabilistic Graphical Models: Principles and Techniques, Cambridge, MA: MIT Press, 2009. [Google Scholar]
  • 57.Jordan MI, Ghahramani Z, Jaakkola TS, Saul LK, et al. An introduction to variational methods for graphical models. Mach Learn 1999; 37(2): 183–233. [Google Scholar]
  • 58.Spirtes P, Glymour CN, Scheines R. Causation, Prediction, and Search, Cambridge, MA: MIT Press, 2000. [Google Scholar]
  • 59.Robinaugh DJ, LeBlanc NJ, Vuletich HA, McNally RJ. Network analysis of persistent complex bereavement disorder in conjugally bereaved adults. J Abnorm Psychol 2014; 123(3): 510. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Maccallum F, Malgaroli M, Bonanno GA. Networks of loss: relationships among symptoms of prolonged grief following spousal and parental loss. J Abnorm Psychol 2017; 126(5): 652. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.McNally R, Mair P, Mugno BL, Riemann BC. Co-morbid obsessive–compulsive disorder and depression: a Bayesian network approach. Psychol Med 2017; 47(7): 1204–1214. [DOI] [PubMed] [Google Scholar]
  • 62.McNally RJ, Robinaugh DJ, Wu GW, Wang L, Deserno MK, Borsboom D. Mental disorders as causal systems: a network approach to posttraumatic stress disorder. Clin Psychol Sci 2015; 3(6): 836–849. [Google Scholar]
  • 63.Aliferis CF, Statnikov A, Tsamardinos I, Mani S, Koutsoukos XD. Local causal and Markov blanket induction for causal discovery and feature selection for classification. Part I: algorithms and empirical evaluation. J Mach Learn Res 2010; 11: 171–234. [Google Scholar]
  • 64.Saxe GN, Statnikov A, Fenyo D, et al. A complex systems approach to causal discovery in psychiatry. PloS One 2016; 11(3): e0151174. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.McNally RJ, Heeren A, Robinaugh DJ. A Bayesian network analysis of posttraumatic stress disorder symptoms in adults reporting childhood sexual abuse. Eur J Psychotraumatol 2017; 8(3): 1341276. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Breiman L. Random forests. Mach Learn. 2001; 45(1): 5–32. [Google Scholar]
  • 67.Statnikov A, Aliferis CF, Hardin DP, Guyon I. A Gentle Introduction to Support Vector Machines in Biomedicine: Volume 1: Theory and Methods, River Edge, NJ: World Scientific Publishing Co. Inc, 2011. [Google Scholar]
  • 68.Breiman L. Bagging predictors. Mach Learn 1996; 24(2): 123–140. [Google Scholar]
  • 69.Cortes C, Vapnik V. Support vector machine. Mach Learn 1995; 20(3): 273–297. [Google Scholar]
  • 70.Schölkopf B, Smola AJ. Learning With Kernels: Support Vector Machines, Regularization, Optimization, and Beyond, Cambridge, MA: MIT Press, 2002. [Google Scholar]
  • 71.Basak D, Pal S, Patranabis DC. Support vector regression. Neural Inf Process Lett Rev 2007; 11(10): 203–224. [Google Scholar]
  • 72.Galatzer-Levy IR, Ma S, Statnikov A, Yehuda R, Shalev AY. Utilization of machine learning for prediction of posttraumatic stress: a re-examination of cortisol in the development of PTSD. Transl Psychiatr 2017; 7(3): e0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Galatzer-Levy IR, Karstoft KI, Statnikov A, Shalev AY. Quantitative forecasting of PTSD from early trauma responses: a machine learning application. J Psychiatr Res 2014; 59: 68–76. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Karstoft K-I, Galatzer-Levy IR, Statnikov A, Li Z, Shalev AY. Bridging a translational gap: using machine learning to improve the prediction of PTSD. BMC Psychiatr 2015; 15(1): 30. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75.Karstoft K-I, Statnikov A, Andersen SB, Madsen T, Galatzer-Levy IR. Early identification of posttraumatic stress following military deployment: application of machine learning methods to a prospective study of Danish soldiers. J Affect Disord 2015; 184: 170–175. [DOI] [PubMed] [Google Scholar]
  • 76.Zhao S, Dong X, Shen W, Ye Z, Xiang R. Machine learning-based classification of diffuse large B-cell lymphoma patients by eight gene expression profiles. Cancer Med 2016; 5(5): 837–852. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.Black AH, Prokasy WF. Classical Conditioning II: Current Research and Theory, New York, NY: Appleton-Century-Crofts, 1972. [Google Scholar]
  • 78.Sutton RS, Barto AG. Reinforcement Learning: An Introduction 1998; Vol. 1 Cambridge, MA: MIT Press. [Google Scholar]
  • 79.Bertsekas DP, Tsitsiklis JN. Neuro-dynamic programming: an overview. In: Proceedings of the 34th IEEE Conference on Decision and Control; December 13–15, 1995; New Orleans, LA.
  • 80.Montague PR, Dayan P, Sejnowski TJ. A framework for mesencephalic dopamine systems based on predictive Hebbian learning. J Neurosci 1996; 16(5): 1936–1947. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81.Schultz W, Dayan P, Montague PR. A neural substrate of prediction and reward. Science 1997; 275(5306): 1593–1599. [DOI] [PubMed] [Google Scholar]
  • 82.Broekens J, Jacobs E, Jonker CM. A reinforcement learning model of joy, distress, hope and fear. Connection Science 2015; 27(3): 215–233. [Google Scholar]
  • 83.Sutton RS. Learning to predict by the methods of temporal differences. Mach Learn 1988; 3(1): 9–44. [Google Scholar]
  • 84.Pani L, Porcella A, Gessa G. The role of stress in the pathophysiology of the dopaminergic system. Mol Psychiatr 2000; 5(1): 14. [DOI] [PubMed] [Google Scholar]
  • 85.Redish AD. Addiction as a computational process gone awry. Science 2004; 306(5703): 1944–1947. [DOI] [PubMed] [Google Scholar]
  • 86.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature 2015; 521(7553): 436–444. [DOI] [PubMed] [Google Scholar]
  • 87.Goodfellow I, Bengio Y, Courville A. Deep Learning, Cambridge, MA: MIT Press, 2016. [Google Scholar]

Articles from Chronic Stress are provided here courtesy of SAGE Publications

RESOURCES