Skip to main content
Journal of the American Medical Informatics Association : JAMIA logoLink to Journal of the American Medical Informatics Association : JAMIA
. 2013 Jul 25;20(e2):e267–e274. doi: 10.1136/amiajnl-2013-001865

Patient-tailored prioritization for a pediatric care decision support system through machine learning

Jeffrey G Klann 1,2,3, Vibha Anand 4,5, Stephen M Downs 4,5
PMCID: PMC3861915  PMID: 23886921

Abstract

Objective

Over 8 years, we have developed an innovative computer decision support system that improves appropriate delivery of pediatric screening and care. This system employs a guidelines evaluation engine using data from the electronic health record (EHR) and input from patients and caregivers. Because guideline recommendations typically exceed the scope of one visit, the engine uses a static prioritization scheme to select recommendations. Here we extend an earlier idea to create patient-tailored prioritization.

Materials and methods

We used Bayesian structure learning to build networks of association among previously collected data from our decision support system. Using area under the receiver-operating characteristic curve (AUC) as a measure of discriminability (a sine qua non for expected value calculations needed for prioritization), we performed a structural analysis of variables with high AUC on a test set. Our source data included 177 variables for 29 402 patients.

Results

The method produced a network model containing 78 screening questions and anticipatory guidance (107 variables total). Average AUC was 0.65, which is sufficient for prioritization depending on factors such as population prevalence. Structure analysis of seven highly predictive variables reveals both face-validity (related nodes are connected) and non-intuitive relationships.

Discussion

We demonstrate the ability of a Bayesian structure learning method to ‘phenotype the population’ seen in our primary care pediatric clinics. The resulting network can be used to produce patient-tailored posterior probabilities that can be used to prioritize content based on the patient's current circumstances.

Conclusions

This study demonstrates the feasibility of EHR-driven population phenotyping for patient-tailored prioritization of pediatric preventive care services.

Keywords: clinical decision support, pediatrics, data mining, Bayesian analysis, phenotype

Background and significance

Despite the acknowledged importance of preventive services and the abundance of authoritative guidelines for use in primary care,1–3 rates of delivery of preventive services remain suboptimal. This is especially true in pediatric primary care settings4 where there are competing demands on physician’s time for addressing acute and chronic issues and for providing anticipatory guidance for normal child growth and development. Furthermore, as the body of evidence-based recommended guidelines keeps growing, it becomes extremely difficult to determine which guidelines may apply to a particular patient and which are of highest priority5–8 within the time constraints of a typical office visit.

Computer decision support systems (CDSS) have been shown to improve rates of delivery of preventive services.9–11 Over the past 8 years, we have developed and deployed an innovative pediatric CDSS—the child health improvement through computer automation (CHICA) system—that uses a scannable paper interface—for use in routine care in our busy primary care pediatric clinics.12 CHICA integrates well into the high volume workflow of our practices by implementing age-appropriate screening of patients in the waiting room, and then combining this information with the patient’s electronic health record (EHR) to generate patient-specific reminders and recommendations for the physician. CHICA has been studied in a number of randomized trials and has been shown to improve rates of both delivery of preventive services as well as management of chronic conditions.13–16 By automating the process of screening and appropriately alerting the physician, CHICA has significantly decreased the burden of screening families and identifying relevant practice guidelines.17

To select which recommendations apply to a patient at a specific encounter, CHICA employs a global static prioritization scheme based on the expected value of the recommendation,18 with priorities assigned by experts based on the product of risk of adverse outcome, severity of the outcome, and the effectiveness of the physician's action to prevent adverse outcome. Selected guidelines result in patient screening questions and (if appropriate) physician prompts.14 18 While this method has yielded good results, its primary limitation is that it relies only on known population risks. It does not adapt to patient-specific risks that may be inferred from data about their changing circumstances. For example, a particular patient may have a higher or lower risk of a health condition or outcome, such as anemia or developmental delay, based on other data in the EHR. This would alter the expected value of screening for these conditions, thus altering the prioritization for that individual. We previously described an alternative, patient-tailored prioritization scheme using Bayesian networks,18 which prioritizes based on all of the available context. The premise of this approach is predicated on the notion that the patient's medical record contains data that can significantly alter or predict the posterior probabilities of variables of interest. In this paper, we continue this work by studying the predictive power of data in the EHR.

Bayesian networks place each variable (eg, clinical information or a screening question) into a directed acyclic graph, in which vertices represent variable nodes and edges represent probabilistic relationships among variables, as in figure 1. Conditional probability tables underlying each node define these relationships. Such networks allow the articulation of complex multivariate associations. To use the network, one can instantiate evidence variables in the network (variables with known values), which revises the probability of each target variable (the variables one desires to predict).

Figure 1.

Figure 1

A Bayesian network for asthma treatment.

A Bayesian network can model complex relationships, such as those between the patient condition and relevant preventive care. They are appealing because of their multivariate approach and because relationships are modularly represented as probability distributions that require no a priori hypotheses. Patient screening prioritization could then be performed by instantiating variables known about the patient’s current state (ie, current condition along with past screening and behavior) and selecting the targets with the highest expected value of information. This type of approach expands on approaches used in a variety of systems for medical diagnosis19 and treatment.20

One challenge to this approach is that it requires experts to articulate both the relationships among variables and the strength of those relationships. Not only is this time consuming and challenging,21 but the necessary data are often non-existent. Our previous work shows that relationships among some variables might be unknown even to the local experts.22 Therefore, in this paper we propose an approach that does not initially require any human input (although expert opinion can later be integrated). The approach uses a well-known, efficient method to discover a Bayesian network from an observational dataset.

Algorithms to perform this structure discovery have been developed relatively recently23 and have become computationally tractable even more recently.24 It is now possible to discover a Bayesian network of several hundred nodes from a relatively small dataset in a matter of hours on a desktop computer. While our previous work using association rule mining discovered meaningful patterns in a different pediatric dataset,22 this work extends this idea to create Bayesian networks that can be used to prioritize pediatric screening questions and interventions.

Objective

In this work, we describe a methodology for assessing individual patient risk of health threats or outcomes, using structure-discovery algorithms. We then evaluate this approach on a dataset of nearly 30 000 patients, analyzing both the discovered relationships and the predictive power of the derived model with the goal of determining its suitability for use in an expected value prioritization of relevant preventive care screening questions and interventions.

Materials and methods

Using data from previous encounters, we used a structure discovery algorithm to build a Bayesian network. We then evaluated the Bayesian network by instantiating the clinical variables with known values (called evidence) for a test set of patients and iteratively examining the posterior probability of each target value. This measures the network's ability to discriminate those patients with the highest probability of certain outcomes. This discriminative power is a sine qua non for the expected value calculations to prioritize questions and reminders. Here we describe the dataset preparation, model generation, and evaluation.

Dataset preparation

To build the model, we used observational data collected by CHICA during 2005–11 from 29 402 unique patients and 177 clinical variables that are recorded by CHICA as coded concept questions and answers. Approximately two-thirds of these patients are below 12 years of age and one-third are between 12 and 21 years of age. We produced a dataset appropriate for a structure-learning algorithm using structured query language.

The variables fell into five broad categories, shown in table 1. The vast majority of the coded concept questions were screening questions (eg, ‘Is there a smoker at home?’) or physician concerns (eg, concern about drug abuse). The remaining questions were as follows: 40 questions were exam and test results; 18 were anticipatory guidance (information on patient history or education—eg, have firearms been discussed?); two were demographic (preferred language and insurance status).

Table 1.

A breakdown of the 177 CHICA variables used in this study

CHICA variable category Count
Patient screening question or physician concern 117
Anticipatory guidance 18
Exam 31
Test result 9
Demographic 2

The majority were patient screening questions or physician concerns. Eighteen were ‘anticipatory guidance’—information on patient history and education. The remaining 42 were demographics, exams, and test results.

CHICA, child health improvement through computer automation.

Some variables were binary, but many had several possible categorical values, which usually included one normal value and several gradations of abnormal (eg, in response to ‘Do any household members smoke?’ possible abnormal answers included ‘relapse’, ‘yes, ready to quit’, and ‘yes, not ready to quit’). To increase the discriminative power of our statistical methods, a CHICA expert recoded each variable into a binary response.

Next, we extracted the most recent known value of each variable for each patient, resulting in a dataset of 29 402 rows and 177 columns, with three possible values: ‘true’, ‘false’, and ‘missing’. All the algorithms we describe below (with the exception of edge orientation) ignore missing values, so our methods are minimally biased toward unrecorded information. We randomly permuted the rows of the dataset and split the permuted data into a training and test set (2/3 and 1/3, respectively). The training set was used for model generation and the test set for model evaluation.

Model generation

We generated a Bayesian network using Java and the freely available Tetrad toolkit,25 in four steps.

First, we generated a ‘network skeleton’ from the training data using the max–min parents and children (MMPC) structure discovery algorithm,24 which is included in Tetrad. A ‘network skeleton’ is an undirected Bayesian network without underlying probabilities. Skeleton generation is becoming a common first step in modern Bayesian structure learning on large datasets.24 26 27 It typically uses tests of statistical association to discover structure. This has performance advantages over graph heuristic methods, and the discovered relationships also usually have a logical meaning to a human viewer. MMPC is one of the best among these skeleton discovery algorithms, partly because it can construct a model ‘faithful to the data’ at small sample sizes.24 28 This means that if the data have no inconsistencies, the underlying structure is always detected. Of course, no real observational data are without inconsistency, but MMPC's small sample size requirement makes it resilient to noisy data. MMPC's underlying statistical test is the G2 test, which is asymptotically equivalent to χ2 but has preferable behavior for structure learning at small sample sizes.27 This implementation of MMPC ignores missing values so that erroneous edges are avoided (eg, a correlation that occurs because edges are missing).

Second, to direct the graph, we implemented a simple greedy search to optimize a global heuristic (the BDeu statistic, also available in Tetrad), which estimates how well the graph explains the data. This follows the example of the max–min hill climbing algorithm,24 which builds on MMPC. The graph-heuristic approach is more robust than other approaches on noisy data. Tetrad's BDeu statistic cannot ignore missing values. This might have unfairly biased edge direction when many values were missing, but studies show that the predictive power of most Bayesian networks is fairly insensitive to non-structural errors.29 It does, however, mean that the resulting directionality might not reflect true causal direction, which must be considered in interpreting the actual graph structure.

Third, we estimated the probabilities in our model (again, with missing values ignored), using maximum likelihood estimation with Dirichlet hyperparameters initialized to α=1. This is a typical method, and it is equivalent to adding one case to each parameter, which can smooth out network parameters learned from noisy data.26

Finally, we exported the graph to XDSL format using Java. XDSL is the preferred format for the SMILE framework,30 a freely available toolkit for network inference.

Evaluation

To evaluate whether these networks could be used for dynamic prioritization, we used SMILE to compute the area under the receiver-operating characteristic curve (AUC) for all screening questions, concerns, and anticipatory guidance given our test set. We used a standard approach that relies on the Wilcoxon statistic's (W) equivalence to the AUC.31 The AUC measures the predictive power of the network. A higher AUC value corresponds to the probability that when an item is true, it is ranked higher than when it is not. To compute W, we performed the following steps: for each target variable, for each applicable test patient (ie, patients for whom this variable was not missing), we ‘instantiated’ that patient (setting values for all of the other known variables in the network for that patient) and measured the posterior probability of the target variable computed by the network. We used this information to compute W and the SE of W.31 Our set of target variables included all variables relevant to patient screening prioritization: the 135 variables in the first two rows of table 1.

Because a higher AUC represents greater ability to discriminate between true and false cases on a test set, it is a sine qua non for prioritization. If the network can accurately predict when a variable is true or false, then the posterior probability that the variable is true or false can be used in an expected value calculation to prioritize screening questions and prompts. The predictive power needed in an expected value calculation is dependent on several factors (see examples at the end of the Discussion section), so lesser predictive ability than is commonly needed may suffice for the proposed application. Furthermore, incorrect prioritization of screening questions is less catastrophic than incorrect prediction of disease. However, the system is unwieldy if it does not do much better than chance (AUC=0.5). We did not select a specific AUC threshold, but we analyzed portions of the network with AUC that were above average on our test set.

We also computed information about the graph structure using Gephi, an open-source graph visualization and manipulation tool.32 We calculated: the Markov blankets of target variables, the degree and betweenness centrality of problems and complaints, and the number of total subgraphs. The Markov blanket of a node is its parents, children, and siblings, and is frequently used as a heuristic for the set of most relevant variables in prediction.33 This was computed by using Gephi's small-world analysis plug-in to count all variables less than three hops away from each target variable and manually removing nodes from this set that were not in the Markov blanket. All other measures were implemented directly by Gephi. The degree of a variable is simply the number of edges connecting it to other variables in the graph. Degree is a proxy of the variable's overall relevance in predicting other variables (because the more connected it is, the more of a role it is likely to play). Betweenness centrality is a somewhat more powerful measure of the variables’ relevance than degree. It is the number of shortest paths on which a variable lies between pairs of other variables.34 Finally, the number of subgraphs are found by searching for components of the graph that have no edges touching other components.35 The graph visualizations in the figures were also created with Gephi.

This retrospective analysis was approved by the institutional review board.

Results

Graph structure

The final graph contained 107 variables (uncorrelated variables or those without predictive power are automatically removed by the algorithms described above). Seventy-eight variables were in the target set for prioritization (screening questions, physician concerns, and anticipatory guidance) and 10 non-target variables were in the Markov blanket of that target set. This filtered graph of 88 variables was made up of 13 subgraphs, of which 10 had three or fewer nodes.

The filtered graph can be seen in figure 2. Each variable is numbered; the number, name, degree, count, and AUC of each variable can be found in supplementary appendix A (available online only). The filtered graph is also included as a supplement in GEXF (graph exchange XML) format (available online only).

Figure 2.

Figure 2

Graph learned from a training set of approximately 30 000 patients, filtered to include only screening questions, anticipatory guidance, and their Markov blankets. It consists of 88 nodes (clinical variables) and 90 edges (relationships). Nodes are colored by subgraph and their size is determined by area under the receiver-operating characteristic curve (larger is higher). The three largest subgraphs are given a unique color; the smaller subgraphs are black. Non-target variables are yellow. Node numbers correspond to the node information table in supplementary appendix A (available online only).

Eight of 10 of the tiny subgraphs contained at least one node of AUC less than 0.60 (ie, not much more predictive than chance). The other two were: the year a home was renovated was predictive of the year it was built (AUC 0.65) and substance abuse in the family was predictive of engaging in high-risk activities under the influence of alcohol (AUC 0.60).

Summary statistics and analysis of most predictive nodes

Average AUC was 0.65 with an average SE of 0.020, and the 39 above-average (AUC ≥ 0.65) variables are listed, along with their SE, in table 2. High AUC tended to have smaller SE than those with low AUC (mean 0.028 vs 0.054). Table 3 shows the 10 non-target variables and their degree (the number of edges attaching that variable to the graph). Betweenness centrality is not shown, because the degree was highly correlated with it—only nodes of degree 2 and higher had a betweenness centrality score greater than zero.

Table 2.

The 39 target variables with above-average AUC (≥0.65, accounting for SE), alongside their SE

# Variable AUC SE
26 Has used tobacco 0.90 0.017
22 Drunk in the past month 0.83 0.036
38 Friends use drugs 0.83 0.027
34 Firearms are kept unloaded 0.80 0.010
54 Looks both ways when crossing street 0.80 0.018
86 Knows what to do in response to a smoke alarm 0.80 0.011
25 Has an escape plan for house fire 0.79 0.015
15 Uses cigarettes or snuff with friends 0.76 0.019
81 Has safety latches installed 0.76 0.011
94 Uses illegal drugs 0.76 0.034
55 Uses low-iron infant formula 0.74 0.039
87 Stops at the curb before crossing 0.74 0.016
51 Knows how to save a choking child 0.73 0.011
44 Has a smoke detector 0.72 0.017
84 Sleeps on side or back 0.72 0.016
98 Wears sports protective gear 0.72 0.017
45 Has stairway gates 0.71 0.012
3 Abuses over-the-counter drugs 0.71 0.047
20 Has driven with a drunk 0.70 0.036
35 Visits homes with firearms 0.70 0.012
12 Knowledgeable about burns 0.69 0.011
32 Firearms have been discussed 0.69 0.017
52 Knows how to swim 0.66 0.011
68 Play area is fenced 0.66 0.009
71 Has the poison control number on phone 0.66 0.008
76 Lives in a pre-1978 home undergoing renovation 0.66 0.049
96 Wears a bike helmet 0.66 0.008
10 Bicycle has coaster breaks 0.65 0.046
19 Doors are secure for child 0.65 0.031
42 Happy with how things are going 0.65 0.023
43 Has had intercourse 0.65 0.023
49 Home was built before 1950 0.65 0.055
79 Had felt sad the past few weeks 0.65 0.026
91 Has tested smoke detector batteries 0.65 0.011
97 Wears a life jacket 0.65 0.020
39 Has had fun in the past 2 weeks 0.64 0.027
69 Play equipment is protected 0.64 0.020
82 School suspension in the past year 0.64 0.019
106 Has seen a dentist 0.64 0.014

AUC is the probability a true instance is correctly ranked higher than a false instance. SE is a measure of the SD of the AUC. The # is the variable number in Figure 2 and supplementary appendix A (available online only).

AUC, area under the receiver-operating characteristic curve.

Table 3.

The 10 non-target variables in the Markov blanket of target variables, alongside their degree

Variable Degree
Standard 18-month developmental screening test 1
Vision screening 1
Normal oral exam 1
Sickle cell disease 1
Has asthma or symptoms of asthma 2
Family history: breastfed 2
Has medication allergies 1
Premature birth 1
English vs spanish speaker 11
ADHD diagnosis 1

Degree is the number of edges connected to this variable, and is related to that variable's importance in prediction.

ADHD, attention deficit hyperactivity disorder.

The top seven target variables had AUC of 0.8 or above. Figure 3 shows an extract of the graph containing these seven nodes and their Markov blankets. These graphs included almost the entirety of the two smaller subgraphs, and only a small portion of the largest one. The relationships between these top seven variables fell into two categories.

Figure 3.

Figure 3

The Markov blankets of high area under the receiver-operating characteristic curve (AUC) nodes from figure 2. Nodes are sized by AUC and the color matches figure 2. The Markov blanket of a node is its parents, children, and siblings, and is frequently used as a heuristic for the set of most relevant variables in prediction.24

Category 1: intuitively correct relationships

Variables that seemed intuitively correlated were reflected in the graph; for example, drug use is correlated to both alcohol use and drug use of friends. One entire subgraph is predominantly alcohol and drug use (see the blue subgraph in figures 2 and 3).

Category 2: less-intuitive relationships

Some unexpected correlations arose. For example, knowing what to do in response to a smoke alarm is partly predicted by behavior at a crosswalk. Similarly, whether firearms are kept unloaded is related to wearing a life jacket (see green and red subgraphs in figures 2 and 3). Correlations with lower AUC were also frequently interesting, such as the relationships between mood, drug use, and sex in adolescents (lower AUC means that the test set did not consistently support the discovered relationship, but its existence in the graph means the training set supported it.).

Discussion

We have employed a statistical method to learn a Bayesian network of pediatric preventive care variables, including screening questions and physician observations, assessments and actions, entirely automatically from previously collected aggregate data, creating a phenotype of the population. We then evaluated the predictive power of the network, laying the groundwork for an expected value approach to prioritize risk assessments and physician prompts based on the current state of the patient. The average AUC (0.65) of the 78 target variables modeled was the same as an earlier expert-derived Bayesian network for childhood asthma detection on CHICA data.36

The top seven of these (AUC≥0.8) fell into two categories. Category 1, teen high-risk behaviors, gives the system good face validity. These behaviors are known to be correlated.37 Category 2 highlights an advantage of a statistically based system: less well recognized relationships (including local correlations that might be true in only the target population) are automatically captured by the system. For example, correlations among safety behaviors have been described in our work with other systems.22

The ability to predict an individual patient's risk of a health state has myriad practical applications, but for the CHICA system, it supports an individualized approach to prioritizing risk assessments and services. The Bright Futures guidelines, the federally recognized authoritative source of preventive care guidelines for children, contain 574 pages of recommendations.3 In a single decade, American Academy of Pediatrics policies generated over 162 discrete, specific recommendations.8 This vastly exceeds what can be done in a typical visit. Without prioritization, the advice followed by pediatricians becomes arbitrary. When the probability of a patient having a particular health risk is known, it becomes possible to move it up or down in priority. For example, knowing a teen is likely to have used tobacco can increase the priority of smoking cessation counseling, and predicting that a family keeps loaded firearms in the home could raise the priority of discussing gun safety.

A demonstration of a full expected value approach to prioritizing questions and alerts in the CHICA system is beyond the scope of the present study because it requires the assessment of severity (utility) of the corresponding adverse outcome. Nonetheless, a couple of examples can illustrate the value of tailoring prioritization to the individual child. Our network offers an AUC of 0.9 for teen smoking. The posterior probability for an individual will vary, but this AUC can be thought of as including (approximately) the point 90% sensitivity and 90% specificity. The baseline rate of teen smoking in the USA is about 20%. This means that the network may distinguish teens with a 69% likelihood of smoking (who certainly would deserve screening and counseling) from those with a 3% likelihood of smoking whose visits might be better spent discussing school performance or asthma.

A less dramatic example might be lead risk (old housing stock); 3.4% of our patients have such risk.17 With an AUC of only 0.66, the network can still identify children with a risk of 6.4% from those with a risk of less than 2% who may not need screening at all.

Limitations and future directions

Like all machine-learning approaches, the model (network) can be negatively biased by the training set and can contain non-generalizable relationships (called over-fitting). The magnitude of this problem is reflected by the predictive power of the network. Ours was good on average but had weak areas—six variables were predicted no better than chance (AUC 0.5). Future work should improve: (1) the training data and (2) the use of those data. For (1), two problems are commonplace. One is that data in medical informatics are often noisy. This is actually unlikely here; our previous work has demonstrated that data collected using CHICA are reliable and meaningful.38 Another problem is that not enough context is provided, leading to transitive relationships (ie, there is a valid statistical relationship but there is a hidden variable, ie, not accounted for). Such transitive relationships frequently occur in EHR data mining.39 We can see this in our network: ‘sickle cell disease’ and ‘friends use drugs’ are connected, probably because both are more common in our African-American population. Although it is computationally intractable to learn networks of more than a few hundred variables, feature selection can be used on a broad dataset to choose relevant contextual variables for the structure-learning algorithm. Many feature selection algorithms exist.40 Although they are commonly quite slow, we are developing fast feature selection methods that use network analysis in correlation graphs.41 For (2), many of the non-target variables were connected to each other rather than to screening questions. It would be preferable to force exam and test results to be linked to screening questions, which can be accomplished by specifying a priori knowledge to the structure-learning algorithm before processing.

The networks in our system are time agnostic. It would be optimal to model not just what was last known to be true but when it was true. For example, if the last concern about drug use was 10 years ago, it is likely not to be a concern now. We mitigate this problem by using the most recent value for each patient; presumably if drug use was a concern many years ago, there would be a more recent instance in which the patient no longer said they used drugs. However, true temporal modeling should be compared to the standard time-agnostic model. In particular, the modeling of temporal patterns, such as increasing values of a parameter or recurrence of a risk factor, may impart greater predictive power. A variety of approaches supports network learning with time, including dynamic Bayesian networks,42 continuous time Bayesian networks,43 and a model we are developing.

Finally, as mentioned earlier, the full prioritization score we previously proposed includes the likelihood of a health risk, its severity (disutility), and the effectiveness of addressing it as a rational way of prioritizing assessment questions or physician reminders.18 By expanding our Bayesian networks into influence diagrams that include decision and value nodes, we can select risk questions and prompts based on their expected value. At that point, a clinical trial of the patient-tailored prioritized questionnaires will be appropriate.

Conclusion

This study demonstrates the feasibility of Bayesian structure learning and inference to build models for prioritizing pediatric screening questions and reminders in a patient-tailored manner, by using discriminative power as a sine qua non for future expected value calculations. These population-phenotype models are built automatically, using only previously collected data without human intervention. The outputs are easy-to-understand diagrams that can be edited by designers and researchers. This preliminary work used a dataset of 29 402 patients and 177 screening and preventive care variables to produce a dynamic model of 107 nodes and 143 edges. Structure analysis revealed groups of related nodes as well as some non-intuitive correlations (eg, firearms and life jackets). Average AUC on a test set was 0.65, with seven nodes above 0.8. Although we have outlined several future improvements to our methodology, we have also shown that even this average AUC can be discriminative enough to prioritize screening questions properly, depending on risk prevalence and the other factors in our previously developed expected value approach.

Acknowledgments

JGK would like to thank Peter Szolovits, Gunther Schadow and Siu Hui for helpng him to understand Bayesian networks, data analysis, and statistical evaluation. This work was performed at the Child Health Informatics and Research Development Lab, Wishard Health System, Indianapolis, IN and at the Massachusetts General Laboratory for Computer Science, Harvard Medical School, Boston, MA.

Footnotes

Contributors: JGK designed the study, executed the study, and wrote the majority of the manuscript. VA contributed her expertise in the CHICA system, provided the data for the study, and edited and contributed to the manuscript. SD devised the idea for the study and advised on its design and execution. He also edited and contributed to the manuscript.

Funding: This work was supported in part by grant 5T15 LM007117-14 from the National Library of Medicine and a Riley Children’s Hospital Foundation grant. Funding for CHICA comes from grants 1R01HS018453 and R01HS017939 from the Agency for Healthcare Research and Quality, and R01LM010031 from the National Library of Medicine.

Competing interests: None.

Ethics approval: This retrospective analysis was approved by the Indiana University Medical School institutional review board.

Provenance and peer review: Not commissioned; externally peer reviewed.

Data sharing statement: Details on the Bayesian network are included in the supplemental files of this manuscript. At this time we do not have permission to share additional data from this study.

References

  • 1.US Preventive Services Task Force guide to preventive services. Alexandria, VA: International Medical Publishing, 1996 [Google Scholar]
  • 2.Walker RD, Curry CES, Hammer LD, et al. Recommendations for preventive pediatric health care. Pediatrics 2007;2007:2901 [Google Scholar]
  • 3.Hagan JF, Shaw JS, Duncan PM. Bright futures: guidelines for health supervision of infants, children, and adolescents. Elk Grove Village, IL: American Academy of Pediatrics, 2008. https://www.pediatriccareonline.org/pco/ub/view/Bright-Futures/135001/all/Front+Matter (accessed 31 Mar 2013) [Google Scholar]
  • 4.Mangione-Smith R, DeCristofaro AH, Setodji CM, et al. The quality of ambulatory care delivered to children in the United States. N Engl J Med 2007;357:1515–23 [DOI] [PubMed] [Google Scholar]
  • 5.Coffield AB, Maciosek MV, McGinnis JM, et al. Priorities among recommended clinical preventive services. Am J Prev Med 2001;21:1–9 [DOI] [PubMed] [Google Scholar]
  • 6.Downs SM, Arbanas JM, Cohen LR. Computer-supported preventive services for children: the CHIP Project. Proceedings of the Annual Symposium on Computer Application in Medical Care; 1995:962 [Google Scholar]
  • 7.Hunt DL, Haynes RB, Hanna SE, et al. Effects of computer-based clinical decision support systems on physician performance and patient outcomes. JAMA 1998;280:1339–46 [DOI] [PubMed] [Google Scholar]
  • 8.Belamarich PF, Gandica R, Stein REK, et al. Drowning in a sea of advice: pediatricians and American Academy of Pediatrics policy statements. Pediatrics 2006;118:e964–78 [DOI] [PubMed] [Google Scholar]
  • 9.Dexter P, Perkins S, Overhage JM, et al. A computerized reminder system to increase the use of preventive care for hospitalized patients. N Engl J Med 2001;345:965–70 [DOI] [PubMed] [Google Scholar]
  • 10.Kaplan B. Evaluating informatics applications—clinical decision support systems literature review. Int J Med Inform 2001;64:15–37 [DOI] [PubMed] [Google Scholar]
  • 11.McDonald CJ. Protocol-based computer reminders, the quality of care and the non-perfectability of man. N Engl J Med 1976;295:1351–5 [DOI] [PubMed] [Google Scholar]
  • 12.Anand V, Biondich PG, Liu G, et al. Child health improvement through computer automation: the CHICA system. Stud Health Technol Inform 2004;107:187–91 [PubMed] [Google Scholar]
  • 13.Carroll AE, Biondich PG, Anand V, et al. Targeted screening for pediatric conditions with the CHICA system. J Am Med Inform Assoc 2011;18:485–90 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Downs SM, Biondich PG, Anand V, et al. Using arden syntax and adaptive turnaround documents to evaluate clinical guidelines. AMIA Annu Symp Proc 2006;2006:214–18 [PMC free article] [PubMed] [Google Scholar]
  • 15.Downs SM, Zhu V, Anand V, et al. The CHICA smoking cessation system. AMIA Annu Symp Proc 2008;2008:166–70 [PMC free article] [PubMed] [Google Scholar]
  • 16.Carroll AE, Biondich P, Anand V, et al. A randomized controlled trial of screening for maternal depression with a clinical decision support system. J Am Med Inform Assoc 2012;20:311–16 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Anand V, Carroll AE, Downs SM. Automated primary care screening in pediatric waiting rooms. Pediatrics 2012;129:e1275–81 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Downs SM, Uner H. Expected value prioritization of prompts and reminders. Proceedings of the AMIA Symposium; 2002:215–19 [PMC free article] [PubMed] [Google Scholar]
  • 19.Shwe MA, Middleton B, Heckerman DE, et al. Probabilistic diagnosis using a reformulation of the INTERNIST-1/QMR knowledge base. I. The probabilistic model and inference algorithms. Methods Inf Med 1991;30:241. [PubMed] [Google Scholar]
  • 20.Shortliffe EH. Computer-based medical consultations, MYCIN. Elsevier, 1976 [Google Scholar]
  • 21.Heckerman DE, Nathwani BN. Toward normative expert systems: part II. Probability-based representations for efficient knowledge acquisition and inference. Methods Inf Med 1992;31:106–16 [PubMed] [Google Scholar]
  • 22.Downs SM, Wallace MY. Mining association rules from a pediatric primary care decision support system. Proceedings of the AMIA Symposium; 2000:200–4 [PMC free article] [PubMed] [Google Scholar]
  • 23.Heckerman D. A tutorial on learning with Bayesian networks. In: Innovations in Bayesian networks. Springer, 2008:33–82 http://dx.doi.org/10.1007/978-3-540-85066-3_3 (accessed 6 Mar 2010) [Google Scholar]
  • 24.Tsamardinos I, Brown LE, Aliferis CF. The max–min hill-climbing Bayesian network structure learning algorithm. Mach Learn 2006;65:31–78 [Google Scholar]
  • 25.Ramsey J.2011. Tetrad Project Homepage. http://www.phil.cmu.edu/projects/tetrad/tetrad4.html (accessed 6 Mar 2010)
  • 26.Koller D, Friedman N. Probabilistic graphical models: principles and techniques. MIT Press, 2009 [Google Scholar]
  • 27.Spirtes P, Glymour CN, Scheines R. Causation, prediction, and search. MIT Press, 2000 [Google Scholar]
  • 28.Aliferis CF, Tsamardinos I, Statnikov A, et al. Causal explorer: a causal probabilistic network learning toolkit for biomedical discovery. In: International Conference on Mathematics and Engineering Techniques in Medicine and Biological Sciences (METMBS'03); 2003:371–6 [Google Scholar]
  • 29.Henrion M, Pradhan M, Del Favero B, et al. Why is diagnosis using belief networks insensitive to imprecision in probabilities. In: Proceedings of the Twelfth Conference on Uncertainty in Artificial Intelligence; 1996:307–14 [Google Scholar]
  • 30.Druzdzel MJ. SMILE: Structural Modeling, Inference, and Learning Engine and GeNIe: a development environment for graphical decision-theoretic models. In: Proceedings of the 16th National Conference on Artificial Intelligence and the 11th Innovative Applications of Artificial Intelligence Conference; Menlo Park, CA, USA: American Association for Artificial Intelligence, 1999:902–3 http://portal.acm.org/citation.cfm?id=315149.315504 (accessed 16 Mar 2011). (ACM ID: 315504) [Google Scholar]
  • 31.Hanley JA, McNeil BJ. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982;143:29–36 [DOI] [PubMed] [Google Scholar]
  • 32.Bastian M, Heymann S, Jacomy M.2009. Gephi: An open source software for exploring and manipulating networks. http://www.aaai.org/ocs/index.php/ICWSM/09/paper/view/154. [DOI] [PMC free article] [PubMed]
  • 33.Tsamardinos I, Aliferis CF. Towards principled feature selection: Relevancy, filters and wrappers. In: Proceedings of the Ninth International Workshop on Artificial Intelligence and Statistics; 2003 [Google Scholar]
  • 34.Brandes U. A faster algorithm for betweenness centrality. J Math Soc 2001;25:163–77 [Google Scholar]
  • 35.Tarjan R. Depth-first search and linear graph algorithms. SIAM J Comput 1972;1:15 [Google Scholar]
  • 36.Anand V, Downs SM. Probabilistic asthma case finding: a noisy or reformulation. AMIA Annu Symp Proc 2008;6–10 [PMC free article] [PubMed] [Google Scholar]
  • 37.Downs SM, Klein JD. Clinical preventive services efficacy and adolescents’ risky behaviors. Arch Pediatr Adolesc Med 1995;149:374–9 [DOI] [PubMed] [Google Scholar]
  • 38.Downs SM, Carroll AE, Anand V, et al. Human and system errors, using adaptive turnaround documents to capture data in a busy practice. AMIA Annu Symp Proc 2005; 2005:211–15 [PMC free article] [PubMed] [Google Scholar]
  • 39.Wright A, Chen ES, Maloney FL. An automated technique for identifying associations between medications, laboratory results and problems. J Biomed Inform 2010;43:891–901 [DOI] [PubMed] [Google Scholar]
  • 40.Aliferis CF, Tsamardinos I, Statnikov A. HITON: a novel markov blanket algorithm for optimal variable selection. In: Proceedings of the AMIA Symposium; 2003:21–5 [PMC free article] [PubMed] [Google Scholar]
  • 41.Klann J. Applications of network analysis to clinical data. In: Proceedings of the AMIA Symposium; 2011:1838 [Google Scholar]
  • 42.Murphy KP.2002. Dynamic Bayesian networks: representation, inference and learning. http://portal.acm.org/citation.cfm?id=937223 (accessed 2 Apr 2010)
  • 43.Nodelman U, Shelton CR, Koller D. Continuous time Bayesian networks. In: Proceedings of the 18th Conference on Uncertainty in Artificial Intelligence; 2002:378–87 [Google Scholar]

Articles from Journal of the American Medical Informatics Association : JAMIA are provided here courtesy of Oxford University Press

RESOURCES