Skip to main content
Journal of the American Medical Informatics Association : JAMIA logoLink to Journal of the American Medical Informatics Association : JAMIA
. 2017 Nov 6;25(5):585–592. doi: 10.1093/jamia/ocx118

Interface, information, interaction: a narrative review of design and functional requirements for clinical decision support

Kristen Miller 1,, Danielle Mosby 1, Muge Capan 2, Rebecca Kowalski 1,2, Raj Ratwani 1, Yaman Noaiseh 3, Rachel Kraft 2, Sanford Schwartz 4, William S Weintraub 5, Ryan Arnold 2,6
PMCID: PMC6018977  PMID: 29126196

Abstract

Objective

Provider acceptance and associated patient outcomes are widely discussed in the evaluation of clinical decision support systems (CDSSs), but critical design criteria for tools have generally been overlooked. The objective of this work is to inform electronic health record alert optimization and clinical practice workflow by identifying, compiling, and reporting design recommendations for CDSS to support the efficient, effective, and timely delivery of high-quality care.

Material and Methods

A narrative review was conducted from 2000 to 2016 in PubMed and The Journal of Human Factors and Ergonomics Society to identify papers that discussed/recommended design features of CDSSs that are associated with the success of these systems.

Results

Fourteen papers were included as meeting the criteria and were found to have a total of 42 unique recommendations; 11 were classified as interface features, 10 as information features, and 21 as interaction features.

Discussion

Features are defined and described, providing actionable guidance that can be applied to CDSS development and policy. To our knowledge, no reviews have been completed that discuss/recommend design features of CDSS at this scale, and thus we found that this was important for the body of literature. The recommendations identified in this narrative review will help to optimize design, organization, management, presentation, and utilization of information through presentation, content, and function. The designation of 3 categories (interface, information, and interaction) should be further evaluated to determine the critical importance of the categories. Future work will determine how to prioritize them with limited resources for designers and developers in order to maximize the clinical utility of CDSS.

Conclusion

This review will expand the field of knowledge and provide a novel organization structure to identify key recommendations for CDSS.

Keywords: design features, CDSS, interface, information, interaction, heuristics

OBJECTIVE

The objective of this work is to inform electronic health record (EHR) alert optimization and clinical practice workflow by identifying, compiling, and reporting design recommendations for clinical decision support systems (CDSSs) to support the efficient, effective, and timely delivery of high-quality care.

BACKGROUND AND SIGNIFICANCE

Provider acceptance and associated patient outcomes are widely discussed in the evaluation of CDSSs, but critical design criteria for tools have generally been overlooked. CDSSs, at the simplest level, are tools to help clinicians and patients make better informed decisions during use of the EHR. In the best examples, successful CDSSs reduce medical errors,1 increase health care quality and efficiency,2 and guide appropriate care decisions.3–5 These are challenging tasks, and thus successful CDSSs remain difficult to develop and implement.6 Indeed, a significant amount of data indicates there are several unintended consequences associated with a high alert burden.7–9 Unsystematic alert management can lead to nonadherence, high override rates, and “alert fatigue,” in which clinicians neglect CDSS and other alerts, thereby diminishing their effectiveness and potential benefits. As a result, CDSS developers and users aspire to improve alert management to achieve better acceptance rates and improved care delivery.

A CDSS is defined as “software that is designed to be a direct aid to clinical decision-making, in which the characteristics of an individual patient are matched to a computerized clinical knowledge base and patient-specific assessments or recommendations are then presented to the clinician for a decision.”10 CDSSs have the potential to enable clinicians to better address rising information needs, providing the opportunity to pick up on subtle early indications of risk or vulnerability while sorting through an avalanche of data. The availability of evidence-based guidelines for clinical care and for CDSS implementation encourages providers to deliver the best evidence-based care available.

Having patient demographic and health information in a medium that affords sophisticated and rapid analysis to identify tailored treatment solutions, and being able to deliver this information to the provider at the point of care, can transform treatment. Unfortunately, many EHRs suffer from poor usability – providers struggle to use these systems efficiently, effectively, and safely. While underlying models and algorithms of CDSSs have been intensively studied,11–16 there remains a lack of evidence-based guidelines in terms of functional and design requirements of the system. When implemented with critical design limitations, there is the potential for difficulty in routine use and mistrust of the recommendations, leading to reduced use and increased provider frustration.6,17,18 These issues limit the potential utility of even well-designed predictive tools. In order to promote the use of CDSSs by providers, we need to determine the best way to design and present information via the user interface. The value of this synthesis is to compile a set of recommendations for CDSS design to better define functional characteristics that impact how the system looks and performs, with the ultimate goal of supporting clinical decision-making. Evaluating the design of these systems from a human factors engineering perspective can result in improved interactions between clinicians and the system, with the potential to eventually improve clinician acceptance and patient outcomes.

MATERIALS AND METHODS

Data sources and searches

This narrative review was completed using a predefined protocol. Our research team searched PubMed and peer-reviewed manuscripts in The Journal of the Human Factors and Ergonomics Society (January 1, 2000, to December 31, 2016) to capture articles specific to clinical design making in health care as well as articles that showcase design features of CDSS and recommendations of these features. The following Medical Subject Headings terms were used for searching publications: “clinical decision support systems” OR “decision support systems” OR “clinical alerts” OR “alert systems” AND “characteristics” OR “features” OR “human factors” OR “usability heuristics” OR “usability factors” OR “recommendation.”

We designed our inclusion/exclusion criteria to identify papers that discuss/recommend design features of CDSSs that have been found to be associated with the success of these systems. Inclusion criteria were as follows: (1) English language, (2) EHR-based, (3) provide specific characteristics in terms of CDSS design, and (4) define success factors used to evaluate CDSS performance. Success factors were defined as “effective” with respect to the practitioner’s perspective (ie, provider usability; provider satisfaction; efficacy evaluation including process of care; compliance with guidelines; knowledge, attitudes, skills; and fit of workflow and cognitive processes) or improved patient outcomes (ie, morbidity or mortality, surrogate outcomes, and indicators of resource use). Exclusion criteria ruled out papers that discuss features without assessment of (1) performance or (2) success factors. Within the included papers, the authors must describe the design of the CDSS, providing key elements. Recommendations had to represent consensus, defined as recommendations identified in a systematic review and/or a design recommendation made by more than one paper (referenced within the article) demonstrating success factors. This process required reviewing all included manuscripts to identify recommendations and considering these recommendations in aggregate.

Consensus and evaluation

Two members of the research team (DM, YN) were responsible for the initial inclusion/exclusion of articles. The entire research team was then responsible for reviewing the final selection of manuscripts. To evaluate the papers, several of the authors read them individually and convened with the group (KM, DM, YN, MC, RK, and RK) to discuss the different opinions until we achieved agreement about the recommendation and its inclusion. To achieve consensus, every identified recommendation was reviewed and the review team discussed relevance and repetition to determine recommendation for inclusion. Recommendations were then organized in a meaningful way through categorization and grouping based on function, content, and design. Similar recommendations were unified into a final set of recommendations and then grouped into 3 main “I” categories: interface, information, and interaction. Interface (presentation) features were further organized into 4 “P” categories: presentation, placement and positioning, and provision of multiple presentation layers. Information (content) features were further organized into 4 “C” categories: clean and concise, content guidance, and consistency. Interaction (function) features were further organized into 5 “F” categories: fast, fit, feedback, forgiveness, and flexible design. The below summary includes recommendations identified from the 14 included papers and references to the original published work that led to the consensus recommendation.

RESULTS

Included manuscripts

The literature search identified a total of 387 unique results, of which 281 were dismissed after abstract review and 79 were dismissed after complete manuscript review. Finally, a total of 27 papers were selected as relevant to our project. Through the consensus process, 14 papers were included as meeting the criteria (Supplementary Table S1). Regarding the types of papers found, several were obtained. We identified 9 literature reviews of recommended CDSS design elements,19–27 3 evaluations of a single specific successful CDSS,28–30 1 comparison of 2 CDSS interfaces,31 and 1 focus group working to design a CDSS.32 Although The Journal of Human Factors and Ergonomics Society was searched, no relevant articles were considered appropriate for this narrative review.

Outcomes representing success

The results and effectiveness of these CDSSs were evaluated using 2 metrics: improved significance and improved outcomes of care. Improved significance (disregarding the importance of outcomes or the magnitude of the improvement) included:

  • Provider usability20,25,29

  • Provider satisfaction20,25,27,29–32

  • Evaluating efficacy, including process of care (eg, diagnosis, treatment, monitoring)19,21,22,24

  • Compliance with guidelines21,31

  • Knowledge, attitudes, skills22

  • Fit of workflow and cognitive processes20,28

Improved outcomes of care included:

  • Morbidity or mortality (eg, blood pressure, clinical events, quality of life)19,21,22

  • Surrogate outcomes (eg, time taken to achieve a stable therapeutic dose)21,22,27

  • Indicators of resource use (eg, duration of hospital stay)22

Design recommendations

A total of 42 unique recommendations were identified in the selected 14 papers. From these unique recommendations, we grouped features into 3 systematized categories: 11 were classified as interface features (Box 1), 10 as information features (Box 2), and 21 as interaction features (Box 3).

Box 1.

Interface features categorized as presentation, placement, positioning, and provision of multiple presentation layers

Interface (Presentation)
Presentation
 Make it simple
 Use appropriate font sizes
 Use meaningful colors
 Ensure acceptable contrast between text and background
 Keep presentation consistent
 Deploy space-filling techniques
 Make icons bold or bigger in size
Placement and positioning
 Display information in prominent positions to ensure that it is  seen
 Allow for reading left to right
 Localize information
Provision of multiple presentation layers
 Avoid using only text

Box 2.

Information features categorized as clean and concise, content guidance, and consistency

Information (Content)
Clean and concise
 Standardize terminology
 Use concise and effective language
Content guidance
 Provide a recommendation, not an assessment
 Justify recommendations
 Suggest alternative recommendations
 Provide additional resources
 Make evidence-based recommendations the default
 Keep recommendations up to date
Consistency
 Recommendations should come from the same place
 Have the same display of basic CDSS for all members of the health  care team

Box 3.

Interaction features categorized as fast, fit, feedback, forgiveness, and flexible design

Interaction (Function)
Fast
 Provide timely feedback
 Reduce the amount of time the user is required to interact with the  CDSS
Fit
 Minimize cognitive load (reduce the number of mouse clicks and  amount of free-text typing; use selection tools, sort options)
 Minimize cognitive load (request information from the provider only when necessary; reduce manual input of values)
 Reduce screens to facilitate navigation and to promote efficient  interactions
 Automatically pull data from the EHR/integrate into the charting  system
 Navigate to appropriate locations
 Initiate intervention and take advantage of interactivity (system provides corollary action)
 Provide a route to get to provider-specific info
 Adapt its behavior according to a subset of relevant actions taken by clinicians
 Incorporate functions supporting the dialog between the CDSS and the clinician
Feedback
 Provide decision support automatically as part of clinician workflow
 Automate alerting
 Request documentation of reasons for not following system  recommendations
Forgiveness
 Allow the user to be able to modify orders
 Integrate a reset button
Flexible design
 Involve the patient
 Utilize adaptive design and feedback
 Provide an indication for all professionals of the availability of  information; the designers may choose the most appropriate  way of indicating the information in the interface
 Incorporate functions to support team awareness about alert  management and its evolution over time (eg, visible access to  how the alert was handled and the reasons for alert override  or rule deactivation if any has been documented)
 Give access upon request to extended information (eg, justification  for the rule, attached scientific documentation) that should be  structured depending on the user profile

Interface features

Presentation

Presentation describes aesthetic appeal and integrity of design. One recommendation is to make it simple.17,20,25,27,31 Words like “easy” and “simple” are recommended frequently with20,31 and without25,27 specific guidance. Simplicity includes only the elements that are most important for communication. Techniques used to promote simplicity include consistent terminology, concise and unambiguous language, effective visualization, improved readability, and reduced density of information.31,33–36 Specific strategies include using “at-a-glance” interface design31,37 and using reduced and concise words.20,31 Readability is achieved when the display is easy to identify and interpret through an appealing and attractive presentation. To improve readability, it is suggested to consider the use of appropriate font sizes,31,33use meaningful colors,31,33,34ensure acceptable contrast between text and background,31,33 and make icons bold or bigger in size.29Space-filling techniques help maximize the amount of information that can displayed in the available display space.31,38,39 Visibility factors take into account human factors and cognitive informatics.

Placement and positioning

Placement and positioning refers to the optimal location of data elements within the design of the user interface. CDSS developers should display information in prominent positions to ensure that it is seen.23,31,35,40 To meet cultural standards and address natural mapping, the system should allow for reading left to right.41–43 With regard to positioning, localizing information31 groups pieces of information together to facilitate on-screen searches.34,35,40 Techniques to strategize positioning include “at a glance,” “at hand,” and “all in one.”30

Provision of multiple presentation layers

The presentation layer represents the interface between users and the rest of the application. If users cannot interact with the application in a way that lets them perform their work in an efficient and effective manner, then the overall success of the application will be severely impaired. It is recommended to avoid using only text20,29,31 and use tables,44 graphs,44 buttons,33 scroll bars,33 and iconic language45–47 to ensure that the density of information is appropriate. Reducing text density makes the CDSS easier to interpret during busy clinician encounters and keeps the attention of the provider. Additionally, information-oriented, systematic graphic design helps providers understand complex information.

Information features

Clean and concise

All components should be designed so their meaning is not ambiguous. Information should demonstrate clarity by being presented as clean and concise. It is recommended to standardize terminology. Designers should use concise and effective language based on a consistent terminology to promote simplicity, consistency, efficient interactions, and effective use of language.31,34–36,40 The most important elements should be easily perceived. Noncritical elements should be minimized so as not to hide critical information.

Content guidance

A CDSS encompasses a variety of tools, including clinical guidelines, diagnostic support, computerized alerts and reminders, and contextually relevant reference information. An effective CDSS must be relevant to those who can act on the information, and the design of recommendations within a CDSS plays an essential role. A CDSS should provide a recommendation, not solely an assessment,19,21,25,32,48,49 and explain and justify the recommendations and their source28,31,34,50 by providing reasons10,25,51 and research evidence,10,25,41 particularly for systems that require a reason for overriding advice.17,19,21,23 It should allow for the provision of advice, suggestions, and alternative recommendations to increase compliance and to respect the autonomy of the physician.17,21,23,31,34 It should provide additional resources or access to additional knowledge when needed,29 including justification of recommendations and rules, and scientific documentation structured depending on the user.28 It should make evidence-based recommendations the default17,21,23 and keep recommendations up to date.21

Consistency

Consistency limits the number of ways recommendations are presented, reducing the cognitive workload by eliminating confusion and following standards and conventions. A CDSS should support situational awareness for an entire team of health care professionals. Therefore, recommendations should come from the same place and the system should have the same display of basic CDSS information on the case at hand for all professionals.28

Interaction features

Fast

A CDSS must be fast with regard to the delivery of recommendations and use.17 It should provide timely feedback.27 A CDSS should be timely upon submission of results23,32 and at the time and location of decision-making.17,19,23,28,48,52–57 It should reduce the amount of time the user is required to interact with the system.20,25,29 The assessment form (data entry) should not be too long or take too long to complete.25 This can be accomplished by reducing the manual input of values,20 but also the number of mouse clicks and amount of free-text typing (eg, provide a standardized selection of options).29 This recommendation is critical for CDSSs that address routine medical problems, but is optional for more complicated assessments.20

Fit

For CDSSs to reach their full potential, complex data must be rapidly accessible and easily understood within a provider’s workflow. Total cognitive load refers to the amount of mental processing power needed to use the CDSS and affects how easily providers enter data and retrieve results. The CDSS should minimize cognitive load.20,21,31 The CDSS should request information from the provider only when necessary to reduce the manual input of values.17,21,23 It should use selection tools (eg, dropdown boxes, field types) and sort options to facilitate ease of use and reduce cognitive load and potentially user error. It should have reduced screens to facilitate navigation and to promote efficient interactions.31,34,35,40 To complement provider workflow and reduce cognitive load, the CDSS should automatically pull data from the EHR/integrate into the charting system,20–22,28,41,53,54,58–62 meaning the CDSS is an integrated component of charting or the order entry system.26 Integration with the EHR allows the CDSS to navigate to appropriate locations and initiate intervention and take advantage of interactivity by initiating interventions, for example, by executing recommendations by noting agreement.20,25 Lastly, the system should be the clinician’s partner.28 It should provide a route to get to provider-specific information.28 The CDSS should be adapted according to a subset of relevant actions taken by clinicians28 and adapt its function to the evaluation of the outcome at risk over time (eg, take into account the evolution of target values to adapt for severity). It should incorporate functions supporting the dialog between the CDSS and the clinician.28

Feedback

Feedback refers to the ability of a system to send information back to the user about what action has been done and what result was accomplished. The system should provide decision support automatically as part of clinician workflow,28,41,53,54,58–62 meaning the CDSS is provided at the point of care or order.26 It should have automated alerting21,30 and develop automatic prompting of users by the decision aid,22,63 as opposed to systems that rely on manual processes where clinicians are required to seek out the advice of the decision support system.26 It should request documentation of reasons for not following CDSS recommendations48,52 by prompting clinicians to record a reason when they do not follow the advised course of action rather than allowing the system advice to be bypassed without recording a reason.26

Forgiveness

Forgiveness refers to flexibility of use, providing immediate and reversible actions. The system should allow the user to be able to modify orders32 and should integrate a reset button.26,29 Errors in data entry and selection occur, and users require the ability to make changes or begin using the tool again.

Flexible design

CDSSs must be flexible and adaptable, able to explore multiple assumptions and incorporate new information as circumstances change. CDSSs offer a process for enhancing health-related decisions and actions with pertinent, organized clinical knowledge and patient information. Information recipients can include patients, clinicians, and others involved in patient care delivery; therefore, it is recommended to involve the patient if possible, in both design and providing advice,19 which would empower patients to become actively involved in their own care (and provide actionable advice outside of a clinical encounter). The system should utilize adaptive design and feedback. Adaptive design maximizes the space and layout of a tool based on available space for the CDSS interface, but can also be applied in a more general sense in adapting to the provider using the CDSS. It should provide assistance without user control over output.22 It should provide an indication for all professionals of the availability of any information, allowing for the integration of new clinical information.28 The system should identify latent needs and inform the end user,17,27 improving the clinical utility of the tool. It should incorporate functions to support team awareness about the alert management and its evolution over time (eg, visible access to how the alert was handled and the reasons for alert override or rule deactivation if any has been documented)28 and give access upon request to extended information (justification of the rule, attached scientific documentation, etc.) that should be structured depending on the user profile.28

DISCUSSION

This narrative review article contributes to the body of literature on CDSS, as we are unaware of previously published articles at this scale and level of detail. This review will expand the field of knowledge and provide a novel organizational structure to identify key recommendations for CDSS.64 There are serious shortcomings in the existing system of alerting providers in a meaningful way and providing actionable, evidence-based clinical decision support.17,65 Lack of knowledge regarding alert presentation to providers has impeded alert optimization. Specifically, little guidance is provided for creating the most effective ways to differentiate alerts, highlighting important pieces of information without adding noise, or creating a universal standard.66

It remains a challenge to implement CDSS effectively without inducing “alert fatigue,”67 defined as a process “wherein clinicians may inadvertently ignore clinically useful alerts, thus diminishing the systems’ effectiveness and possibly leading to serious adverse consequences for patients.”68 Researchers following all of the recommendations found in this narrative review may still not prevent alert fatigue, which may arguably be the most important barrier to achieving an effective CDSS. Research exists regarding the number of alerts providers receive through the EHR system,69,70 but to date, no “best practices” in the design of these alerts have been identified. The recommendations identified in this narrative review will help optimize design, organization, management, presentation, and utilization of information through presentation, content, and function. The designation of 3 categories (interface, information, and interaction) should be further evaluated to determine the critical importance of the categories. Future work will determine how to prioritize them, given limited resources for designers and developers in order to maximize the clinical utility of CDSS.

Human factors engineering principles suggest that the format and presentation of CDSS may not be readily applied in the busy clinical setting and may fail to instill confidence in clinical staff. Effective presentation of an alert, including what is displayed and how it is displayed, may offer better cognitive support during busy patient encounters and may help providers extract information quickly. Following good human factors principles, alerts should signal an important matter and inform and guide the provider.71 Our review identified best practices based on effective CDSSs, including provider preference and improved patient outcomes. These recommendations are specific and actionable, meaning they can easily be applied by developers and CDSS designers for future implementations.

Limitations of this work include the subjective basis that defines the study design within a narrative review based on our opinion, which leads to a selection bias. As this review was qualitative in nature and focused on recommendations provided by the authors, our review did not focus on any quantitative methods the papers may have also provided. Another important limitation of this narrative review was that some of the recommendations we found are not very actionable, such as “use appropriate fonts,” “use meaningful colors,” and “identify latent needs.” Additional guidance is needed to determine what is actionable, as well as what recommendations may be more important to follow than others.

There is mounting proof that health information technology interventions improve patient outcomes through early diagnosis and initiation of evidence-based protocols for critically ill patients.72 The rapid acceleration of technology and the convergence of predictive analytics and human factors address the Centers for Medicare and Medicaid Services stage 3 meaningful use requirements.73 The process of integrating real-time analytics into clinical workflow will be significantly affected by changes the Centers for Medicare and Medicaid Services has planned for EHRs. A shift toward more agile and collaborative infrastructure building is expected to be a key feature of health information technology strategies, such as CDSS design, in the future. As interoperability and big data analytics capabilities become increasingly central to crafting the health care information systems of the future, there is a critical need to optimize the flow of health information and communication to support clinical decision-making.

From a policy perspective, this review highlights the need to establish basic design standards, with a foundation in human factors, to guide the design, development, implementation, and customization of CDSSs. Policymakers must recognize the importance of usability design standards and should develop guidelines for EHR developers and adopters to follow. It is critical that these standards be adhered to by EHR vendors, health care providers, and other stakeholders. By prioritizing usability and adopting a common set of design standards to ensure consistency across EHR platforms, we can begin to improve usability, enhance functionality, and develop technology standards that allow full integration of the EHR into the care system to improve clinical decision-making. Federal agencies, like the Office of the National Coordinator for Health Information Technology, should consider policy mechanisms to encourage creation of, adoption of, and adherence to design standards to ensure the safe and effective use of CDSSs. These policies must be designed and implemented in a way that promotes safety, while also being careful not to stifle innovation.

CONCLUSION

Human factors, usability, and human-computer interaction principles are fundamental to successful CDSSs. We found that by compiling a set of recommendations for CDSS design, we were able to synthesize the design and functional characteristics that impact the performance of the system. Evaluating the recommendations derived from the papers included in this qualitative narrative review, we classified 3 “I” categories – interface, information, and interaction – further organized by 5 “P” categories (presentation, placement and positioning, and provision of multiple presentation layers; 4 “C” categories (clean and concise, content guidance, and consistency; and 5 “F” categories (fast, fit, feedback, forgiveness, and flexible design). We compiled this list of recommendations as general guidance for the design of successful CDSSs, and we believe that testing and validating such a unified set of recommendations could eventually enrich the literature and the field by providing domain-specific design heuristics.

DISCLAIMER

The authors declare no conflicts of interest.

FUNDING

This work is supported by an Institutional Development Award from the National Institute of General Medical Sciences of the National Institutes of Health under grant number U54-GM104941 (PI: Binder-Macleod), and by the National Library of Medicine of the National Institutes of Health under grant number 1R01LM012300-01A1 (PI: KM).

CONTRIBUTORS

KM was responsible for study design and overall management, and drafted the manuscript. DM conducted literature searches, retrieved publications, provided overall management of the project, participated in consensus evaluation, and revised the manuscript. MC contributed to study design, participated in consensus evaluation, and revised the manuscript. RR revised the manuscript. YM contributed to study design and literature searches, retrieved publications, and participated in consensus evaluation. RK and RK participated in consensus evaluation and revised the manuscript. SS and WW contributed to the study design and revised the manuscript. RA contributed to the study design and revised the manuscript. All authors approved the version of the manuscript to be published.

SUPPLEMENTARY MATERIAL

Supplementary material is available at Journal of the American Medical Informatics Association online.

Supplementary Material

Supplementary Table

REFERENCES

  • 1. Bates D, Cohen M, Leape L, Overhage J, Shabot M, Sheridan T. Reducing the frequency of errors in medicine using information technology. J Am Med Inform Assoc. 2001;84:299–308. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. Teich J, Wrinn M. Clinical decision support systems come of age. MC Comput. 2000;171:43–46. [PubMed] [Google Scholar]
  • 3. Souza N, Sebaldt R, Mackay J et al. , ; Team CSR. Computerized clinical decision support systems for primary preventive care: a decision-maker-researcher partnership systematic review of effects on process of care and patient outcomes. Implement Sci. 2011;6:87. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Sahota N, Lloyd R, Ramakrishna A et al. ; Team CSR. Computerized clinical decision support systems for acute care management: a decision-maker-researcher partnership systematic review of effects on process of care and patient outcomes. Implement Sci. 2011;6:91. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Roshanov P, Misra S, Gerstein H et al. , Computerized clinical decision support systems for chronic disease management: a decision-maker-researcher partnership systematic review. Implement Sci. 2011;6:92. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Ash J, Berg M, Coiera E. Some unintended consequences of information technology in health care: the nature of patient care information system–related errors. J Am Med Inform Assoc. 2004;112:104–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Murphy D, Reis B, Sittig D, Singh H. Notifications received by primary care practitioners in electronic health records: a taxonomy and time analysis. Am J Med. 2012;1252:209.e1–7. [DOI] [PubMed] [Google Scholar]
  • 8. van der Sijs H, Aarts J, Vulto A, Berg M. Overriding of drug safety alerts in computerized physician order entry. J Am Med Inform Assoc. 2006;132:138–47. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Weingart S, Toth M, Sands D, Aronson M, Davis R, Phillips R. Physicians’ decisions to override computerized drug alerts in primary care. Arch Intern Med. 2003;16321:2625–31. [DOI] [PubMed] [Google Scholar]
  • 10. Sim I, Gorman P, Greenes R et al. , Clinical decision support systems for the practice of evidence-based medicine. J Am Med Inform Assoc. 2001;86:527–34. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Moja L, Kwag K, Lytras T et al. , Effectiveness of computerized decision support systems linked to electronic health records: a systematic review and meta-analysis. Am J Public Health. 2014;10412:e12–22. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Ninh A. DocBot: a novel clinical decision support algorithm. In: Conf Proc IEEE Eng Med Biol Soc 2014:2014;6290–93. [DOI] [PubMed] [Google Scholar]
  • 13. Tenorio JM, Hummel AD, Cohrs FM, Sdepanian VL, Pisa IT, Marin H de F. Artificial intelligence techniques applied to the development of a decision-support system for diagnosing celiac disease. Int J Med Inform. 2011;8011:793–802. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Yet B, Perkins Z, Rasmussen T, Tai N, Marsh D. Combining data and meta-analysis to build Bayesian networks for clinical decision support. J Biomed Inform. 2014;52:373–85. [DOI] [PubMed] [Google Scholar]
  • 15. van Belle VMCA, van Calster B, Timmerman D et al. , A mathematical model for interpretable clinical decision support with applications in gynecology. PLoS One. 2012;73:e34312. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Gultepe E, Green JP, Nguyen H, Adams J, Albertson T, Tagkopoulos I. From vital signs to clinical outcomes for patients with sepsis: a machine learning basis for a clinical decision support system. J Am Med Inform Assoc. 2014;21:315–25. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Bates D, Kuperman G, Wang S. Ten commandments for effective clinical decision support: making the practice of evidence-based medicine a reality. J Am Med Inform Assoc. 2003;10:523–30. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Campbell E, Sittig DF, Ash JS, Guappone KP, Dykstra RH. Types of unintended consequences related to computerized provider order entry. J Am Med Inform Assoc. 2006;135:547–56. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Roshanov P, Fernandes N, Wilczynski J et al. , Features of effective computerized clinical decision support systems: meta regression of 162 randomized trials. BMJ. 2013;346:1657. [DOI] [PubMed] [Google Scholar]
  • 20. Martinez-Perez B, de la Torre-Diez I, Lopez-Coronado M, Sainz-de-Abajo B, Robles M, Garcia-Gomez J. Mobile clinical decision support systems and applications: a literature and commercial review. J Med Syst. 2014;381:4. [DOI] [PubMed] [Google Scholar]
  • 21. Mack E, Wheeler D, Embi P. Clinical decision support systems in the pediatric intensive care unit. Pediatrc Crit Care Med. 2009;101:23–28. [DOI] [PubMed] [Google Scholar]
  • 22. Nies J, Colombet I, Degoulet P, Durieux P. Determinants of success for computerized clinical decision support systems integrated into cpoe systems: a systematic review. AMIA Annu Symp Proc. 2006;2006:594–98. [PMC free article] [PubMed] [Google Scholar]
  • 23. Kawamoto K, Houlihan C, Balas E, Lobach D. Improving clinical practice using clinical decision support systems: a systematic review of trials to identify features critical to success. BMJ. 2005;330:765. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Kawamoto K, Lobach D. Clinical decision support provided within physician order entry systems: a systematic review of features effective for changing clinician behavior. AMIA Annu Symp Proc. 2003;2003:361–65. [PMC free article] [PubMed] [Google Scholar]
  • 25. Lee S. Features of computerized clinical decision support systems support of nursing practice. Comput Inform Nurs. 2013;3110:477–95. [DOI] [PubMed] [Google Scholar]
  • 26. Chase J, Andreassen S, Jensen K, Shaw G. Impact of human factors on clinical protocol performance: a proposed assessment framework and case examples. J Diabetes Sci Technol. 2008;23:409–16. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Wright M, Robicsek A. No clinical decision support systems and infection prevention: to know is not enough. Am J Infect Control. 2015;436:554–58. [DOI] [PubMed] [Google Scholar]
  • 28. Pelayo S, Marcilly R, Bernonville S, Leroy N, Beuscart-Zephir M. Human factors based recommendations for the design of medication related clinical decision support systems (CDSS). Stud Heal Technol Inform. 2011;169:412–16. [PubMed] [Google Scholar]
  • 29. Devine E, Lee C, Overby C et al. , Usability evaluation of pharmacogenomics clinical decision support aids and clinical knowledge resources in a computerized provider order entry system: a mixed methods approach. Int J Med Inform. 2014;837:473–83. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Kanstrup A, Christiansen M, Nohr C. Four principles for user interface design of computerized clinical decision support systems. Stud Heal Technol Inform. 2011;166:65–73. [PubMed] [Google Scholar]
  • 31. Tsopra R, Jais J, Venot A, Duclos C. Comparison of two kinds of interface, based on guided navigation or usability principles, for improving the adoption of computerized decision support systems: application to the prescription of antibiotics. J Am Med Inform Assoc. 2014;21(e1):e107–16. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Bright T. Transforming user needs into functional requirements for an antibiotic clinical decision support system. Appl Clin Inform. 2013;44:618–35. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Belden J, Grayson R, Barnes J. Defining and testing EMR usability: principles and proposed methods of EMR usability evaluation and rating. Healthc Inform Manag Syst Soc. 2009. https://mospace.umsystem.edu/xmlui/bitstream/handle/10355/3719/DefiningTestingEMR.pdf;sequence=1. [Google Scholar]
  • 34. Horsky J, Schiff G, Johnston D, Mercincavage L, Bell D, Middleton B. Interface design principles for usable decision support: a targeted review of best practices for clinical prescribing interventions. J Biomed Inform. 2012;456:1202–16. [DOI] [PubMed] [Google Scholar]
  • 35. Khajouei R, Jaspers M. CPOE system design aspects and their qualitative effect on usability. Stud Heal Technol Inform. 2008;136:309–14. [PubMed] [Google Scholar]
  • 36. Middleton B, Bloomrosen M, Dente M et al. , Enhancing patient safety and quality of care by improving the usability of electronic health record systems: recommendations from AMIA. J Am Med Inform Assoc. 2013;20(e1):e2–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Tsopra R, Lamy J, Venot A, Duclos C. Design of an original interface that facilitates the use of clinical practice guidelines of infection by physicians in primary care. Stud Heal Technol Inform. 2012;180:93–97. [PubMed] [Google Scholar]
  • 38. Chazard E, Puech P, Gregoire M, Beuscart R. Using Treemaps to represent medical data. Stud Heal Technol Inform. 2006;124:522–27. [PubMed] [Google Scholar]
  • 39. Johnson B, Shneiderman B. Tree-Maps: a space-filling approach to the visualization of hierarchical information structures. Proc 2nd Conf Vis ’91. 1991;284–91. [Google Scholar]
  • 40. Khajouei R, Jaspers M. The impact of CPOE medication systems’ design aspects on usability, workflow and medication orders: a systematic review. Methods Inform Med. 2010;49:3–19. [DOI] [PubMed] [Google Scholar]
  • 41. Shiffman R, Liaw Y, Brandt C, Corb G. Computer-based guideline implementation systems: a systematic review of functionality and effectiveness. J Am Med Inform. Assoc 1999;62:104–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Maviglia S, Zielstorff R, Paterno M, Teich J, Bates D, Kuperman G. Automating complex guidelines for chronic disease: lessons learned. J Am Med Inform Assoc. 2003;102:154–65. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Linder J, Schnipper J, Tsurikova R et al. , Documentation-based clinical decision support to improve antibiotic prescribing for acute respiratory infections in primary care: a cluster randomised controlled trial. Inform Prim Care. 2009;174:231–40. [DOI] [PubMed] [Google Scholar]
  • 44. Starren J, Johnson S. An object-oriented taxonomy of medical data presentations. J Am Med Inform Assoc. 2000;71:1–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45. Lamy J, Duclos C, Bar-Hen A, Ouvrard P, Venot A. An iconic language for the graphical representation of medical concepts. BMC Med Inform Decis Mak. 2008;8:16. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Lamy J, Venot A, Bar-Hen A, Ouvrard P, Duclos C. Design of a graphical and interactive interface for facilitating access to drug contraindications, cautions for use, interactions and adverse effects. BMC Med Inform Decis Mak. 2008;8:21. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Lamy J, Soulalmia L, Kerdelhue G, Venot A, Duclos C. Validating the semantics of a medical iconic language using ontological reasoning. J Biomed Inform. 2013;461:56–67. [DOI] [PubMed] [Google Scholar]
  • 48. Morris A. Developing and implementing computerized protocols for standardization of clinical decisions. Ann Intern Med. 2000;132:373–83. [DOI] [PubMed] [Google Scholar]
  • 49. Stock J, Waud C, Coderre J et al. , Clinical reporting to primary care physicians leads to increased use and understanding of bone densitometry and affects the management of osteoporosis. A randomized trial. Ann Intern Med. 1998;128(12 Pt 1):996–99. [DOI] [PubMed] [Google Scholar]
  • 50. Mollon B, Chong J, Holbrook A, Sung M, Thabane L, Foster G. Features predicting the success of computerized decision support for prescribing: a systematic review of randomized controlled trials. BMC Med Inform Decis Mak. 2009;9:11. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51. Wendt T, Knaup-Gregori P, Winter A. Decision support in medicine: a survey of problems of user acceptance. Stud Heal Technol Inform. 2000;77:852–56. [PubMed] [Google Scholar]
  • 52. Bennett J, Glasziou P. Computerised reminders and feedback in medication management: a systematic review of randomised controlled trials. Med J Aust. 2003;1785:217–22. [DOI] [PubMed] [Google Scholar]
  • 53. Kaplan B. Evaluating informatics applications – some alternative approaches: theory, social interactionism, and call for methodological pluralism. Int J Med Inform. 2001;641:39–56. [DOI] [PubMed] [Google Scholar]
  • 54. Bodenheimer T, Grumbach K. Electronic technology: a spark to revitalize primary care? JAMA. 2003;2902:259–64. [DOI] [PubMed] [Google Scholar]
  • 55. Tierney W. Improving clinical decisions and outcomes with information: a review. Int J Med Inform. 2001;621:1–9. [DOI] [PubMed] [Google Scholar]
  • 56. Heathfield H, Wyatt J. Philosophies for the design and development of clinical decision-support systems. Methods Inform Med. 1993;321:9–17. [PubMed] [Google Scholar]
  • 57. Wyatt J. Lessons learnt from the field trial of ACORN, an expert system to advise on chest pain. Proc Sixth World Conf Med Informatics. Singapore, 1989;111–5. [Google Scholar]
  • 58. Wetter T. Lessons learnt from bringing knowledge-based decision support into routine use. Artif Intell Med. 2002;243:195–203. [DOI] [PubMed] [Google Scholar]
  • 59. Payne T. Computer decision support systems. Chest. 2000;118(2 Suppl):47S–52S. [DOI] [PubMed] [Google Scholar]
  • 60. Trivedi M, Kern J, Marcee A et al. , Development and implementation of computerized clinical guidelines: barriers and solutions. Methods Inform Med. 2002;415:435–42. [PubMed] [Google Scholar]
  • 61. Aronsky D, Chan K, Haug P. Evaluation of a computerized diagnostic decision support system for patients with pneumonia: study design considerations. J Am Med Inform Assoc. 2001;85:473–85. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62. Hersh W. Medical informatics: improving health care through information. JAMA. 2002;28816:1955–58. [DOI] [PubMed] [Google Scholar]
  • 63. Garg A, Adhikari N, McDonald H et al. , No Effects of computerized clinical decision support systems on practitioner performance and patient outcomes: a systematic review. JAMA. 2005;29310:1233–38. [DOI] [PubMed] [Google Scholar]
  • 64. Byrne J. Improving the peer review of narrative literature reviews. Res Integr Peer Rev 2016;1:12. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65. Phansalkar S, Zachariah M, Seidling HM, Mendes C, Volk L, Bates DW. Evaluation of medication alerts in electronic health records for compliance with human factors principles. J Am Med Inform Assoc. 2014;21:e332–40. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66. Miller R, Waitman L, Chen S, Rosenbloom S. The anatomy of decision support during inpatient care provider order entry (CPOE): empirical observations from a decade of CPOE experience at Vanderbilt. J Biomed Inform. 2005;386:469–85. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67. Russ A, Zillich A, McManus M, Doebbeling B, Saleem J. Prescribers’ interactions with medication alerts at the point of prescribing: a multi-method, in situ investigation of the human-computer interaction. Int J Med Inform. 2012;814:232–43. [DOI] [PubMed] [Google Scholar]
  • 68. Kesselheim A, Cresswell K, Phansalkar S, Bates D, Sheikh A. Clinical decision support systems could be modified to reduce “alert fatigue” while still minimizing the risk of litigation. Health Aff. 2011;3012:2310–17. [DOI] [PubMed] [Google Scholar]
  • 69. Kizzier-Carnahan V, Artis K, Mohan V, Gold J. Frequency of passive EHR Alerts in the ICU: another form of alert fatigue? J Patient Saf. 2016. https://www.ncbi.nlm.nih.gov/pubmed/27331600. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70. Alert Fatigue and Patient Risk: An Effective Drug Decision Support System Could Eliminate Both. Manag Healthc Exec [Internet]. 2016. http://managedhealthcareexecutive.modernmedicine.com/managed-healthcare-executive/news/alert-fatigue-and-patient-risk-effective-drug-decision-support-system-could-eliminate-both [Google Scholar]
  • 71. Engineering Equipment and Materials Users Association. Alarm Systems: A Guide to Design, Management, and Procurement. London: EEMUA, 1999. [Google Scholar]
  • 72. Chaudhry B, Wang J, Wu S et al. , Systematic review: impact of health information technology on quality, efficiency, and costs of medical care. Ann Intern Med. 2006;14410:742–52. [DOI] [PubMed] [Google Scholar]
  • 73. Davenport T, McNeill D. Analytics in Healthcare and the Life Sciences: Strategies, Implementation Methods, and Best {ractices. Upper Saddle River, NJ: Pearson, 2013. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Table

Articles from Journal of the American Medical Informatics Association : JAMIA are provided here courtesy of Oxford University Press

RESOURCES