Abstract
A Bayesian network is a graphical model that uses probability theory to represent relationships among its variables. The model is a directed acyclic graph whose nodes represent variables, such as the presence of a disease or an imaging finding. Connections between nodes express causal influences between variables as probability values. Bayesian networks can learn their structure (nodes and connections) and/or conditional probability values from data. Bayesian networks offer several advantages: (a) they can efficiently perform complex inferences, (b) reason from cause to effect or vice versa, (c) assess counterfactual data, (d) integrate observations with canonical (“textbook”) knowledge, and (e) explain their reasoning. Bayesian networks have been employed in a wide variety of applications in radiology, including diagnosis and treatment planning. Unlike deep learning approaches, Bayesian networks have not been applied to computer vision. However, hybrid artificial intelligence systems have combined deep learning models with Bayesian networks, where the deep learning model identifies findings in medical images and the Bayesian network formulates and explains a diagnosis from those findings. One can apply a Bayesian network's probabilistic knowledge to integrate clinical and imaging findings to support diagnosis, treatment planning, and clinical decision-making. This article reviews the fundamental principles of Bayesian networks and summarizes their applications in radiology.
Keywords: Bayesian Network, Machine Learning, Abdominal Imaging, Musculoskeletal Imaging, Breast Imaging, Neurologic Imaging, Radiology Education
Supplemental material is available for this article.
© RSNA, 2023
Keywords: Bayesian Network, Machine Learning, Abdominal Imaging, Musculoskeletal Imaging, Breast Imaging, Neurologic Imaging, Radiology Education
Summary
Bayesian networks are graphical models that use probability theory to integrate clinical and imaging findings for diagnosis and clinical decision-making, with the advantages of explainability and ability to train on small datasets.
Essentials
■ Bayesian networks have been used in radiology to integrate clinical and imaging findings for differential diagnosis and clinical decision-making.
■ Efficient algorithms enable Bayesian networks to compute posterior probabilities (updated probabilities given additional information), analyze the value of information (to identify the most useful next diagnostic test), perform sensitivity analysis, and generate explanations of the model's reasoning.
■ Bayesian networks offer advantages over various other artificial intelligence approaches, including explainability and accurate inference from small datasets and in the setting of missing data.
■ Although Bayesian networks have not been applied to computer vision, hybrid artificial intelligence systems have combined deep learning models with Bayesian networks; the deep learning model identifies and extracts findings in medical images, and the Bayesian network formulates and explains a diagnosis from those findings and other clinical and laboratory data.
Introduction
Diagnosis and decision-making in radiology require that physicians integrate an imaging study's findings with large amounts of information, including the patient's clinical symptoms, laboratory values, comorbidities, prior medical history, and the likelihood of various diseases. Artificial intelligence (AI) systems can complement physician decision-making. These systems can store large amounts of data and perform complex calculations consistently, quickly, and without biases. A Bayesian network is an AI formalism, first introduced in 1985, that can learn, store, and apply knowledge in the form of probability values to reason under conditions of uncertainty (1). Bayesian networks have been used in radiology to integrate clinical and imaging findings for differential diagnosis and clinical decision-making (2–4). This article offers a brief refresher of probability theory, introduces Bayesian networks as a graphical representation of probabilistic knowledge, reviews current applications of Bayesian networks in radiology, and discusses future applications. A glossary of Bayesian network terms (Table S1), Bayesian networks’ advantages and disadvantages (Table S2), and additional resources (Table S3) are included.
Probability: Background and Definitions
Bayesian networks use probability to reason in settings of uncertainty. Probability theory offers a set of methods to conceptualize and quantify uncertainty. Probability values obey three fundamental properties: (a) an event's probability value must be greater than or equal to 0, (b) the values of all possible events must sum to 1, and (c) the probability of A or B is the sum of their individual probabilities if the two events are mutually exclusive.
Conditional probability expresses the likelihood of one event given another and is defined as probability that the two events, A and B, occur together (their joint probability) divided by the probability of the event conditioned upon. Thus, the conditional probability of A given B is defined as P(A|B) = P (A,B)/P(B). Conditional probabilities are familiar concepts in diagnostic medical testing, such as sensitivity, specificity, positive predictive value, and negative predictive value. Conditional probabilities of A and B are related using Bayes theorem,
![]() |
Bayes theorem allows probabilistic reasoning in both directions, that is, from the probability of B given A to the probability of A given B, or vice versa. In the setting of a diagnostic test, where A and B might represent a disease and a positive test result, respectively, one could compute the probability of the test result given the disease, or that of the disease given the test result.
Bayesian Networks
In medical applications, Bayesian networks typically express relationships between uncertain variables, such as the presence of a disease, and observable variables, such as demographic data, symptoms, vital signs, laboratory values, and imaging findings. Bayesian networks can integrate large amounts of probabilistic knowledge in a mathematically robust way that eliminates many of the heuristic biases of human decision-making. A Bayesian network model consists of a graph and an associated set of conditional probability tables (1,5,6). The graph incorporates a set of nodes and a set of edges, or arcs, that connect pairs of nodes. The nodes in a Bayesian network represent stochastic, or random, variables. Variables can have continuous or discrete values; we limit our discussion to variables with discrete values, as they are seen most frequently in medical applications. Continuous variables, such as age or laboratory values, are typically discretized. Each node thus has two or more states that each have an associated probability value. The states must be exhaustive (they cover all possibilities) and mutually exclusive (only one can be true). Thus, the probability values for any node's states always sum to 1. The edges of a Bayesian network form a directed acyclic graph, as the edges have specified direction and cannot form a closed-loop cycle.
Each node has a conditional probability table that specifies its states’ probabilities given the values of the node's incoming edges. For nodes without incoming edges, the conditional probability table specifies the prior probability values of the nodes’ states. An edge between two nodes represents probabilistic influence between the two variables, expressed as a set of conditional probabilities. By convention, the orientation of the edge conveys the direction of causal influence.
To illustrate, Figure 1 presents a simple Bayesian network with three variables: (a) age, a continuous variable that has been discretized into five states (eg, 40–49 years); (b) breast cancer, with states present and absent, influenced by the patient's age; and (c) mammogram, with states normal and abnormal, influenced by the presence or absence of breast cancer. Prior probabilities have been assigned for age; conditional probabilities have been specified for breast cancer and mammogram.
Figure 1:
An example Bayesian network model with three nodes. Tan-colored nodes are inferred from known information; nodes in gray have been set (instantiated) to a specific value, typically in response to an observation. An arrow indicates the direction of probabilistic influence; if an arrow points from node A to node B, then the probabilities of B's states are conditioned on A. For example, the probability of abnormal mammographic findings is conditioned on the presence or absence of breast cancer. (A) The conditional probability tables for breast cancer and mammogram are shown beneath those nodes; here the sensitivity of mammography for breast cancer, P (mammogram = abnormal | breast cancer = present), is 90%. Prior probabilities have been assigned to the age ranges. Before any evidence is presented, there is a 1.26% probability that breast cancer is present. (B) If we know that the patient is of age 70 or greater, the model infers a 2.60% probability that breast cancer is present. (C) If that patient has an abnormal mammogram, the probability of breast cancer rises to 32.5%. Note that although the arrow points from breast cancer to mammogram, information can flow in each direction. As one would expect, the result of the mammogram influences the patient's likelihood of breast cancer.
Inference
Inference is the computational procedure that applies a Bayesian network's probabilistic knowledge. Highly efficient algorithms enable Bayesian networks to compute posterior probabilities, analyze the value of information, perform sensitivity analysis, and generate explanations of the model's reasoning.
Posterior Probabilities
The most fundamental Bayesian network inference task is to compute posterior probabilities, which are the probabilities of the model's states after evidence is presented. Given a set of evidence E, specified by fixing the states of any subset of nodes in the network, one can calculate the marginal posterior probabilities, P(□|E), of all other nodes. Because evidence in Bayesian networks can be transmitted in either direction across a link, Bayesian networks can be used to perform predictive inference (from causes to effects), diagnostic inference (from effects to causes), or any combination thereof. Efficient algorithms can propagate beliefs throughout the network as additional evidence is acquired (6,7). Additionally, one can analyze a counterfactual query, such as “If A were true (when it is known to be false), what is the probability of C given the other known information?” (8).
Value of Information
In settings in which evidence can be obtained incrementally (eg, through additional laboratory tests or imaging examinations) a user may wish to know which piece of additional information would have the greatest influence on the probability of the hypothesis node or nodes. In other words, which test result will add the greatest certainty to the diagnosis? A Bayesian network can calculate the value of information by computing the mutual information (or cross-entropy) of the hypothesis node H given a particular observation node O, written I(H;O). The most informative evidence for that hypothesis is the one with the highest value of I(H;O) (9).
Sensitivity Analysis
When using a Bayesian network in practice, a natural question is how much confidence one should have in the network's output. In other words, how sensitive are the results to small changes in the network's parameters? A model's sensitivity can be determined by calculating the derivatives of the posterior probability distributions of the set of target nodes (eg, diagnostic hypotheses) over each of the conditional probability table entries in the Bayesian network (10–12). A large derivative means that small changes in the parameter may lead to large changes in the target nodes’ posterior probabilities. In most cases, one can identify a small subset of parameters, called the sensitivity set, that influences the probability of the target nodes.
Explanation
Although deep neural networks have been applied successfully to numerous image- and text-based applications in radiology, their “black-box” property has rekindled interest in explainable AI (13). Bayesian networks offer strong advantages over neural networks; one can exploit a Bayesian network's structure to describe direct and indirect influences between variables. A Bayesian network model explicitly specifies probabilistic relationships between variables and can quantify those relationships. Two types of explanations, microlevel and macrolevel, have been described (14). Microlevel explanations identify the pieces of evidence that had the greatest influence on the hypothesis of interest. Explanation is generated using a computation similar to that for value of information; one can quantify the impact of omitting that evidence (15). Macrolevel explanations identify the network paths through which evidence flows to influence the hypothesis. The technique identifies all paths through the network from evidence nodes to the hypothesis and measures the strength of each, defined as the amount by which the evidence influences the probability of each node along the path. The idea is that the path is as strong as its weakest link. The paths can then be displayed and ranked in terms of strength (eg, Fig 2D) (15,16).
Figure 2:
As illustrated schematically in independent examples, Bayesian networks can (A) perform exact and efficient computation over models with hundreds or thousands of variables, (B) reason “forwards” from causes to effects as well as “backwards” from effects to causes, (C) learn their structure and probability values from manual input and/or directly from data, and (D) explain their reasoning.
Constructing a Model
There is considerable flexibility in the construction of Bayesian networks; one can apply manual and/or automated approaches for structure learning and parameter learning. Structure learning determines the structure of the Bayesian network's graphical model, including the dependence and independence of variables and the placement of the edges of the graph. Structure learning methods include scoring-based search algorithms (using optimization techniques), constraint-based algorithms (to test for conditional dependence in the data), and Bayesian methods (to generate an ensemble of possible structures). These methods are beyond the scope of this review and are discussed in detail elsewhere (17,18). Parameter learning applies a given dataset to determine the conditional probability values at each node. For example, Burnside et al (19) trained a Bayesian network from a large database of patient demographics and mammographic findings to accurately predict the presence of breast cancer. Several approaches have been developed to automatically deduce both the structure and the parameters of a Bayesian network from data (20,21).
One can combine manual and automated approaches to parameter learning by blending probability values provided by experts and those learned from data. Laplace smoothing is a common approach that adds a smoothing parameter, λ > 0, to each state's count to assure a nonzero value (22). For example, for a variable with three states, Laplace smoothing would assign an initial count of λ to each state to give each state a probability of one-third. As actual data are accrued, each state's count is incremented accordingly. Laplace smoothing provides a simple way to avoid overfitting by pulling the probability estimates away from zero and toward the uniform distribution, and the effect of smoothing wanes as more data are accrued.
Applications in Clinical Radiology
Abdominal Imaging
Several diagnostic models have been developed in abdominal imaging. Decision Aid for Diagnosing Liver Lesions (DAFODILL) aided in diagnosis of 14 types of liver lesions using eight MRI-based features and four clinical features (23). A model for diagnosis of gallbladder disease incorporated probability data from the peer-reviewed literature to evaluate hepatobiliary diseases based on a patient's age and sex, imaging findings, physical examination findings, and laboratory values (24). This model served as a test bed for a system to explain a Bayesian network model's reasoning. A model for diagnosis of acute appendicitis in children applied four US imaging features (appendiceal diameter, wall hyperemia, appearance of periappendiceal mesentery, presence of appendicolith) and nine clinical and laboratory variables (including duration of symptoms, presence of fever, white blood cell count, and C-reactive protein) (25). This model was derived from existing clinical data, published literature, and expert opinions without requiring a large dataset to train on. As an example, in a pediatric patient with an appendix of 11 mm, appendicolith present, wall hyperemia, fever, and 48 hours of symptoms, the Bayesian network determined a 97% probability of acute appendicitis. In a pediatric patient with nonvisualized appendix, normal-appearing mesentery, normal white blood cell count, no fever but with rebound tenderness and nausea, the Bayesian network calculated a 34% probability of acute appendicitis. The model was able to assign a probability even if certain imaging or clinical data were missing or not available.
Musculoskeletal Imaging
Primary bone tumors are rare and can present a diagnostic challenge. A Bayesian network model was able to correctly discriminate among five benign and five malignant neoplasms of the appendicular skeleton using age, sex, and 17 radiographic characteristics (26). A naive Bayesian approach used two clinical and 16 qualitative radiographic features to diagnose 29 bone tumor types based on data of 1664 cases. This system correctly identified the diagnosis in 44% of cases and listed the correct condition among the three most likely conditions in 60% of cases (27). These examples show how Bayesian networks can be applied in settings with relatively few cases from which to learn. Unlike neural network approaches, which generally require large datasets for training, Bayesian networks can apply knowledge from the medical literature.
Breast Imaging
Breast imaging reports contain highly structured data that are particularly suitable for diagnostic decision support. The MammoNet system incorporated seven clinical features and 15 mammographic findings to predict the presence of breast cancer. It achieved an area under the receiver operating characteristic curve of 0.88 on 67 test cases (28,29). Burnside et al (19) analyzed a database of more than 48 000 screening and diagnostic mammograms, matched with outcomes data from a breast cancer registry, to create a tree-augmented naive Bayes model of breast cancer risk that achieved an area under the receiver operating characteristic curve of 0.96. A natural language processing tool that extracted Breast Imaging Reporting and Data System (BI-RADS) descriptors from textual mammography reports applied this Bayesian network model to predict BI-RADS assessment category with 98% accuracy, and the model was shown to reduce false-positive interpretations by 29% (30,31). Another Bayesian network model of 16 epidemiologic and clinical characteristics, morphologic MRI features, and quantitative MRI parameters was developed to predict the risk of triple-negative breast cancer (32).
Neurologic Imaging
Bayesian networks have been applied in a variety of neuroscience and neuroimaging studies, including resting-state functional MRI studies to predict states of belief and disbelief and early diagnosis of Alzheimer disease from structural MRI (33–35). Bayesian networks have been used in an array of multimodal MRI analyses for limited but complex diagnostic tasks in neuroradiology, including differentiating among atypical meningiomas, glioblastomas, and metastases (36); predicting glioma grade (37); and assisting in segmentation of multiple sclerosis lesions (38).
Advances in deep learning, such as convolutional neural networks, have enabled automated image analysis methods that make possible fully automated diagnostic systems to go from raw images to differential diagnoses. Recent studies explored Bayesian networks for diagnosis of 19 diseases involving the cerebral hemispheres (39) and 36 diseases involving deep gray matter (40). These efforts generated differential diagnoses with high accuracy by combining deep learning to detect and characterize abnormalities on multimodal brain MR images with Bayesian networks to integrate these imaging abnormalities with clinical features. These systems achieved performance similar to that of subspecialists and better performance than that of general radiologists and neuroradiology fellows. The approach accurately models the steps taken by radiologists when generating an image-based differential diagnosis, namely recognizing imaging abnormalities and then integrating imaging findings with clinical context.
Radiology Education
Because Bayesian networks can invoke the causal connections and probability values to explain their reasoning, researchers have explored Bayesian networks in radiology education. The Bayesian Network Tutoring and Explanation (BANTER) system tutored users in diagnosis and selection of optimal diagnostic procedures. It computed posterior probabilities based on given evidence, determined the best diagnostic procedure to affirm or exclude a potential diagnosis, quizzed users on their selection of optimal diagnostic procedures, and generated explanations of its reasoning (16,41,42). The Adaptive Radiology Interpretation and Education System, or ARIES, allowed users to interact with Bayesian network models to compute probabilities of disease, identify the features most strongly associated with a given diagnosis, and highlight the most discriminative features through sensitivity analyses (43,44).
Discussion
Advantages of Bayesian Networks
Bayesian networks offer several advantages and can be used instead of or in concert with other radiology AI approaches, such as deep neural networks, decision trees, support vector machines, and clustering algorithms. Bayesian networks can apply a variety of highly efficient algorithms to learn a model's structure and/or conditional probability parameters directly from data. It can also apply a network's probabilistic knowledge to determine the most likely diagnosis, identify the most useful diagnostic test, and determine the sensitivity of the model's conclusions to its variables. Bayesian networks can make accurate inferences from relatively small datasets and can perform well in the setting of missing data. Although convolutional neural networks and other deep learning algorithms have been applied to numerous imaging tasks, such as detection and diagnosis of abnormalities, a frequent concern is their inability to explain their reasoning (13,45). In addition, other AI learning methods often require large datasets—on the order of 1000 or 10 000 cases—to achieve useful results (46). Recent work combining deep learning to detect findings and Bayesian networks to relate those findings to diagnoses can tap the advantages of both technologies (39,40). Some key features of Bayesian networks are summarized schematically in Figure 2.
Bayesian network models can be updated continuously to learn from experience and play a role in learning health systems. The ability to directly examine a Bayesian network's probability tables can help assure that the model is performing as expected (47). Researchers have extended the capabilities of Bayesian networks into models that evolve over time and integrated them into logic-based models for automated reasoning. A dynamic Bayesian network, a model that repeats the static interactions of a conventional Bayesian network over time, has been developed to predict lung cancer risk from clinical and CT findings (17,48,49).
Several efforts have sought to integrate the probabilistic reasoning of Bayesian networks with other AI frameworks, such as logic-based knowledge representation and neural network learning mechanisms. A generalized logic-programming approach constructs a Bayesian network to answer queries based on a context-sensitive probabilistic knowledge base (50). Ontologies apply a logic-based framework to enable automated reasoning through the relationships defined between concepts (51). Several groups have explored approaches to augment ontologies with Bayesian networks’ probabilistic knowledge and ability to reason under uncertainty (52–54). Bayesian neural networks consist of a stochastic artificial neural network trained using Bayesian inference (55).
Limitations of Bayesian Networks
Bayesian networks have several limitations. Unlike deep neural networks, Bayesian networks cannot be applied as directly to tasks in computer vision and natural language processing. An important limitation is the need to discretize continuous variables; automated approaches for discretization pose a technical challenge, and the only continuous distribution that Bayesian networks can work with is a conditional Gaussian distribution (56). Furthermore, the scope of Bayesian networks has been limited, especially in biomedical applications. A recent review found that almost 60% of Bayesian-network health care applications were focused in only four areas: heart conditions, cancer, psychologic disorders, and lung disease (57). As outlined by Kyrimi et al (58), despite hundreds of articles published on Bayesian networks, there is still a large implementation gap, as Bayesian networks have rarely been applied in routine clinical practice.
Future Applications
One key opportunity for Bayesian networks lies in their integration with deep learning models, as has been shown for diagnosis in neuroradiology (39,40). Such hybrid approaches use the relative strengths of each technology. Deep neural networks can effectively recognize patterns and extract features, such as increased fluid-attenuated inversion recovery signal intensity in a certain spatial distribution, and Bayesian networks can interpret and integrate the extracted image features, in combination with relevant clinical features such as age, sex, and laboratory values, to provide an explainable diagnosis. For rare diseases, a particular advantage of Bayesian networks is that one can specify a model's probability tables with values gleaned from the medical literature rather than having to present thousands of examples to train a deep learning system.
Bayesian networks are poised to play an important role in learning health systems. Because a model's conditional probability tables can be updated directly based on its experience, accrual of real-life data will allow a Bayesian network to model its population more precisely and to automatically adjust its reasoning over time to match a changing environment. For example, once a specific diagnosis is identified, the patient's clinical data can then be inputted to refine the model. One can imagine collecting data on a larger scale from a health system's laboratory, pathology, and genomics databases to continually update probabilistic models for imaging diagnosis. Standardized data dictionaries, such as the Observational Medical Outcomes Partnership Common Data Model (59) and the radiology community's Common Data Elements (60), will support such efforts.
Conclusion
Bayesian networks continue to form an important AI approach in radiology. They offer powerful machine learning and decision support capabilities across a large number of applications. Their ability to learn from small datasets and provide explainable decisions make them especially useful in a medical field that integrates multimodal data for diagnosis and medical decision-making. One can combine Bayesian networks with other AI approaches. For example, one can use deep learning techniques to identify findings in images and use Bayesian networks to integrate those findings with other data and to explain the system's reasoning. There are opportunities for greater adoption of Bayesian networks in clinical practice.
Authors declared no funding for this work.
Disclosures of conflicts of interest: S.X.M. No relevant relationships. A.H.D. No relevant relationships. J.D.R. Research grant in artificial intelligence from the American Society of Neuroradiology Foundation; medical advisor to Galileo CDS; Radiology: Artificial Intelligence trainee editorial board alumni. A.M.R. Grants from Carestream Health/RSNA to institution; recipient of Research Scholar Grant ASNR Trainee Award; Radiology: Artificial Intelligence trainee editorial board alumni. D.R.W. No relevant relationships. P.H. Support from Mahidol University Office of International Relations; senior consultant on Radiology: Artificial Intelligence editorial board. C.E.K. Editor of Radiology: Artificial Intelligence; chair of the Radiology Informatics Committee, 2016–2019.
Abbreviations:
- AI
- artificial intelligence
- BI-RADS
- Breast Imaging Reporting and Data System
References
- 1. Pearl J . Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference . San Mateo, CA: : Morgan Kaufmann Publishers; , 1988. . [Google Scholar]
- 2. Kahn CE Jr . Artificial intelligence and radiology decision support . In: Hendee WR , Trueblood JH , eds. Digital Imaging . Madison, WI: : Medical Physics Publishing; , 1993. ; 281 – 300 . [Google Scholar]
- 3. Burnside ES . Bayesian networks: computer-assisted diagnosis support in radiology . Acad Radiol 2005. ; 12 ( 4 ): 422 – 430 . [DOI] [PubMed] [Google Scholar]
- 4. Kyrimi E , McLachlan S , Dube K , Neves MR , Fahmi A , Fenton N . A comprehensive scoping review of Bayesian networks in healthcare: Past, present and future . Artif Intell Med 2021. ; 117 : 102108 . [DOI] [PubMed] [Google Scholar]
- 5. Jensen FV . An introduction to Bayesian networks . New York: : Springer; , 1996. . [Google Scholar]
- 6. Jensen FV . Bayesian networks and decision graphs . New York: : Springer; , 2001. . [Google Scholar]
- 7. Mengshoel OJ . Understanding the scalability of Bayesian network inference using clique tree growth curves . Artif Intell 2010. ; 174 ( 12-13 ): 984 – 1006 . [Google Scholar]
- 8. Balke A , Pearl J . Probabilistic evaluation of counterfactual queries . https://escholarship.org/uc/item/6vh9k0cf. Published October 24, 2011. Accessed November 12, 2021.
- 9. Jensen FV , Liang J . drHugin. A system for value of information in Bayesian networks . http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.30.9692. Published 1994. Accessed December 2021.
- 10. Kjærulff U , van der Gaag LC . Making sensitivity analysis computationally efficient . arXiv 1301.3868 [preprint] https://arxiv.org/abs/1301.3868. Published January 16, 2013. Accessed December 2021. [Google Scholar]
- 11. Chan H , Darwiche A . Sensitivity analysis in Bayesian Networks: from single to multiple parameters . arXiv 1207.4124 [preprint] https://arxiv.org/abs/1207.4124. Published July 11, 2012. Accessed December 2021. [Google Scholar]
- 12. Castillo E , Gutierrez JM , Hadi AS . Sensitivity analysis in discrete Bayesian networks . IEEE Trans Syst Man Cybern A Syst Hum 1997. ; 27 ( 4 ): 412 – 423 . [Google Scholar]
- 13. Reyes M , Meier R , Pereira S , et al . On the interpretability of artificial intelligence in radiology: Challenges and opportunities . Radiol Artif Intell 2020. ; 2 ( 3 ): e190043 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Lacave C. Díez FJ . A review of explanation methods for Bayesian networks . Knowl Eng Rev 2003. ; 17 ( 2 ): 107 – 127 . [Google Scholar]
- 15. Suermondt HJ , Cooper GF . An evaluation of explanations of probabilistic inference . Comput Biomed Res 1993. ; 26 ( 3 ): 242 – 254 . [DOI] [PubMed] [Google Scholar]
- 16. Haddawy P , Jacobson J , Kahn CE Jr . BANTER: a Bayesian network tutoring shell . Artif Intell Med 1997. ; 10 ( 2 ): 177 – 200 . [DOI] [PubMed] [Google Scholar]
- 17. Koller D , Friedman N . Probabilistic graphical models: principles and techniques . http://mcb111.org/w06/KollerFriedman.pdf . Cambridge, MA: : MIT Press; , 2009. . [Google Scholar]
- 18. Cheng J , Greiner R , Kelly J , Bell D , Liu W . Learning Bayesian networks from data: An information-theory based approach . Artif Intell 2002. ; 137 ( 1-2 ): 43 – 90 . [Google Scholar]
- 19. Burnside ES , Davis J , Chhatwal J , et al . Probabilistic computer model developed from clinical data in national mammography database format to classify mammographic findings . Radiology 2009. ; 251 ( 3 ): 663 – 672 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. de Campos LM . A scoring function for learning Bayesian networks based on mutual information and conditional independence tests . J Mach Learn Res 2006. ; 7 : 2149 – 2187 . [Google Scholar]
- 21. Heckerman D , Geiger D , Chickering D . Learning Bayesian networks: The combination of knowledge and statistical data . Mach Learn 1995. ; 20 ( 3 ): 197 – 243 . [Google Scholar]
- 22. Provost F , Domingos P . Tree induction for probability-based ranking . Mach Learn 2003. ; 52 ( 3 ): 199 – 215 . [Google Scholar]
- 23. Tombropoulos R , Shiffman S , Davidson C . A decision aid for diagnosis of liver lesions on MRI . Proc Annu Symp Comput Appl Med Care 1993. ; 439 – 443 . [PMC free article] [PubMed] [Google Scholar]
- 24. Haddawy P , Kahn CE Jr , Butarbutar M . A Bayesian network model for radiological diagnosis and procedure selection: work-up of suspected gallbladder disease . Med Phys 1994. ; 21 ( 7 ): 1185 – 1192 . [DOI] [PubMed] [Google Scholar]
- 25. Reddan T , Corness J , Harden F , Hu W , Mengersen K . Bayesian approach to predicting acute appendicitis using ultrasonographic and clinical variables in children . Healthc Inform Res 2019. ; 25 ( 3 ): 212 – 220 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Kahn CE Jr , Laur JJ , Carrera GF . A Bayesian network for diagnosis of primary bone tumors . J Digit Imaging 2001. ; 14 ( 2 Suppl 1 ): 56 – 57 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Do BH , Langlotz C , Beaulieu CF . Bone tumor diagnosis using a naïve Bayesian model of demographic and radiographic features . J Digit Imaging 2017. ; 30 ( 5 ): 640 – 647 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Kahn CE Jr , Roberts LM , Shaffer KA , Haddawy P . Construction of a Bayesian network for mammographic diagnosis of breast cancer . Comput Biol Med 1997. ; 27 ( 1 ): 19 – 29 . [DOI] [PubMed] [Google Scholar]
- 29. Kahn CE Jr , Roberts LM , Wang K , Jenks D , Haddawy P . Preliminary investigation of a Bayesian network for mammographic diagnosis of breast cancer . Proc Annu Symp Comput Appl Med Care 1995. ; 208 – 212 . [PMC free article] [PubMed] [Google Scholar]
- 30. Bozkurt S , Gimenez F , Burnside ES , Gulkesen KH , Rubin DL . Using automatically extracted information from mammography reports for decision-support . J Biomed Inform 2016. ; 62 : 224 – 231 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31. Zeng J , Gimenez F , Burnside ES , Rubin DL , Shachter R . A probabilistic model to support radiologists’ classification decisions in mammography practice . Med Decis Making 2019. ; 39 ( 3 ): 208 – 216 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32. Huang Y , Zheng C , Zhang X , et al . The usefulness of Bayesian network in assessing the risk of triple-negative breast cancer . Acad Radiol 2020. ; 27 ( 12 ): e282 – e291 . [DOI] [PubMed] [Google Scholar]
- 33. Bielza C , Larrañaga P . Bayesian networks in neuroscience: a survey . Front Comput Neurosci 2014. ; 8 : 131 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Douglas PK , Harris S , Yuille A , Cohen MS . Performance comparison of machine learning algorithms and number of independent components used in fMRI decoding of belief vs. disbelief . Neuroimage 2011. ; 56 ( 2 ): 544 – 553 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Diciotti S , Ginestroni A , Bessi V , et al . Identification of mild Alzheimer's disease through automated classification of structural MRI features . Annu Int Conf IEEE Eng Med Biol Soc 2012. ; 2012 : 428 – 431 . [DOI] [PubMed] [Google Scholar]
- 36. Svolos P , Tsolaki E , Theodorou K , et al . Classification methods for the differentiation of atypical meningiomas using diffusion and perfusion techniques at 3-T MRI . Clin Imaging 2013. ; 37 ( 5 ): 856 – 864 . [DOI] [PubMed] [Google Scholar]
- 37. Tsolaki E , Svolos P , Kousi E , et al . Automated differentiation of glioblastomas from intracranial metastases using 3T MR spectroscopic and perfusion data . Int J CARS 2013. ; 8 ( 5 ): 751 – 761 . [DOI] [PubMed] [Google Scholar]
- 38. Herskovits EH , Bryan RN , Yang F . Automated Bayesian segmentation of microvascular white-matter lesions in the ACCORD-MIND study . Adv Med Sci 2008. ; 53 ( 2 ): 182 – 190 . [DOI] [PubMed] [Google Scholar]
- 39. Rauschecker AM , Rudie JD , Xie L , et al . Artificial intelligence system approaching neuroradiologist-level differential diagnosis accuracy at brain MRI . Radiology 2020. ; 295 ( 3 ): 626 – 637 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Rudie JD , Rauschecker AM , Xie L , et al . Subspecialty-level deep gray matter differential diagnoses with deep learning and Bayesian networks on clinical brain MRI: a pilot study . Radiol Artif Intell 2020. ; 2 ( 5 ): e190146 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41. Haddawy P , Jacobson J , Kahn CE Jr . An educational tool for high-level interaction with Bayesian networks . Proceedings of the 6th IEEE International Conference on Tools with Artificial Intelligence . Los Alamitos, CA: : IEEE Computer Society Press; , 1994. ; 578 – 584 . [Google Scholar]
- 42. Haddawy P , Jacobson J , Kahn CE Jr . Generating explanations and tutorial problems from Bayesian networks . Proc Annu Symp Comput Appl Med Care 1994. ; 770 – 774 . [PMC free article] [PubMed] [Google Scholar]
- 43. Chen PH , Botzolakis E , Mohan S , Bryan R , Cook T . Feasibility of streamlining an interactive Bayesian-based diagnostic support tool designed for clinical practice . SPIE Medical Imaging . San Diego, CA: : SPIE; , 2016. . [Google Scholar]
- 44. Duda J , Botzolakis E , Chen PH , et al . Bayesian network interface for assisting radiology interpretation and education . SPIE Medical Imaging . Houston, TX: : SPIE; , 2018. . [Google Scholar]
- 45. Soffer S , Ben-Cohen A , Shimon O , Amitai MM , Greenspan H , Klang E . Convolutional neural networks for radiologic images: A radiologist's guide . Radiology 2019. ; 290 ( 3 ): 590 – 606 . [DOI] [PubMed] [Google Scholar]
- 46. Alwosheel A , van Cranenburgh S , Chorus CG . Is your dataset big enough? Sample size requirements when using artificial neural networks for discrete choice analysis . Journal of Choice Modelling 2018. ; 28 : 167 – 182 . [Google Scholar]
- 47. Krumholz HM . Big data and new knowledge in medicine: the thinking, training, and tools needed for a learning health system . Health Aff (Millwood) 2014. ; 33 ( 7 ): 1163 – 1170 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48. van Gerven MA , Taal BG , Lucas PJ . Dynamic Bayesian networks as prognostic models for clinical patient management . J Biomed Inform 2008. ; 41 ( 4 ): 515 – 529 . [DOI] [PubMed] [Google Scholar]
- 49. Petousis P , Han SX , Aberle D , Bui AA . Prediction of lung cancer incidence on the low-dose computed tomography arm of the National Lung Screening Trial: A dynamic Bayesian network . Artif Intell Med 2016. ; 72 : 42 – 55 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50. Ngo L , Haddawy P . Answering queries from context-sensitive probabilistic knowledge bases . Theor Comput Sci 1997. ; 171 ( 1-2 ): 147 – 177 . [Google Scholar]
- 51. Schulz S , Jansen L . Formal ontologies in biomedical knowledge representation . Yearb Med Inform 2013. ; 8 : 132 – 146 . [PubMed] [Google Scholar]
- 52. Carvalho RN . Probabilistic ontology: representation and modeling methodology . Fairfax, VA: : George Mason University; , 2011. ; 343 . [Google Scholar]
- 53. da Costa PCG , Laskey KB , Laskey KJ . PR-OWL. A Bayesian ontology language for the Semantic Web . Uncertainty Reasoning for the Semantic Web I . Springer; , 2008. ; 88 – 107 . [Google Scholar]
- 54. Yang Y , Calmet J . OntoBayes: An ontology-driven uncertainty model . International Conference on Computational Intelligence for Modelling, Control and Automation and International Conference on Intelligent Agents, Web Technologies and Internet Commerce . Vienna, Austria: : IEEE; , 2005. ; 457 – 463 . [Google Scholar]
- 55. Jospin LV , Buntine W , Boussaid F , Laga H , Bennamoun M . Hands-on Bayesian neural networks – a tutorial for deep learning users . arXiv 2007.06823 [preprint] https://arxiv.org/abs/2007.06823. Updated January 3, 2022. Accessed June 2022. [Google Scholar]
- 56. Nojavan AF , Qian SS , Stow CA . Comparative analysis of discretization methods in Bayesian networks . Environ Model Softw 2017. ; 87 : 64 – 71 . [Google Scholar]
- 57. McLachlan S , Dube K , Hitman GA , Fenton NE , Kyrimi E . Bayesian networks in healthcare: Distribution by medical condition . Artif Intell Med 2020. ; 107 : 101912 . [DOI] [PubMed] [Google Scholar]
- 58. Kyrimi E , Dube K , Fenton N , et al . Bayesian networks in healthcare: What is preventing their adoption? Artif Intell Med 2021. ; 116 : 102079 . [DOI] [PubMed] [Google Scholar]
- 59. Overhage JM , Ryan PB , Reich CG , Hartzema AG , Stang PE . Validation of a common data model for active safety surveillance research . J Am Med Inform Assoc 2012. ; 19 ( 1 ): 54 – 60 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60. Kohli M , Alkasab T , Wang K , et al . Bending the artificial intelligence curve for radiology: Informatics tools from ACR and RSNA . J Am Coll Radiol 2019. ; 16 ( 10 ): 1464 – 1470 . [DOI] [PubMed] [Google Scholar]