Abstract
Artificial intelligence (AI) is catalyzing a new era in cardiovascular care. Innovations include the detection of latent signatures enabling early diagnosis, real-time nudges for optimized clinical decision-making, and digital twin simulations to personalize treatment. The integration of AI in health systems has the potential to improve patient outcomes and enhance the efficiency of care delivery across diverse populations. However, many challenges remain—including interoperability and data privacy, algorithmic fairness and sustainability, and system considerations such as workflow integration and policy. Here, we draw examples from state-of-the-art AI applications to examine the interplay between AI and learning health systems to augment continuous measurement and feedback, and implementation science to robustly evaluate the efficacy and safety of AI interventions deployed into clinical workflows. We share a vision for “AI in action”, transitioning cardiovascular health systems from sparse, reactive, and hospital-centric episodes to a multimodal, connected ecosystem continuously learning to improve patient outcomes.
Key words: artificial intelligence, implementation science, learning health system, precision medicine
Central Illustration
Highlights
-
•
AI-enabled technologies can act as a copilot for cardiovascular care, digitally enhancing perception, cognition, and action in a learning health system.
-
•
Implementation science principles can be applied to rigorously evaluate AI, establish guardrails against harmful learning, and operationalize tools for diffusion across cardiovascular care pathways.
-
•
We envision “AI in action” as the democratized implementation of trustworthy, adaptive AI within multimodal, connected ecosystems that continuously measure, learn, and improve cardiovascular outcomes.
implementing a (deep) learning health care system
Artificial intelligence (AI) systems have evolved to perform tasks once thought to require human cognition, including visual perception, understanding language, problem-solving through reasoning, and predicting outcomes from a range of actions. AI relies on pattern recognition algorithms that use vast data sets as a fundamental departure from older, rule-based engines. One form of AI, deep learning, particularly excels at identifying patterns in complex medical data.1 AI applications in health care are propelled by the large and growing amount of digital, multimodal data—electronic health records (EHRs), the digitalization of diagnostic testing, and connected devices such as wearables—that train and power these models.2 As such, AI seems a key and important tool in improving patient care and health system efficiency in cardiovascular care.3, 4, 5
However, a wide gap separates developments in health care AI from safe clinical adoption. For example, more than 40% of the AI-enabled devices cleared by the U.S. FDA (Food and Drug Administration) received regulatory authorization without clinical validation data,6 with recall rates exceeding 15%.7 In addition, rapid AI adoption globally has not followed implementation science best practices, which limits the broadest impact. More than 300 hospitals across China have installed on-site versions of the DeepSeek AI language model—integrated into diagnostics, decision support, and patient education—even as governance and regulatory approaches remain nascent.8 In the United Kingdom, 1 in 5 physicians reports using generative AI tools in clinical practice.9 Taken together, this depicts an urgent challenge, integrating best implementation practices to deploy AI safely and effectively for widespread clinical use in real-world settings.
A learning health system (LHS) is a concept that aims to improve health care by integrating scientific evidence directly into clinical practice, while simultaneously generating new knowledge from real-world care experiences.10 It is often described as a cyclical process where data are converted into knowledge, this knowledge is translated into practice, and changes in practice then generate more data, thus fostering continuous improvement.11 Accomplishing the goals of an LHS requires technological innovation, aligned organizational incentives,12 and a key focus on implementation science, which is a dynamic field dedicated to systematically studying methods to promote the integration of evidence-based practices, interventions, and policies into routine clinical settings.13 The fundamental purpose of implementation science is to bridge the gap between what is known from research and what is put into practice.14 This gap can be substantial; only 1 in 5 evidence-based interventions makes it to routine clinical practice, and the evidence generated from clinical research may take 17 years to be implemented.15 Applying LHS principles to AI implementation in cardiovascular care will involve prospective deployment, iterative evaluations of patient and process outcomes, and feedback loops that update models based on real-world performance and stakeholder experiences. This operationalizes implementation as a safe, effective, and scalable learning cycle. Several frameworks and comprehensive guidelines have been proposed for the implementation of AI in an LHS (Table 1).
Table 1.
Selected Theories, Models, and Frameworks Proposed to Bridge the Implementation Gap for AI in a Learning Health System
| Framework | Summary of Guidelines |
|---|---|
| FUTURE-AI16 | Fairness, universality, traceability, usability, robustness, and explainability. Defines 30 best practices addressing technical, clinical, and socioethical dimensions. |
| TRAIN-Europe17 | Trustworthy and responsible AI network. Aimed at enhancing the quality of health care AI while promoting international collaboration and democratizing responsible AI. Proposes 5 areas to assess a hospital’s preparedness for responsible AI: AI model inventory, real-world data testing, bias assessment, governance, adoption and scalability. |
| NAM AICC18 | National Academy of Medicine AI code of conduct. Aimed at harmonizing AI principles and mapping these to the National Academy of Medicine’s LHS shared commitments, with a focus on AI lifecycle, governance model, and stakeholder roles |
| ESC DCAI19 | European Society of Cardiology’s digital cardiology and AI committee’s roadmap. Ensuring data quality and integrity, standardizing disease definitions and outcomes, creating a robust evaluation framework |
| CFIR 2.020 | Consolidated framework for implementation research. Focuses on domains like implementation effectiveness, intervention characteristics, and settings. |
| RE-AIM21 | Reach, efficacy, adoption, implementation, and maintenance. Guides the evaluation of process and health outcomes. |
| Diffusion of innovations22 | Describes how a new practice spreads through a population, with key players being innovators, early adopters, early majority, late majority and laggards |
| TEHAI Framework23 | Translational evaluation of health care AI. Emphasizes capability, utility, and adoption, with 15 subcomponents addressing technical validity, ethical safeguards, generalizability, safety, privacy, and real-world integration. |
| AI-TAM24 | AI technology acceptance model. Extends the technology acceptance model with constructs for trust in AI, perceived AI output quality, and collaborative intention. |
AI = artificial intelligence; LHS = learning health system.
In this review, we use state-of-the-art cardiovascular AI applications as case studies to illustrate the key themes of this special issue—AI for augmenting an LHS and implementation science for robust evaluation of AI interventions (Central Illustration). The ultimate vision is of “AI in action”: maximizing the utility of AI in an LHS with democratized access to widely implemented and trustworthy medical AI.
Central Illustration.
Implementing a (Deep) Learning Health System for Cardiovascular Care
AI as a copilot for cardiovascular care
AI tools can automate measurement and reporting across both structured and unstructured data, uncovering latent patterns that capture current disease trends and forecast future risk. Together, these capabilities enable high-fidelity clinical data abstraction, system-wide real-time quality measurement, and a shift from reactive to proactive care within an LHS (Figure 1).
Figure 1.
AI as a Copilot for Cardiovascular Care
AI augments the learning health system across four layers—measurement, reporting, detection, and prediction. For measurement and reporting tasks, NLP, LLMs, and RAG convert unstructured EHR text and imaging into structured abstractions, and image-to-text models can auto-generate diagnostics, streamlining quality measurement and documentation. For detection, CNNs infer latent disease signatures from routine signals. For prediction, feed-forward and transformer models—alone or fused with multimodal data such as EHR and CMR—forecast clinical trajectories and deliver individualized risk assessments. Collectively, these capabilities shift cardiovascular care from reactive episodes to proactive, system-wide decision support. CMR = cardiac magnetic resonance; ECG = electrocardiogram; EHR = electronic health record.
Autocompletion of measurement and reporting tasks
To evaluate care outcomes, health systems rely on process and outcome measures that span individual patients to hospital-level aggregate measures. Measurements rely on structured data (eg, patient billing codes for conditions, medication records, and laboratory values) and unstructured data (eg, text from clinical notes and medical images). Although structured data are easy to query, extracting key clinical information from unstructured data requires manual abstraction, with personnel and infrastructure costing millions of dollars annually.25
AI systems can support these measurement tasks, aligning directly with the core objectives of an LHS, which aims to learn from daily experience and continuously improve care delivery.26 Specifically, AI algorithms can automate data collection from disparate sources, enabling health systems to gain a comprehensive, accurate view of factors influencing health and disease, while reducing the burden on health care professionals.10,11,18 In addition, AI algorithms can identify nonlinear relationships between these factors that traditional statistical methods might otherwise overlook1,27,28 and facilitate better tracking of long-term outcomes and cross-institute benchmarking,18,29 supporting an LHS’s goal of assessing and enhancing health quality.10
Advances in natural language processing (NLP) have enabled parsing of information that is traditionally captured in clinical notes. NLP further enhances the ability to not only extract words when they are mentioned but also provides broader inference from the context of the presented information. This is accomplished through neural networks with attention mechanisms to encode text into representations with meaning (semantic representations). These models are generally task-specific, such as identifying patients with aortic stenosis30 or subtypes of heart failure (HF),31, 32, 33, 34 quantifying disease severity (eg, grading the functional status35 or phenogroup36 of patients with HF), or determining clinical endpoints (eg, adjudicating HF hospitalizations37).
Newer implementations of language models include transformer-based large language models (LLMs), such as generative pre-trained transformer and Llama. These are general-purpose models trained on large and diverse corpora of text, building even more comprehensive semantic embeddings—numeric maps of how words and ideas relate. Using these maps, LLMs can generate representation-informed answers to tasks based on meaning and context, rather than simple keyword matches. LLMs have also been deployed to extract nonclinical data, such as social determinants of health, and to summarize clinical documentation.38,39 Although LLMs have been developed using trillions of bytes of data across the internet, they require explicit adaptation for health system applications, including fine-tuning on medical text, incorporating their use in the context of the broader available health system data, and being grounded in medical literature to enable their accurate use. This limits their scalability due to the large computational and operational systems expertise needed to tune the model for each health care system and task.
One approach to this limitation is the use of retrieval-augmented generation (RAG) techniques. RAG is a query-based enhancement of LLMs that can enable medical adaptation.40 RAG-LLMs store contextually-aware data embeddings that can be optimized for clinical text, allowing interoperable querying across database architectures. RAG-LLMs could be deployed across 2 different health system architectures to assess clinical quality measures such as stroke risk and bleeding risk in patients with atrial fibrillation (AF)40 or identify reasons where therapeutic compliance is determined by patient preference or other subjective features. In addition to improved sensitivity and accuracy over structured data, the RAG-LLM also provided reasoning for its decisions. RAG-LLMs have also demonstrated scalability across an entire health system to accurately identify sepsis clinical quality care measures at the point of care.41
Beyond the reporting of quality-of-care assessments, other AI applications can further enhance the workflow of clinicians. Computer vision, that is, AI tools for imaging, can automate the inference of human-readable labels, such as the detection of severe aortic stenosis from echocardiography,42 or the identification of major coronary vessels and regions of stenosis on coronary angiographic images.43 Furthermore, emerging innovations in image-to-text AI models can generate accurate, comprehensive reports from electrocardiograms (ECGs)44 or echocardiograms,45,46 streamlining diagnostic workflows. Together, these advances illustrate how AI can augment an LHS by alleviating reporting tasks and supporting health care workers.47,48 The automation of reporting tasks using AI technologies enables clinicians to refocus on direct patient care, potentially enhancing professional well-being and supporting a culture of continuous learning within an LHS.10
Superhuman AI for the detection of latent disease signatures
AI, particularly via implementations of deep learning, excels at identifying subtle disease signatures within massive, complex, and high-dimensional datasets, often beyond the perceptual limits of human experts.1,29 This “superhuman” capability to detect and predict disease status is poised to augment a health system as an engine for knowledge generation, enabling the extraction of meaningful insights from the vast stream of health data generated daily.11,28 Furthermore, by accelerating the bidirectional evidence-to-practice learning cycle,49 AI tools can potentially improve care delivery.27
There are specific implementations of deep learning, such as convolutional neural networks, that enable automated pattern recognition in images, applying cascades of learnable convolutional filters and pooling layers to transform pixel data into hierarchical representations of features within images. The inference of complex disease signatures can maximize the diagnostic utility of existing data streams and enable early detection. One example is the diagnosis of left ventricular systolic dysfunction (LVSD) via ECGs. Traditionally, LVSD is diagnosed based on mechanical measurements of ejection fraction from the echocardiogram. Recently, AI has demonstrated an ability to interpret subtle electrical signals on an ECG that correlate to LVSD.50 In addition, from a normal sinus rhythm ECG, it is possible for tools to discern latent signals of rhythm disturbances, such as AF or concealed long QT syndrome.51,52 Similarly, AI can assist with the diagnosis of cardiac amyloidosis by inferring signals from echocardiogram videos, including point-of-care ultrasound, differentiating from other causes of increased left ventricular thickness as well as capturing preclinical progression, paving the way for widespread screening.53, 54, 55, 56
AI algorithms can also be used for “opportunistic sensing”; extracting additional, clinically useful information from a diagnostic modality that was not the primary reason the data were collected, without additional testing. For example, signatures of atherosclerotic cardiovascular disease risk are detectable on chest X-rays,57 noncardiac computed tomography images,58 mammography scans,59,60 and retinal images.61 Some of the defining features, such as vascular calcification, are interpretable by expert human readers as well. However, the ability of AI tools to reliably identify on all testing, broadens the identification of diseases when competing diagnostic workflows may confound their detection in routine care.
Another “superhuman” capability of AI is the ability to forecast future clinical outcomes from a single, static diagnostic test. Convolutional neural network models trained on concurrent disease labels have displayed the ability to predict incident disease. For example, the application of an AI-ECG model trained to learn signatures of concurrent LVSD identified screen-positive individuals without HF at baseline linked with a >4-fold risk of developing LVSD in subsequent years.62 The same model also defined the risk of HF using a single baseline ECG, arguably identifying subtle structural anomalies that progress to HF in the future.50 Similarly, an AI-Echo model trained to detect concurrent aortic stenosis could also predict the development and progression of severe aortic stenosis from a normal echocardiogram video.63 In the future, AI-enhanced diagnostic modeling may harness multiple data sources and modalities in existing EHR data to uncover clinically actionable information.
Emerging AI model applications can also transform the whole temporal arc of a patient’s care into a digital narrative that can be used to infer a range of insights about their care and outcomes. An emerging class of AI models called transformers is particularly adept at encoding both structured EHR timelines and text into digital sequences that form foundational data layers for predictive risk models in an LHS.64 Such an application enables robust longitudinal health modeling and can forecast disease states.65 This formulation enables high-fidelity prediction of individualized disease trajectories, as well as supporting virtual trials and exploration of counterfactual interventions.66 The framework also extends to specialized tasks, such as aligning structured EHR elements with other data sources, such as imaging, to capture a multimodal representation of disease states. For example, such a unified framework leveraged the EHR and cardiac magnetic resonance (CMR) to predict the risk of arrhythmic death in hypertrophic cardiomyopathy, with superior discrimination over established clinical guidelines.67
Overall, AI models can support the realization of an LHS by converting existing clinical data points and free-text narratives into structured, forward-looking risk vectors.10,27 In addition, AI-enhanced early detection and trajectory prediction have the potential to reconfigure care pathways from reactive to proactive management.1,28,29 The combined impact may lead to improvements in important clinical endpoints—such as disability and mortality—as well as gains in workflow efficiency and patient experience.2,16
Digitally augmented perception and action in a learning health system
An LHS can be conceptualized as having a “sensory” arm and a “motor” arm, which together translate information to improved care (Figure 2).68 We discuss the sensing and nudging components of an LHS in detail below.
Figure 2.
Digitally Augmented Perception, Action, and Cognition in a Learning Health System
AI-enabled capabilities have the potential to augment the sensory, motor, and cognitive aspects of an LHS. Perception expands sensing through: 1) opportunistic sensing from routine tests and reports; 2) remote monitoring via wearables and implanted sensors that stream physiologic data; and 3) vocal biomarkers and ambient listening that convert clinician–patient dialogue into structured information. Action translates signals into decisions via: 1) digital nudges that support guideline-concordant therapy and quality improvement; 2) risk alerts that surface early deterioration to care teams; and 3) diagnostic dialogues that return individualized, evidence-grounded recommendations. Cognition closes the loop by: 1) simulation with digital twins for procedural planning and forecasting; 2) LLM-assisted trial recruitment; and 3) trial emulation and adaptive enrichment using real-world data to generate timely evidence.
Perception
Sensing in an LHS represents continuously capturing expansive, multimodal data from the health care environment.11,49,69 AI-enhanced cardiovascular innovations can broaden data streams for richer perception.68
An emerging sensory layer of an LHS combines voice recording capabilities with generative AI to streamline the workload of medical documentation.18 Ambient listening tools deploy LLMs on real-time clinician–patient dialogue to generate context-aware clinical notes and surface orders discussed for rapid verification.70 Their use shortens note-completion time, lessening time spent after-hours on documentation, and enables physicians to focus on the uniquely human elements of their profession.48 Consequently, ambient AI scribes have been shown to improve clinician efficiency and reduce administrative burden.39,71 In addition, AI-enhanced speech analysis can also be used as a “vocal biomarker”. For example, classifying patients with HF as “wet” or “dry” by identifying subtle voice alterations linked to pulmonary fluid overload72 and assessing the risk of HF hospitalization and mortality.73
AI-enabled extraction of data from wearable devices enables home-based assessment, bypassing the traditional framework of data collection at clinical encounters.10,27,29,74 The use of the single-lead ECG capability of wearable devices, such as the Apple Watch or Fitbit, paired with AI facilitates the detection of cardiovascular disease.75, 76, 77 Moreover, multimodal AI models can distil rich patterns in time series data to infer advanced physiological metrics that presently require specialized laboratory testing. AI-enhanced wearable sensors can be used to infer peak oxygen uptake78 or pulmonary capillary wedge pressure,79,80 enabling assessment in the community. A nationally representative survey of U.S. adults found that most would be willing to share wearable data with clinicians.81 On the other hand, recent qualitative work has shown that such data streams can be converted into commercial assets with little patient involvement.82 Scaling up to enable population health screening for cardiovascular disease in the community will require thoughtful governance.83
Remote patient monitoring using implanted devices is another way that the uptake of digital health technologies grows the sensory capabilities of an LHS.10,27,29,84 In recipients of implantable cardioverter defibrillators, AI-enabled analysis of continuous data revealed a sedentary, poor-sleep profile linked with significantly higher rates of malignant ventricular arrhythmias.85 Furthermore, a novel implantable sensor—measuring pulmonary artery pressure—significantly reduced HF hospitalizations by streaming daily hemodynamic data to clinicians.86,87 The use of remote patient monitoring has increased ten-fold among Medicare enrollees between 2019 and 2022,88 opening new avenues for timely clinician intervention while deepening patients’ engagement in their care.10,27
Action
The use of AI tools to enhance precise interventions can be conceptualized as a “motor” arm of an LHS,68 translating validated evidence into action via decision support or quality improvement.11,49,69 For example, AI can detect early signals of impending clinical deterioration, then trigger an alert as a clinical decision support tool.10 In a pragmatic, single-blind randomized clinical trial (RCT) conducted in emergency and inpatient departments, mortality risk inferred using an AI application for ECGs was associated with reductions in 90-day all-cause mortality when the AI risk prediction was visible to the treating clinician. Exploratory analyses attributed this benefit to earlier escalation of care: higher rates of admission to intensive care units, amiodarone use, and evaluation with echocardiography or N-terminal pro-B-type natriuretic peptide testing.89 In a pragmatic cluster-RCT spanning acute and intensive care units across 2 health systems, hourly AI alerts—derived from nursing-surveillance metadata—reduced in-hospital mortality when surfaced without interrupting workflows in the EHR to the care team. Further exploratory analyses suggest the survival benefit arose from earlier interventions for AI-detected high-risk patients.90
A “nudge” is an implementation strategy that influences behavior predictably without restricting choice.91 In cardiovascular care, electronic “nudges” use behavioral economic principles to steer physician or patient choices toward optimal care, for example, by defaulting to guideline-directed medical therapy (GDMT) in order sets.92 However, across multiple randomized trials and registries, embedding nudges into clinician decision support or patient messaging has achieved only modest gains in achieving uptake of optimal therapies and improving medication adherence in cardiovascular care.93, 94, 95, 96, 97, 98 There needs to be a focus on improving the quality and appropriateness of nudges to improve their impact.99 Individualized nudges could be supported by an AI-driven framework identifying the reasons for GDMT underutilization. For example, in patients with AF, a RAG-LLM can calculate stroke and bleeding risk scores from structured and unstructured EHR more rapidly and reliably than traditional approaches, as well as identifying reasons for nonadherence, including patient preferences or contraindications.40 Similarly, an AI-enabled opportunistic sensing approach can be used to improve GDMT rates in hypertension. By scanning echocardiogram reports, an NLP model can identify mentions of left-ventricular hypertrophy in patients not yet prescribed an antihypertensive agent—even when hypertension management was not the focus of the study or care setting—to then automatically notify clinicians and prompt therapy initiation.100
AI-enhanced analysis can also interpret continuous, multimodal physiologic data streams in real-time, enabling feedback loops.10 In a 1-year randomized trial of adults with both hypertension and type 2 diabetes, an AI algorithm that analyzed daily home blood pressure uploads together with continuous glucose monitor data led to hypertension remission (<140/90 mm Hg and no antihypertensive medication) in 50% of participants vs 0% under usual care.101 In the future, a clinician may be able to engage in an AI-enabled diagnostic dialogue, querying an LLM with comprehensive access to multimodal investigation reports, that return individualized recommendations and subspecialist-level differential diagnoses and management plans.102
Overall, AI can generate precise and individualized electronic nudges, translating new evidence into action in real-time. Widespread implementation will require clear guidance around clinical actionability16 and consideration of the lead-time over which the AI model is evaluated.103
Imaginative AI to forecast care and close the learning loop
Clinicians routinely imagine clinical trajectories, visualize complex procedures and reason through the evidence bases for competing therapies. Emerging AI approaches are poised to amplify these higher-order cognitive tasks, delivering high-fidelity digital simulations and rapid evidence generation across the LHS (Figure 2).
Simulation
The continuous improvement of care within an LHS can be advanced through digital twins, which are virtual representations of physical subjects that use real-time data to mirror the world experienced by the subject. In health care, digital twins can exist at various levels, ranging from individual patients to entire hospital systems, to simulate clinical scenarios, to inform medical decision-making, and forecast outcomes in real-time for the subject.104,105 For example, an abnormality is detected on a patient’s wearable device and fed into the patient’s digital twin that is stored in an LHS. The digital twin then simulates possible outcomes from interventions and alerts the patient’s clinician of the event and their recommendation.106 Notably, the substantial computational power required to develop digital twins—and the associated cost—must be balanced against equity as a key component of an LHS.
Applications of cardiovascular digital twins include simulation for procedural planning or guidance during the procedures. For instance, in AF ablation procedures, a digital twin model leverages CMR images to generate a 3D reconstruction of the atria, predefining fibrotic areas as targets for ablation, avoiding lengthy electrical mapping.107 In a randomized trial, combining AI-guided target ablation with standard pulmonary vein isolation achieved higher 1-year freedom from documented AF than pulmonary vein isolation alone.108 Digital twins can also be used to provide intraprocedural guidance for coronary angiography.109 Real-time guidance of percutaneous coronary intervention is delivered by superimposing a coronary roadmap onto live fluoroscopy images, reducing contrast volume, radiation dose, and procedure time.110 Another intraprocedural digital twin model uses angiographic images to simulate the physiological results of stent placement faster than measurement of fractional flow reserve.111
The ability of digital twins to simulate multiple patient characteristics in parallel permits granular and personalized forecasting of disease progression and modification.112 In patients with a myocardial infarction, construction of 3D heart models from CMR images enables estimation of arrhythmia risk, potentially reducing the rate of sudden cardiac death and also avoiding unnecessary implantable cardioverter defibrillator insertions.113 In diabetes care, a digital twin modeling glucose dynamics to automatically titrate insulin delivery in patients with type 1 diabetes reached its primary outcome of improved time in glucose range.114
Digital twins typically use multimodal data available in an LHS, with substantial complexity involved in model building and deriving individualized parameters. In cardiovascular care, for example, digital twins mechanistically model processes from cells to organ systems. Accordingly, most models to date have focused on single organ systems and have been limited to a few dozen individuals. Recently, neural networks and Gaussian process emulation have reduced the computational space to derive individual parameters, allowing for the creation of digital twins of a large cohort of participants.115 Although a majority of applications have focused on individual patients, system-level models are poised to have a major impact on health care workflows, efficiency, and patient outcomes.106 For example, a nation-wide digital twin of emergency call centers simulated call center distribution, cellular service availability, and call volume to optimize the service quality of response calls.116
Evidence generation
The application of generative AI to an LHS can also contribute to evidence generation through enhancing recruitment to clinical trials and emulating treatment effects in real-world populations. Historically, identifying eligible patients for clinical trials has been inefficient and costly, often relying on patient referrals from clinicians who may not be consistently aware or motivated.10 For example, a national registry found that <5% of eligible patients were enrolled in acute coronary syndrome trials.117
An LHS can improve inefficiencies in trial recruitment by leveraging EHRs to streamline patient identification and trial execution. LLM-based tools can facilitate the rapid screening of the EHR to identify eligible participants than manual review.118, 119, 120 Trialists could also foster uptake by applying less restrictive eligibility criteria, prospectively embedding implementation science principles of stakeholder engagement and workflow integration within the trial design.121 Indeed, AI-driven emulation of oncology trials using real-world EHR data has demonstrated that many eligibility rules can be safely relaxed, doubling the number of eligible patients while preserving estimated treatment effects.122 AI tools can also be used for more intelligent trial recruitment. For example, identifying potential participants with a higher mortality risk helps identify a population more likely to reach a mortality endpoint.123 Alternatively, a phenomapping approach embedded in an adaptive trial design learns responder signatures at interim analyses to subsequently enroll participants conditioned on predicted benefit, reducing the number of participants required to achieve the same treatment effect.124 Lastly, AI can also be used to augment clinical trials with in silico control populations.125 An LHS can provide the patient population for these innovations and, in turn, learn from their findings.
Given the cost and duration of RCTs, trial emulation in observational EHR data sets represents the next frontier for evidence generation.126 Targeted trial emulation takes advantage of the large sample sizes and rich longitudinal data available within an LHS, using advanced statistical techniques to balance confounders between treatment arms and estimate treatment effects.127, 128, 129
Finally, emerging applications of AI can enable inference from RCTs targeted for specific populations. Notably, RCT-measured treatment effects often generalize imperfectly to heterogeneous real-world populations, likely due to the under-representation of demographic groups enrolled in RCTs.130,131 Digital twins provide an AI-enabled way to perform trials in hard-to-enroll populations—such as children—by supplying virtual control arms and reducing exposure, cost, and time while preserving inferential power.132 Another reason for the limited generalizability of RCT treatment effects may be greater prognostic heterogeneity among real-world patients.133 Generative adversarial models such as RCT-Twin-GAN learn EHR covariate distributions to create a digital twin of the original trial, translating its treatment effects to real-world patients while preserving the estimated effect size.134 These synthetic cohorts are abstractions—performing model-based augmentation to reweight and translate trial effects under explicit assumptions to a specified target population—but provide some representation where none exists. Taken together, AI approaches may close the learning loop using newly generated data to refine and improve future care.2,49 Although the use of AI for evidence generation remains in its early stages, with limited impact on guidelines and clinical care so far, continued methodological research and application to important clinical questions is critical to realizing its full potential.
Guardrails against unwanted learning: maintaining fairness, privacy, and sustainability
AI algorithms can learn unintended patterns potentially leading to harm. For example, embedding systemic biases present in training data, enabling patient reidentification that compromises privacy, and drifting toward artifactual signals after deployment. In this section, we discuss these risks and outline guardrails for fair, privacy-preserving, and sustainable AI implementation (Figure 3).
Figure 3.
A Roadmap for Implementing “AI in Action”
A roadmap for “AI in action”, transforming cardiovascular health systems from sparse, reactive, hospital-centric episodes to a multimodal, connected learning ecosystem that continuously measures, learns, and improves cardiovascular outcomes. Guardrails ensure models learn the right signals and remain safe over time. Operationalization requires AI models to be generalizable, interoperable, and usable. Diffusion will require explainability, thoughtful regulation, and engaged stakeholders for scaling of responsible use. Together, these protections support the implementation of trustworthy AI within a learning health system, toward system-wide adoption and sustained improvement.
Fairness
A core principle for trustworthy AI is fairness, which mandates that AI tools maintain consistent performance across population subgroups.16 However, EHR data are inconsistent, incomplete, and prone to biases from underlying systemic issues.135 These stem from nonrepresentative sampling and systematically poorer data quality for historically oppressed and excluded groups,30,136 driven by inequitable health care access, socioeconomic disparities, and related structural factors.137,138 Moreover, most AI models are developed to optimize aggregate performance, so they can appear accurate globally, but still perform poorly for under-represented subgroups, further perpetuating health disparities.139 Therefore, it is essential to incorporate health equity frameworks to address this issue. Ideally, this should be addressed upstream by training AI tools on representative real-world data that adequately capture variations encountered in clinical practice.16 Practical steps will include lowering access barriers to health systems where most data are collected, broadening data collection to alternative sources such as wearables, and addressing mistrust about how health data are used for care and research.
Algorithmic bias is a major obstacle to the implementation of fair medical AI. One issue is “demographic shortcutting,” where deep learning models can base predictions on confounders, such as age, race, or ethnicity, instead of using task-relevant clinical features.140 The algorithmic encoding of demographic attributes produces quantifiable unfairness, for example, a higher false negative rate in certain age groups, reflecting substantial underdiagnosis.141 The reliance on demographic shortcuts may be suppressed during model training by specific technical adaptations. These include adding an extra output layer to predict demographic attributes and applying either an adversarial penalty141 or gradient reversal142 to negate demographic prediction while optimizing the model backbone. Post-training calibration of model thresholds can further equalize error rates across demographic groups.141 Optimal approaches account for bias remain uncertain and likely will vary across data sources and use cases.
The adoption of generative AI in an LHS may present an additional form of algorithmic bias. LLMs were presented with a series of emergency department cases, each presented in 32 variations with socioeconomic demographic changes, but clinical information held constant. Cases labeled as Black, unhoused, or LGBTQIA+ were more often routed to urgent care or mental health assessment than controls. In addition, there was an inflation of advanced imaging referrals for high-income labels, whereas low-income cases were steered toward basic or no testing. These disparities were not supported by guidelines, indicating algorithmic bias driven by demographic proxies that reflect structural inequities rather than clinical need.143 Future LLM workflows may address this via grounding—tying each AI model output back to verifiable evidence, such as radiology images, text notes, or clinical guidelines, allowing users to trace the provenance of the information18—and guardrails—layered controls designed to detect or block unsafe or ungrounded outputs from AI systems—forming a critical safety net.47
Biases with diagnostic labels are another key consideration.144 A widely used care-management algorithm meant to flag patients with complex needs was found to exhibit significant racial bias.145 Instead of identifying clinical need, it had learned to optimize for future health care costs as a proxy label, therefore systematically underestimating risk for Black patients, who often incur lower costs due to access and insurance barriers. Emerging solutions include using direct health measures—laboratory values, diagnoses, and clinical endpoints—as AI model targets rather than proxy labels that can embed structural equities.146 Fairness can also be explicitly assessed using frameworks such as probability-based checks; clinically meaningful, decision-analytic metrics is another strategy.147 Implementation science methods then provide guardrails for deployment, ensuring that the benefits of AI are equitably distributed across all patient populations and health care settings, especially in under-resourced areas.13
Data privacy
Public trust is essential for the successful implementation and sustained operation of an LHS, particularly in the AI era.2,148 This depends on strong data privacy and security frameworks. However, AI algorithms are data hungry. Richard Sutton, a founding father of reinforcement learning, wrote in 2019 of the “bitter lesson” made evident from 7 decades of AI research: expertly crafted, domain-specific models are outperformed over the long run by general-purpose algorithms, which simply harness more compute and data.149 AI scaling laws empirically demonstrate Sutton’s thesis, showing that model performance approximates power-law gains as the volume of training data and compute grow.150 Data scaling is the clear trajectory for medical AI,151 as reflected by a recent report of a generative model granted access to EHR data from 57 million patients.152
Data privacy approaches based on simple deidentification are fragile, as linking “anonymized” health records with basic demographic information is sufficient to reidentify individuals.153,154 Federated learning circumvents this by sharing model parameters across sites, rather than raw data.155 Federated models may be augmented with differential privacy, a technique that injects carefully calibrated noise so that the presence or absence of any single patient’s data cannot be inferred.156 Another solution may be provided by generating synthetic data using generative adversarial networks or probabilistic graphical models to preserve real distributions and dependencies.157,158 Taken together, these AI-driven solutions to data privacy enable model fitting while minimizing patient identifiability and privacy exposure.
Adaptive learning
Unlike traditional models—which are static, rule-based, and require manual updates—AI models can adapt to maintain or improve performance postdeployment.2,26,29,49 As patient demographics, clinical workflows, and data-capture practices drift over time, the population distribution that was used for model development does not match reality, which can result in the degradation of model performance and risk harm.18 Although the adaptive property of AI modeling is well-suited to an LHS principle of continuous learning,148,159 there are a number of challenges. Sensing the need for a model update requires differentiating real and artifactual performance degradation, with the latter arising in the setting of the model appropriately modifying care pathways. This may require the development of novel counterfactual analysis methods to assess performance accurately.160 Another challenge is deciding when to implement model updates. Implementation science outlines the need for continuous monitoring, periodic auditing, and iterative updating of AI systems postdeployment.17 For example, the “tight-loose-tight” framework, adapted for health care AI, explicitly includes a “monitor and report outcomes” phase to ascertain if the intended results are achieved.18 An alternative to this “freeze-the-model” approach may be a “dynamic deployment” framework, using continuous assessment of local performance.161 Lastly, further research is required to determine how to implement updates. A simulation study indicated that, once an intervention is triggered, retraining or codeploying additional models degraded model specificity at fixed sensitivity, meaning no universal update rule existed.162 The complete lifecycle of an AI model must likely be tracked to sustain performance, as well as ensuring traceability and trust.16
Operationalizing AI in the learning health system
A spectrum of implementation challenges shapes whether AI will deliver value in cardiovascular care. To operationalize AI, the central questions are as follows: will the model generalize? can it operate with local data and systems? and will clinicians adopt it? This section outlines AI universality, interoperability, and usability as determinants of success in an LHS (Figure 3).
Universality
The generalizability of AI models to new populations, clinical sites, and equipment remains a key challenge for integration into an LHS.2 Implementation science calls for robust, prospective validation in diverse, real-world clinical workflows. However, most AI algorithms fall short of this standard, despite showing promise in retrospective studies.163 Recent progress includes external benchmarking of AI-ECG models for LVSD detection164 and AI-Echo models for amyloidosis.165 However, more evidence to support the effectiveness, safety, and value of AI applications is needed, while ensuring that evidentiary standards for these applications are comparable to other screening or intervention tools. For example, in an analytical model study, initiating anticoagulation in patients with wearable-detected subclinical AF led to negligible benefit and substantially higher bleeding risk.166 Evaluations of costs and potential harms of this kind are essential for stakeholder trust. AI tools that fail to meet prespecified performance and utility thresholds should be retired.
AI systems can be exquisitely sensitive to modest, often imperceptible variations in input data. Accordingly, the intentional adaptation of models to account for local context may be a key feature of effective AI performance, rather than the pursuit of a single static “generalized” model. Input variations can arise from a range of factors, such as differences in measurement devices or data acquisition protocols167,168 or targeted adversarial attacks.169 One solution is to evaluate local clinical validity, then recalibrate or fine-tune the model if needed.16 This approach has been applied to locally calibrate AI-ECG models for hypertrophic cardiomyopathy170 and LVSD.171 Another solution is equipment-specific fine-tuning. For example, an AI-Echo model for detecting aortic stenosis preserved accuracy on CMR by applying a cross-modal adapter layer that reshaped CMR videos into the model’s native echo format.63 Lastly, preclinical silent trials, which run in the background without influencing clinical care, provide a novel way to evaluate AI tools in new populations. These trials have been endorsed by ethical AI implementation frameworks2 and may assist with successful downstream implementation.172
Interoperability
Interoperability remains a major challenge to the broad implementation of AI in cardiovascular care.11 Disparate systems store data in incompatible formats, structures, and vocabularies, so a digital tool developed in one system struggles to understand the information received from another.173 The Fast Healthcare Interoperability Resources standard stresses the need for rigorous syntactic and semantic definitions to promote the exchange of health care-related data.174 These include using standard languages for medical ontologies such as SNOMED (Systematized Nomenclature of Medicine) codes, common data models like OMOP (Observational Medical Outcomes Partnership), and interoperability protocols including DICOM (Digital Imaging and Communications in Medicine) and HL7 FHIR (Health Level Seven Fast Healthcare Interoperability Resources).16 In parallel, emerging “zero-shot” approaches, in which a transformer model can adapt to a new hospital’s data schema with minimal additional tuning, may provide an AI-based solution to data interoperability.175 Resolving fragmentation between noninteroperable systems could yield higher-quality more representative training data27 and accelerate the knowledge generation cycle10 across an LHS.
Usability
Effective AI deployment critically depends on usability and workflow integration.2,11 The optimization of these elements helps ensure that new technologies and evidence-based practices are adopted and demonstrably improve both patient care and clinician experience.16,29 Implementation science and human-computer interaction methods, such as rapid-cycle A/B testing, provide structured ways to evaluate usability and iteratively refine interventions.176 Implementation studies show that real-world pressures—such as time pressure, alert fatigue, and data overload—can erode the usability of a system.177,178 User feedback is therefore vital for refining workflow integration, ensuring that critical information is surfaced and delivered to patients and clinicians in their preferred format.179,180 Multifaceted implementation strategies that combine technology updates, staff training, and policy changes can also accelerate the adoption of evidence-based practices.181
AI introduces specific challenges, notably the need to reconcile algorithmic recommendations with clinician judgment. For example, in an implementation across 3 emergency departments, an AI-informed triage assistant increased sensitivity for identifying patients requiring critical care and improved patient flow. However, nurse agreement with the AI recommendation varied—high-agreement nurses outperformed the AI alone, whereas low-agreement nurses underperformed it—underscoring the need to calibrate the balance between artificial and human intelligence.182 Interviews with physicians and nurses using an AI-alert system for sepsis suggest a partnership model may be a good way to build trust over time while supporting clinician autonomy.183
Accelerating the diffusion of AI into clinical practice
For the diffusion of AI into cardiovascular care pathways, health systems must combine early ongoing stakeholder engagement; transparency in AI-guided decisions; and governance aligned with evolving frameworks for learning algorithms. Together, these pillars foster a culture of shared responsibility that enables trustworthy, adaptive implementation (Figure 3).
Explainability
Many AI models function as “black boxes”, offering little insight into how they reach a conclusion. This lack of transparency stymies adoption in evidence-based medicine, where decisions can carry life-or-death consequences. Consequently, explainability—the ability of a system to reveal clinically meaningful rationale for its predictions—has become a key research focus.16
Explainability approaches—such as SHapley Additive exPlanations, Local Interpretable Model-agnostic Explanations, and Gradient-weighted Class Activation Mapping—highlight the key input elements driving model predictions. For example, in AI-enabled analyses of clinical notes, these explainability methods can highlight the text features which are the most influential in the model’s decision-making.40 In AI-ECG applications, they can offer insight into which pixels or waveform segments are used by algorithms to detect disease.50,184 Although current explainability methods can aid system-level audit, they may be unreliable and can mislead at the patient level.185
As these approaches mature, they have the potential to create transparent feedback loops that improves clinician confidence and potentially even sharpens human diagnostic skills. Exposure to AI-ECG mortality scores improved cardiologists’ ability to identify high-risk ECGs.186 In another study, researchers clustered SHapley Additive exPlanations values for an LVSD model to derive 6 human-readable ECG patterns. After a short explainer video, this approach also raised cardiologists’ reading accuracy.187 Lastly, generative AI provided an interactive tool, synthesizing counterfactual ECGs to simulate the textbook waveform and rhythm changes associated with hyperkalemia and AF, thereby enhancing clinician trust.188
Regulatory and policy
The regulatory landscape for AI in health care is evolving rapidly.189 Traditional policy frameworks—built for static medical devices—are being reimagined to accommodate the continuous learning cycles enabled by AI tools embedded in an LHS.190 The FDA proposed Software as a Medical Device guidance191 defines an iterative evidence cycle: 1) valid clinical association between the software output and target condition, 2) technical validation proving the software processes inputs accurately and reliably, and 3) clinical validation confirming the output improves care in the intended population. Pertaining to AI-enabled software, the Software as a Medical Device framework emphasizes oversight of the entire lifecycle. This includes the need for postdeployment performance monitoring for drift and the use of a predetermined change control plan to streamline future model updates without necessitating full re-review. The National Academy of Medicine’s AI Code of Conduct reinforces this emphasis on continuous evaluation.18 However, adaptive accountability appears to remain the key bottleneck framework for safely implementing AI tools. A recent review of nearly 1000 FDA-cleared AI medical devices found that most postmarket recalls stemmed from software-performance defects and labeling errors, underscoring the need for continuous real-world surveillance to trigger timely model updates.7 Building empirical evidence may prove essential to shape AI governance frameworks, including policies that best harness commercial resources and expertise while maintaining safeguards to protect data privacy and clinician autonomy.
Stakeholder engagement
Successfully integrating AI into an LHS depends on meaningful engagement with stakeholders involved in care delivery and improvement processes.192, 193, 194 Key stakeholders—including clinicians, patients, administrators, payers, information technology vendors, regulators, and researchers—should collaboratively shape the responsible implementation of AI across health systems.195
Clinicians express concerns around over-reliance on AI systems, including ownership over decision-making, “automation bias”, and deskilling.183,196,197 A “human-in-the-loop” system, relying on expert human input to adjudicate the AI model, could help mitigate these concerns by surfacing high-stakes decisions for clinician review and quantifying model uncertainty.198 Scaling human-in-the-loop workflows, while keeping clinician workload manageable, may require automation of low-risk cases and defining clear escalation pathways. Patients are wary of the depersonalization of care interactions and data privacy risks.199, 200, 201 Conversely, patients are more accepting of transparent, clinician-supervised health care AI.202 Interestingly, AI tools can also be used to facilitate patient engagement, for example, by using LLMs to identify care priorities.203 Health care leaders and researchers focus on the organizational changes required for widespread AI adoption, including infrastructure readiness, workforce training, and ethical governance.204,205 Cost is another concern requiring consideration of funding, reimbursement models, and resource allocation.2,13,159 The concern about AI replacing the clinical workforce should be weighed against AI’s potential to democratize care amid physician shortages in the United States and worldwide.206
Stakeholder engagement should be integrated at every stage of the AI tool lifecycle.14 At the design stage, stakeholders can thoughtfully codesign AI applications in alignment with local priorities.196,207 At the implementation stage, the design of AI clinical workflows can be iteratively refined according to user feedback and real-world constraints.183 Across deployment and evaluation phases, ongoing collaboration between AI developers, clinicians, and patients is essential to ensure models remain performant, safe, and equitable as workflows and data evolve.208
AI in action
This state-of-the-art review outlines how AI can serve as a copilot for cardiovascular care within an LHS—from the automation of care outcome measures and diagnostic reporting tasks to advances in disease phenotyping and risk forecasting. The operationalization and diffusion of AI tools within an LHS requires the uniform application of core principles of evidence-based practice—explicitly addressing a range of logistical, system, and human factors—and guardrails to ensure efficient, safe, and effective care. Successful translation of AI into action demands the alignment of augmented capabilities with continuous learning via feedback loops to ensure sustained improvements in patient outcomes, health care quality, and equity.
Funding support and author disclosures
Dr Biswas is a co-inventor on U.S. patent applications 2025/0084489 and 2022/0136063. Dr Ogunniyi has received institutional research support grants from AstraZeneca, Boehringer Ingelheim, Cardurion Pharmaceuticals, and Pfizer, outside the submitted work. Dr Thomas Maddox reports serving as a consultant to Elion (advisory board member). Dr Ahmad has received research support from Pfizer, Atman Health, Tempus, AstraZeneca, and Abiomed; consulting fees from Alnylam Pharmaceuticals; and speaking honoraria from AstraZeneca and Alnylam Pharmaceuticals. Dr Khera is supported by the National Institutes of Health (R01AG089981, R01HL167858, K23HL153775) and the Doris Duke Charitable Foundation (Award 2022060), with additional support from the Blavatnik Foundation through the Blavatnik Fund for Innovation at Yale; he is an Associate Editor of JAMA; through Yale, he has received research support from Bristol Myers Squibb, BridgeBio, and Novo Nordisk; he is a co-inventor on patent applications WO2023230345A1, US20220336048A1, 63/484,426, 63/508,315, 63/580,137, 63/606,203, 63/619,241, and 63/562,335, and 63/346,610; and he is a cofounder of Evidence2Health, and Ensight-AI. The funders had no role in the preparation, review, or approval of the paper, and decision to submit the paper for publication. All other authors have reported that they have no relationships relevant to the contents of this paper to disclose.
Footnotes
The authors attest they are in compliance with human studies committees and animal welfare regulations of the authors’ institutions and Food and Drug Administration guidelines, including patient consent where appropriate. For more information, visit the Author Center.
References
- 1.Norgeot B., Glicksberg B.S., Butte A.J. A call for deep-learning healthcare. Nat Med. 2019;25:14–15. doi: 10.1038/s41591-018-0320-3. [DOI] [PubMed] [Google Scholar]
- 2.Steel P.A.D., Wardi G., Harrington R.A., Longhurst C.A. Learning health system strategies in the AI era. Npj Health Syst. 2025;2:1–9. [Google Scholar]
- 3.Khera R., Oikonomou E.K., Nadkarni G.N., et al. Transforming cardiovascular care with artificial intelligence: from discovery to practice: JACC state-of-the-art review. J Am Coll Cardiol. 2024;84:97–114. doi: 10.1016/j.jacc.2024.05.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Elias P., Jain S.S., Poterucha T., et al. Artificial intelligence for cardiovascular care-part 1: advances: JACC review topic of the week. J Am Coll Cardiol. 2024;83:2472–2486. doi: 10.1016/j.jacc.2024.03.400. [DOI] [PubMed] [Google Scholar]
- 5.Jain S.S., Elias P., Poterucha T., et al. Artificial intelligence in cardiovascular care-part 2: applications: JACC review topic of the week. J Am Coll Cardiol. 2024;83:2487–2496. doi: 10.1016/j.jacc.2024.03.401. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Chouffani El Fassi S., Abdullah A., Fang Y., et al. Not all AI health tools with regulatory authorization are clinically validated. Nat Med. 2024;30:2718–2720. doi: 10.1038/s41591-024-03203-3. [DOI] [PubMed] [Google Scholar]
- 7.Lee B., Patel S., Favorito C., Sandri S., Jennings M.R., Dai T. Development and commercialization pathways of AI medical devices in the United States: implications for safety and regulatory oversight. NEJM AI. 2025;2 [Google Scholar]
- 8.Zeng D., Qin Y., Sheng B., Wong T.Y. DeepSeek’s “low-cost” adoption across china’s hospital systems: too fast, too soon? JAMA. 2025;333:1866–1869. doi: 10.1001/jama.2025.6571. [DOI] [PubMed] [Google Scholar]
- 9.Blease C.R., Locher C., Gaab J., Hägglund M., Mandl K.D. Generative artificial intelligence in primary care: an online survey of UK general practitioners. BMJ Health Care Inform. 2024;31 doi: 10.1136/bmjhci-2024-101102. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Maddox T.M., Albert N.M., Borden W.B., et al. The learning healthcare system and cardiovascular care: a scientific statement from the American heart association. Circulation. 2017;135:e826–e857. doi: 10.1161/CIR.0000000000000480. [DOI] [PubMed] [Google Scholar]
- 11.Sheikh A., Anderson M., Albala S., et al. Health information technology and digital innovation for national learning health and care systems. Lancet Digit Health. 2021;3:e383–e396. doi: 10.1016/S2589-7500(21)00005-4. [DOI] [PubMed] [Google Scholar]
- 12.Brown A., Reid R.J. The surprising politics of learning health systems. Learn Health Syst. 2025;9 doi: 10.1002/lrh2.70008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Khan M.S., Rashid A.M., Van Spall H.G.C., et al. Integrating cardiovascular implementation science research within healthcare systems. Prog Cardiovasc Dis. 2025 doi: 10.1016/j.pcad.2025.04.005. [DOI] [PubMed] [Google Scholar]
- 14.Handley M.A., Gorukanti A., Cattamanchi A. Strategies for implementing implementation science: a methodological overview. Emerg Med J. 2016;33:660–664. doi: 10.1136/emermed-2015-205461. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Rubin R. It takes an average of 17 years for evidence to change practice-the burgeoning field of implementation science seeks to speed things up. JAMA. 2023;329:1333–1336. doi: 10.1001/jama.2023.4387. [DOI] [PubMed] [Google Scholar]
- 16.Lekadir K., Frangi A.F., Porras A.R., et al. FUTURE-AI: international consensus guideline for trustworthy and deployable artificial intelligence in healthcare. BMJ. 2025;388 doi: 10.1136/bmj-2024-081554. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.van Genderen M.E., Kant I.M.J., Tacchetti C., Jovinge S. Moving toward implementation of responsible artificial intelligence in health care: the european TRAIN initiative. JAMA. 2025;333:1483–1484. doi: 10.1001/jama.2025.1335. [DOI] [PubMed] [Google Scholar]
- 18.National Academy of Medicine . An Artificial Intelligence Code of Conduct for Health and Medicine: Essential Guidance for Aligned Action. The National Academies Press; 2025. [DOI] [PubMed] [Google Scholar]
- 19.Asselbergs F.W., Lüscher T.F. Trustworthy implementation of artificial intelligence in cardiology: a roadmap of the european society of cardiology. Eur Heart J. 2025;46:677–679. doi: 10.1093/eurheartj/ehae748. [DOI] [PubMed] [Google Scholar]
- 20.Damschroder L.J., Reardon C.M., Widerquist M.A.O., Lowery J. The updated consolidated framework for implementation research based on user feedback. Implement Sci. 2022;17:75. doi: 10.1186/s13012-022-01245-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Glasgow R.E., Vogt T.M., Boles S.M. Evaluating the public health impact of health promotion interventions: the RE-AIM framework. Am J Public Health. 1999;89:1322–1327. doi: 10.2105/ajph.89.9.1322. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Anon. Diffusion of innovations. By Everett M. rogers The free press of Glencoe division of the Macmillan co., 60 fifth avenue, New York 11, N. y., 1962. Xiii+367pp. 14×21cm. Price $6.50. J Pharm Sci. 1963;52:612. [Google Scholar]
- 23.Reddy S., Rogers W., Makinen V.-P., et al. Evaluation framework to guide implementation of AI systems into healthcare settings. BMJ Health Care Inform. 2021;28 doi: 10.1136/bmjhci-2021-100444. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Baroni I., Re Calegari G., Scandolari D., Celino I. AI-TAM: a model to investigate user acceptance and collaborative intention inhuman-in-the-loop AI applications. Hum Comput. 2022;9:1–21. [Google Scholar]
- 25.Saraswathula A., Merck S.J., Bai G., et al. The volume and cost of quality metric reporting. JAMA. 2023;329:1840–1847. doi: 10.1001/jama.2023.7271. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Angus D.C. Fusing randomized trials with big data: the key to self-learning health care systems? JAMA. 2015;314:767–768. doi: 10.1001/jama.2015.7762. [DOI] [PubMed] [Google Scholar]
- 27.Armoundas A.A., Narayan S.M., Arnett D.K., et al. Use of artificial intelligence in improving outcomes in heart disease: a scientific statement from the American Heart Association. Circulation. 2024;149:e1028–e1050. doi: 10.1161/CIR.0000000000001201. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Krumholz H.M. Big data and new knowledge in medicine: the thinking, training, and tools needed for a learning health system. Health Aff (Millwood) 2014;33:1163–1170. doi: 10.1377/hlthaff.2014.0053. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Naylor C.D. On the prospects for a (deep) learning health care system. JAMA. 2018;320:1099–1100. doi: 10.1001/jama.2018.11103. [DOI] [PubMed] [Google Scholar]
- 30.Biswas D., Wu J., Brown S., et al. Racial and ethnic disparities in aortic stenosis within a universal healthcare system characterized by natural language processing for targeted intervention. Eur Heart J Digit Health. 2025;18:2025. doi: 10.1093/ehjdh/ztaf018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Nargesi A.A., Adejumo P., Dhingra L.S., et al. Automated identification of heart failure with reduced ejection fraction using deep learning-based natural language processing. JACC Heart Fail. 2024;9:2024. doi: 10.1016/j.jchf.2024.08.012. [DOI] [PubMed] [Google Scholar]
- 32.Wu J., Biswas D., Ryan M., et al. Artificial intelligence methods for improved detection of undiagnosed heart failure with preserved ejection fraction. Eur J Heart Fail. 2024;26:302–310. doi: 10.1002/ejhf.3115. [DOI] [PubMed] [Google Scholar]
- 33.Cheema B., Mutharasan R.K., Sharma A., et al. Augmented intelligence to identify patients with advanced heart failure in an integrated health system. JACC: Adv. 2022;1 doi: 10.1016/j.jacadv.2022.100123. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Wu J., Biswas D., Brown S., et al. Artificial intelligence methods to detect Heart failure with preserved ejection fraction (AIM-HFpEF) within electronic Health records: an equitable disease detection model. Eur Heart J Digit Health. 2025 doi: 10.1093/ehjdh/ztaf107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Adejumo P., Thangaraj P.M., Dhingra L.S., et al. Natural language processing of clinical documentation to assess functional status in patients with heart failure. JAMA Netw Open. 2024;7 doi: 10.1001/jamanetworkopen.2024.43925. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Brown S., Soltani F., Wu J., et al. Artificial intelligence enabled phenogrouping of heart failure with preserved ejection fraction depicts early and end-stage trajectories. medRxiv. 2025 [Google Scholar]
- 37.Cunningham J.W., Singh P., Reeder C., et al. Natural language processing for adjudication of heart failure in a multicenter clinical trial: a secondary analysis of a randomized clinical trial. JAMA Cardiol. 2024;9:174–181. doi: 10.1001/jamacardio.2023.4859. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Guevara M., Chen S., Thomas S., et al. Large language models to identify social determinants of health in electronic health records. NPJ Digit Med. 2024;7:6. doi: 10.1038/s41746-023-00970-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Van Veen D., Van Uden C., Blankemeier L., et al. Adapted large language models can outperform medical experts in clinical text summarization. Nat Med. 2024;30:1134–1142. doi: 10.1038/s41591-024-02855-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Adejumo P., Thangaraj P., Shankar S.V., Dhingra L.S., Aminorroaya A., Khera R. Retrieval-Augmented generation for extracting CHA2DS2-VASc risk factors from unstructured clinical notes in patients with atrial fibrillation. medRxiv. 2024 doi: 10.1101/2024.09.19.24313992. [DOI] [Google Scholar]
- 41.Boussina A., Krishnamoorthy R., Quintero K., et al. Large language models for more efficient reporting of hospital quality measures. NEJM AI. 2024;1 doi: 10.1056/aics2400420. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Holste G., Oikonomou E.K., Mortazavi B.J., et al. Severe aortic stenosis detection by deep learning applied to echocardiography. Eur Heart J. 2023;44:4592–4604. doi: 10.1093/eurheartj/ehad456. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Yang S., Kweon J., Roh J.-H., et al. Deep learning segmentation of major vessels in X-ray coronary angiography. Sci Rep. 2019;9 doi: 10.1038/s41598-019-53254-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Khunte A., Sangha V., Oikonomou E.K., et al. Automated diagnostic reports from images of electrocardiograms at the point-of-care. medRxiv. 2024 doi: 10.1101/2024.02.17.24302976. [DOI] [Google Scholar]
- 45.Holste G., Oikonomou E.K., Tokodi M., Kovács A., Wang Z., Khera R. Complete AI-enabled echocardiography interpretation with multitask deep learning. JAMA. 2025 doi: 10.1001/jama.2025.8731. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Zhang J., Gajjala S., Agrawal P., et al. Fully automated echocardiogram interpretation in clinical practice. Circulation. 2018;138:1623–1635. doi: 10.1161/CIRCULATIONAHA.118.034338. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Maddox T.M., Embí P., Gerhart J., Goldsack J., Parikh R.B., Sarich T.C. Generative AI in medicine - evaluating progress and challenges. N Engl J Med. 2025 doi: 10.1056/NEJMsb2503956. [DOI] [PubMed] [Google Scholar]
- 48.Duggan M.J., Gervase J., Schoenbaum A., et al. Clinician experiences with ambient scribe technology to assist with documentation burden and efficiency. JAMA Netw Open. 2025;8 doi: 10.1001/jamanetworkopen.2024.60637. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Greene S.M., Reid R.J., Larson E.B. Implementing the learning health system: from concept to action. Ann Intern Med. 2012;157:207–210. doi: 10.7326/0003-4819-157-3-201208070-00012. [DOI] [PubMed] [Google Scholar]
- 50.Sangha V., Nargesi A.A., Dhingra L.S., et al. Detection of left ventricular systolic dysfunction from electrocardiographic images. Circulation. 2023;148:765–777. doi: 10.1161/CIRCULATIONAHA.122.062646. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Bos J.M., Attia Z.I., Albert D.E., Noseworthy P.A., Friedman P.A., Ackerman M.J. Use of artificial intelligence and deep neural networks in evaluation of patients with electrocardiographically concealed long QT syndrome from the surface 12-lead electrocardiogram. JAMA Cardiol. 2021;6:532–538. doi: 10.1001/jamacardio.2020.7422. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Attia Z.I., Noseworthy P.A., Lopez-Jimenez F., et al. An artificial intelligence-enabled ECG algorithm for the identification of patients with atrial fibrillation during sinus rhythm: a retrospective analysis of outcome prediction. Lancet. 2019;394:861–867. doi: 10.1016/S0140-6736(19)31721-0. [DOI] [PubMed] [Google Scholar]
- 53.Oikonomou E.K., Vaid A., Holste G., et al. Artificial intelligence-guided detection of under-recognised cardiomyopathies on point-of-care cardiac ultrasonography: a multicentre study. Lancet Digit Health. 2025;7:e113–e123. doi: 10.1016/S2589-7500(24)00249-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Slivnick J.A., Hawkes W., Oliveira J., et al. Cardiac amyloidosis detection from a single echocardiographic video clip: a novel artificial intelligence-based screening tool. Eur Heart J. 2025 doi: 10.1093/eurheartj/ehaf387. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Grogan M., Lopez-Jimenez F., Guthrie S., et al. Value of artificial intelligence for enhancing suspicion of cardiac amyloidosis using electrocardiography and echocardiography: a narrative review. J Am Heart Assoc. 2025;14 doi: 10.1161/JAHA.124.036533. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Oikonomou E.K., Sangha V., Vasisht S.S., et al. Artificial intelligence-enabled electrocardiography and echocardiography to track preclinical progression of transthyretin amyloid cardiomyopathy. Eur Heart J. 2025 doi: 10.1093/eurheartj/ehaf450. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Gallone G., Iodice F., Presta A., et al. Detection of subclinical atherosclerosis by image-based deep learning on chest x-ray. Eur Heart J Digit Health. 2025 doi: 10.1093/ehjdh/ztaf033. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Hughes-Austin J.M., Dominguez A., 3rd, Allison M.A., et al. Relationship of coronary calcium on standard chest CT scans with mortality. JACC Cardiovasc Imaging. 2016;9:152–159. doi: 10.1016/j.jcmg.2015.06.030. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Iribarren C., Chandra M., Lee C., et al. Breast arterial calcification: a novel cardiovascular risk enhancer among postmenopausal women. Circ Cardiovasc Imaging. 2022;15 doi: 10.1161/CIRCIMAGING.121.013526. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Allen T.S., Bui Q.M., Petersen G.M., et al. Automated breast arterial calcification score is associated with cardiovascular outcomes and mortality. JACC: Adv. 2024;3 doi: 10.1016/j.jacadv.2024.101283. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Li L.Y., Isaksen A.A., Lebiecka-Johansen B., et al. Prediction of cardiovascular markers and diseases using retinal fundus images and deep learning: a systematic scoping review. Eur Heart J Digit Health. 2024;5:660–669. doi: 10.1093/ehjdh/ztae068. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Dhingra L.S., Aminorroaya A., Sangha V., et al. Heart failure risk stratification using artificial intelligence applied to electrocardiogram images: a multinational study. Eur Heart J. 2025;13:2025. doi: 10.1093/eurheartj/ehae914. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Oikonomou E.K., Holste G., Yuan N., et al. A multimodal video-based AI biomarker for aortic stenosis development and progression. JAMA Cardiol. 2024;9:534–544. doi: 10.1001/jamacardio.2024.0595. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Kraljevic Z., Bean D., Shek A., et al. Foresight-a generative pretrained transformer for modelling of patient timelines using electronic health records: a retrospective modelling study. Lancet Digit Health. 2024;6:e281–e290. doi: 10.1016/S2589-7500(24)00025-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Shmatko A., Jung A.W., Gaurav K., et al. Learning the natural history of human disease with generative transformers. Nature. 2025 doi: 10.1038/s41586-025-09529-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Li Y., Wehbe R.M., Ahmad F.S., Wang H., Luo Y. Clinical-Longformer and Clinical-BigBird: transformers for long clinical sequences. arXiv. 2022 [Google Scholar]
- 67.Lai C., Yin M., Kholmovski E.G., et al. Multimodal AI to forecast arrhythmic death in hypertrophic cardiomyopathy. Nat Cardiovasc Res. 2025;4:891–903. doi: 10.1038/s44161-025-00679-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Khera R., Wiens J. Summertime for cardiovascular AI. Circ Cardiovasc Qual Outcomes. 2024;17 doi: 10.1161/CIRCOUTCOMES.123.010404. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Friedman C., Rubin J., Brown J., et al. Toward a science of learning systems: a research agenda for the high-functioning learning Health system. J Am Med Inform Assoc. 2015;22:43–50. doi: 10.1136/amiajnl-2014-002977. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70.Gandhi T.K., Classen D., Sinsky C.A., et al. How can artificial intelligence decrease cognitive and work burden for front line practitioners? JAMIA Open. 2023;6 doi: 10.1093/jamiaopen/ooad079. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Asgari E., Montaña-Brown N., Dubois M., et al. A framework to assess clinical safety and hallucination rates of LLMs for medical text summarisation. NPJ Digit Med. 2025;8:274. doi: 10.1038/s41746-025-01670-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Amir O., Abraham W.T., Azzam Z.S., et al. Remote speech analysis in the evaluation of hospitalized patients with acute decompensated heart failure. JACC Heart Fail. 2022;10:41–49. doi: 10.1016/j.jchf.2021.08.008. [DOI] [PubMed] [Google Scholar]
- 73.Maor E., Perry D., Mevorach D., et al. Vocal biomarker is associated with hospitalization and mortality among heart failure patients. J Am Heart Assoc. 2020;9 doi: 10.1161/JAHA.119.013359. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.Pedroso A.F., Khera R. Leveraging AI-enhanced digital health with consumer devices for scalable cardiovascular screening, prediction, and monitoring. NPJ Cardiovasc Health. 2025;2:34. doi: 10.1038/s44325-025-00071-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75.Khunte A., Sangha V., Oikonomou E.K., et al. Detection of left ventricular systolic dysfunction from single-lead electrocardiography adapted for portable and wearable devices. NPJ Digit Med. 2023;6:124. doi: 10.1038/s41746-023-00869-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.Dhingra L.S., Aminorroaya A., Pedroso A.F., et al. Artificial intelligence-enabled prediction of heart failure risk from single-lead electrocardiograms. JAMA Cardiol. 2025 doi: 10.1001/jamacardio.2025.0492. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Aminorroaya A., Dhingra L.S., Pedroso A.F., et al. Development and multinational validation of an ensemble deep learning algorithm for detecting and predicting structural heart disease using noisy single-lead electrocardiograms. Eur Heart J Digit Health. 2025;10:2025. doi: 10.1093/ehjdh/ztaf034. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78.Moayedi Y., Foroutan F., Gao Y., et al. Developments in digital wearable in Heart failure and the rationale for the design of TRUE-HF (Ted rogers understanding of exacerbations in Heart failure) apple CPET study. Circ Heart Fail. 2025;18 doi: 10.1161/CIRCHEARTFAILURE.124.012204. [DOI] [PubMed] [Google Scholar]
- 79.Klein L., Fudim M., Etemadi M., et al. Noninvasive pulmonary capillary wedge pressure estimation in heart failure patients with the use of wearable sensing and AI. JACC Heart Fail. 2025;13 doi: 10.1016/j.jchf.2025.102513. [DOI] [PubMed] [Google Scholar]
- 80.Franklin D., Tzavelis A., Lee J.Y., et al. Synchronized wearables for the detection of haemodynamic states via electrocardiography and multispectral photoplethysmography. Nat Biomed Eng. 2023;7:1229–1241. doi: 10.1038/s41551-023-01098-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Aminorroaya A., Dhingra L.S., Nargesi A.A., Oikonomou E.K., Krumholz H.M., Khera R. Use of smart devices to track cardiovascular health goals in the United States. JACC: Adv. 2023;2 doi: 10.1016/j.jacadv.2023.100544. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82.Spithoff S., Vesely L., McPhail B., Rowe R.K., Mogic L., Grundy Q. The primary care medical record industry in Canada and its data collection and commercialization practices. JAMA Netw Open. 2025;8 doi: 10.1001/jamanetworkopen.2025.7688. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83.Biswas D., Aminorroaya A., Croon P.M., Batinica B., Pedroso A.F., Khera R. Transforming population health screening for atherosclerotic cardiovascular disease with AI-enhanced ECG analytics: opportunities and challenges. Curr Atheroscler Rep. 2025;27:86. doi: 10.1007/s11883-025-01337-4. [DOI] [PubMed] [Google Scholar]
- 84.Pedroso A.F., Lin Z., Ross J.S., Khera R. National patterns of remote patient monitoring service availability at US hospitals. Circ Cardiovasc Qual Outcomes. 2025;18 doi: 10.1161/CIRCOUTCOMES.125.012034. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85.Kolk M.Z.H., Frodi D.M., Langford J., et al. Deep behavioural representation learning reveals risk profiles for malignant ventricular arrhythmias. NPJ Digit Med. 2024;7:250. doi: 10.1038/s41746-024-01247-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86.Abraham W.T., Adamson P.B., Bourge R.C., et al. Wireless pulmonary artery haemodynamic monitoring in chronic heart failure: a randomised controlled trial. Lancet. 2011;377:658–666. doi: 10.1016/S0140-6736(11)60101-3. [DOI] [PubMed] [Google Scholar]
- 87.Brugts J.J., Radhoe S.P., Clephas P.R.D., et al. Remote haemodynamic monitoring of pulmonary artery pressures in patients with chronic heart failure (MONITOR-HF): a randomised clinical trial. Lancet. 2023;401:2113–2123. doi: 10.1016/S0140-6736(23)00923-6. [DOI] [PubMed] [Google Scholar]
- 88.Anon . Office of Inspector General | Government Oversight. US Department of Health and Human Services; 2024. Additional oversight of remote patient monitoring in medicare is needed. [Google Scholar]
- 89.Lin C.-S., Liu W.-T., Tsai D.-J., et al. AI-enabled electrocardiography alert intervention and all-cause mortality: a pragmatic randomized clinical trial. Nat Med. 2024;30:1461–1470. doi: 10.1038/s41591-024-02961-4. [DOI] [PubMed] [Google Scholar]
- 90.Rossetti S.C., Dykes P.C., Knaplund C., et al. Real-time surveillance system for patient deterioration: a pragmatic cluster-randomized controlled trial. Nat Med. 2025:1895–1902. doi: 10.1038/s41591-025-03609-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 91.Patel M.S., Volpp K.G., Asch D.A. Nudge units to improve the delivery of health care. N Engl J Med. 2018;378:214–216. doi: 10.1056/NEJMp1712984. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 92.Breathett K., Lewsey S., Brownell N.K., et al. Implementation science to achieve equity in heart failure care: a scientific statement from the American heart association. Circulation. 2024;149:e1143–e1163. doi: 10.1161/CIR.0000000000001231. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 93.Ahmad F.S., Persell S.D. Nudging to improve cardiovascular care-clinicians, patients, or both. JAMA Cardiol. 2023;8:31–32. doi: 10.1001/jamacardio.2022.4382. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 94.Lama G., Yamamoto Y., Riello R.J., et al. Electronic alerts to improve heart failure therapy in outpatient practice. J Am Coll Cardiol. 2022;79:2203–2213. doi: 10.1016/j.jacc.2022.03.338. [DOI] [PubMed] [Google Scholar]
- 95.Ho P.M., Glorioso T.J., Allen L.A., et al. Personalized patient data and behavioral nudges to improve adherence to chronic cardiovascular medications: a randomized pragmatic trial. JAMA. 2025;333:49–59. doi: 10.1001/jama.2024.21739. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 96.Shah N.N., Ghazi L., Yamamoto Y., et al. Pragmatic Trial of messaging to providers about treatment of HyperLIPIDemia (PROMPT-LIPID): A randomized clinical trial. Circ Cardiovasc Qual Outcomes. 2024;17 doi: 10.1161/CIRCOUTCOMES.123.010335. [DOI] [PubMed] [Google Scholar]
- 97.Mukhopadhyay A., Reynolds H.R., Phillips L.M., et al. Cluster-randomized trial comparing ambulatory decision support tools to improve heart failure care. J Am Coll Cardiol. 2023;81:1303–1316. doi: 10.1016/j.jacc.2023.02.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 98.Adusumalli S., Kanter G.P., Small D.S., et al. Effect of nudges to clinicians, patients, or both to increase statin prescribing: a cluster randomized clinical trial. JAMA Cardiol. 2023;8:23–30. doi: 10.1001/jamacardio.2022.4373. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 99.Fox C.R., Doctor J.N., Goldstein N.J., Meeker D., Persell S.D., Linder J.A. Details matter: predicting when nudging clinicians will succeed or fail. BMJ. 2020;370 doi: 10.1136/bmj.m3256. [DOI] [PubMed] [Google Scholar]
- 100.Berman A.N., Hidrue M.K., Ginder C., et al. Leveraging preexisting cardiovascular data to improve the detection and treatment of hypertension: the NOTIFY-LVH randomized clinical trial. JAMA Cardiol. 2025 doi: 10.1001/jamacardio.2025.0871. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101.Shamanna P., Joshi S., Dharmalingam M., et al. Digital twin in managing hypertension among people with type 2 diabetes: 1-year randomized controlled trial. JACC: Adv. 2024;3 doi: 10.1016/j.jacadv.2024.101172. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 102.O’Sullivan J.W., Palepu A., Saab K., et al. Towards democratization of subspeciality medical expertise. arXiv. 2024 [Google Scholar]
- 103.Byrd T.F., 4th, Southwell B., Ravishankar A., et al. Validation of a proprietary deterioration index model and performance in hospitalized adults. JAMA Netw Open. 2023;6 doi: 10.1001/jamanetworkopen.2023.24176. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 104.Thangaraj P.M., Benson S.H., Oikonomou E.K., Asselbergs F.W., Khera R. Cardiovascular care with digital twin technology in the era of generative artificial intelligence. Eur Heart J. 2024 doi: 10.1093/eurheartj/ehae619. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 105.Sadée C., Testa S., Barba T., et al. Medical digital twins: enabling precision medicine and medical artificial intelligence. Lancet Digit Health. 2025;7 doi: 10.1016/j.landig.2025.02.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 106.Vallée A. Digital twin for healthcare systems. Front Digit Health. 2023;5 doi: 10.3389/fdgth.2023.1253050. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 107.Boyle P.M., Zghaib T., Zahid S., et al. Computationally guided personalized targeted ablation of persistent atrial fibrillation. Nat Biomed Eng. 2019;3:870–879. doi: 10.1038/s41551-019-0437-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 108.Deisenhofer I., Albenque J.-P., Busch S., et al. Artificial intelligence for individualized treatment of persistent atrial fibrillation: a randomized controlled trial. Nat Med. 2025;31:1286–1293. doi: 10.1038/s41591-025-03517-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 109.Aminorroaya A., Biswas D., Pedroso A.F., Khera R. Harnessing artificial intelligence for innovation in interventional cardiovascular care. J Soc Cardiovasc Angiogr Interv. 2025;4 doi: 10.1016/j.jscai.2025.102562. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 110.Piayda K., Phinicarides R., Afzal S., et al. Dynamic coronary roadmap in percutaneous coronary intervention: results from an open-label, randomized trial. JACC Cardiovasc Interv. 2021;14:2523–2525. doi: 10.1016/j.jcin.2021.08.068. [DOI] [PubMed] [Google Scholar]
- 111.Gosling R.C., Morris P.D., Silva Soto D.A., Lawford P.V., Hose D.R., Gunn J.P. Virtual coronary intervention: a treatment planning tool based upon the angiogram. JACC Cardiovasc Imaging. 2019;12:865–872. doi: 10.1016/j.jcmg.2018.01.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 112.Fisher C.K., Smith A.M., Walsh J.R., et al. Machine learning for comprehensive forecasting of alzheimer’s disease progression. Sci Rep. 2019;9 doi: 10.1038/s41598-019-49656-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 113.Arevalo H.J., Vadakkumpadan F., Guallar E., et al. Arrhythmia risk stratification of patients after myocardial infarction using personalized heart models. Nat Commun. 2016;7 doi: 10.1038/ncomms11437. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 114.Kovatchev B.P., Colmegna P., Pavan J., et al. Human-machine co-adaptation to automated insulin delivery: a randomised clinical trial using digital twin technology. NPJ Digit Med. 2025;8:253. doi: 10.1038/s41746-025-01679-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 115.Qian S., Ugurlu D., Fairweather E., et al. Developing cardiac digital twin populations powered by machine learning provides electrophysiological insights in conduction and repolarization. Nat Cardiovasc Res. 2025;4:624–636. doi: 10.1038/s44161-025-00650-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 116.Penverne Y., Martinez C., Cellier N., et al. A simulation based digital twin approach to assessing the organization of response to emergency calls. NPJ Digit Med. 2024;7:385. doi: 10.1038/s41746-024-01392-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 117.Udell J.A., Wang T.Y., Li S., et al. Clinical trial participation after myocardial infarction in a national cardiovascular data registry. JAMA. 2014;312:841. doi: 10.1001/jama.2014.6217. [DOI] [PubMed] [Google Scholar]
- 118.Unlu O., Shin J., Mailly C.J., et al. Retrieval-augmented generation–enabled GPT-4 for clinical trial screening. NEJM AI. 2024;1 [Google Scholar]
- 119.Unlu O., Varugheese M., Shin J., et al. Manual vs AI-assisted prescreening for trial eligibility using large language models-A randomized clinical trial. JAMA. 2025 doi: 10.1001/jama.2024.28047. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 120.Gupta S., Basu A., Nievas M., et al. PRISM: Patient records interpretation for semantic clinical trial matching system using large language models. NPJ Digit Med. 2024;7:305. doi: 10.1038/s41746-024-01274-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 121.Van Spall H.G.C., Desveaux L., Finch T., et al. A guide to implementation science for phase 3 clinical trialists: designing trials for evidence uptake. J Am Coll Cardiol. 2024;84:2063–2072. doi: 10.1016/j.jacc.2024.08.068. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 122.Liu R., Rizzo S., Whipple S., et al. Evaluating eligibility criteria of oncology trials using real-world data and AI. Nature. 2021;592:629–633. doi: 10.1038/s41586-021-03430-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 123.Jering K.S., Campagnari C., Claggett B., et al. Improving clinical trial efficiency using a machine learning-based risk score to enrich study populations. Eur J Heart Fail. 2022;24:1418–1426. doi: 10.1002/ejhf.2528. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 124.Oikonomou E.K., Thangaraj P.M., Bhatt D.L., et al. An explainable machine learning-based phenomapping strategy for adaptive predictive enrichment in randomized clinical trials. NPJ Digit Med. 2023;6:217. doi: 10.1038/s41746-023-00963-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 125.Popat S., Liu S.V., Scheuer N., et al. Addressing challenges with real-world synthetic control arms to demonstrate the comparative effectiveness of pralsetinib in non-small cell lung cancer. Nat Commun. 2022;13:3500. doi: 10.1038/s41467-022-30908-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 126.Makary M.A., Prasad V. Priorities for a new FDA. JAMA. 2025 doi: 10.1001/jama.2025.10116. [DOI] [PubMed] [Google Scholar]
- 127.Khera R., Aminorroaya A., Dhingra L.S., et al. Comparative effectiveness of second-line antihyperglycemic agents for cardiovascular outcomes: a multinational, federated analysis of LEGEND-T2DM. J Am Coll Cardiol. 2024;84:904–917. doi: 10.1016/j.jacc.2024.05.069. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 128.Suchard M.A., Schuemie M.J., Krumholz H.M., et al. Comprehensive comparative effectiveness and safety of first-line antihypertensive drug classes: a systematic, multinational, large-scale analysis. Lancet. 2019;394:1816–1826. doi: 10.1016/S0140-6736(19)32317-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 129.Wang S.V., Schneeweiss S., RCT-DUPLICATE Initiative, et al. Emulation of randomized clinical trials with nonrandomized database analyses: results of 32 clinical trials. JAMA. 2023;329:1376–1385. doi: 10.1001/jama.2023.4221. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 130.Reddy K.P., Faggioni M., Eberly L.A., et al. Enrollment of older patients, women, and racial and ethnic minority individuals in valvular heart disease clinical trials: a systematic review. JAMA Cardiol. 2023;8:871–878. doi: 10.1001/jamacardio.2023.2098. [DOI] [PubMed] [Google Scholar]
- 131.Cho L., Vest A.R., O’Donoghue M.L., et al. Increasing participation of women in cardiovascular trials: JACC council perspectives. J Am Coll Cardiol. 2021;78:737–751. doi: 10.1016/j.jacc.2021.06.022. [DOI] [PubMed] [Google Scholar]
- 132.Pammi M., Shah P.S., Yang L.K., Hagan J., Aghaeepour N., Neu J. Digital twins, synthetic patient data, and in-silico trials: can they empower paediatric clinical trials? Lancet Digit Health. 2025;7 doi: 10.1016/j.landig.2025.01.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 133.Orcutt X., Chen K., Mamtani R., Long Q., Parikh R.B. Evaluating generalizability of oncology trial results to real-world patients using machine learning-based trial emulations. Nat Med. 2025;31:457–465. doi: 10.1038/s41591-024-03352-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 134.Thangaraj P.M., Shankar S.V., Huang S., et al. A novel digital twin strategy to examine the implications of randomized clinical trials for real-world populations. medRxiv. 2024 doi: 10.1101/2024.03.25.24304868. [DOI] [PubMed] [Google Scholar]
- 135.Ahmad F.S., Chan C., Rosenman M.B., et al. Validity of cardiovascular data from electronic sources: the multi-ethnic study of atherosclerosis and HealthLNK. Circulation. 2017;136:1207–1216. doi: 10.1161/CIRCULATIONAHA.117.027436. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 136.Brown S., Biswas D., Wu J., et al. Race- and ethnicity-related differences in heart failure with preserved ejection fraction using natural Language processing. JACC: Adv. 2024;3 doi: 10.1016/j.jacadv.2024.101064. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 137.Lobb R., Colditz G.A. Implementation science and its application to population health. Annu Rev Public Health. 2013;34:235–251. doi: 10.1146/annurev-publhealth-031912-114444. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 138.Ogunniyi M.O., Commodore-Mensah Y., Ferdinand K.C. Race, ethnicity, hypertension, and heart disease: JACC focus seminar 1/9. J Am Coll Cardiol. 2021;78:2460–2470. doi: 10.1016/j.jacc.2021.06.017. [DOI] [PubMed] [Google Scholar]
- 139.Gianfrancesco M.A., Tamang S., Yazdany J., Schmajuk G. Potential biases in machine learning algorithms using electronic health record data. JAMA Intern Med. 2018;178:1544–1547. doi: 10.1001/jamainternmed.2018.3763. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 140.Gichoya J.W., Banerjee I., Bhimireddy A.R., et al. AI recognition of patient race in medical imaging: a modelling study. Lancet Digit Health. 2022;4:e406–e414. doi: 10.1016/S2589-7500(22)00063-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 141.Yang Y., Zhang H., Gichoya J.W., Katabi D., Ghassemi M. The limits of fair medical imaging AI in real-world generalization. Nat Med. 2024;30:2838–2848. doi: 10.1038/s41591-024-03113-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 142.Brown A., Tomasev N., Freyberg J., Liu Y., Karthikesalingam A., Schrouff J. Detecting shortcut learning for fair medical AI using shortcut testing. Nat Commun. 2023;14:4314. doi: 10.1038/s41467-023-39902-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 143.Omar M., Soffer S., Agbareia R., et al. Sociodemographic biases in medical decision making by large language models. Nat Med. 2025;31:1873–1881. doi: 10.1038/s41591-025-03626-6. [DOI] [PubMed] [Google Scholar]
- 144.Croon P.M., Dhingra L.S., Biswas D., Oikonomou E.K., Khera R. Phenotypic selectivity of artificial intelligence-enhanced electrocardiography in cardiovascular diagnosis and risk prediction. Circulation. 2025 doi: 10.1161/CIRCULATIONAHA.125.076279. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 145.Obermeyer Z., Powers B., Vogeli C., Mullainathan S. Dissecting racial bias in an algorithm used to manage the health of populations. Science. 2019;366:447–453. doi: 10.1126/science.aax2342. [DOI] [PubMed] [Google Scholar]
- 146.Hasanzadeh F., Josephson C.B., Waters G., Adedinsewo D., Azizi Z., White J.A. Bias recognition and mitigation strategies in artificial intelligence healthcare applications. NPJ Digit Med. 2025;8:154. doi: 10.1038/s41746-025-01503-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 147.Matos J., Van Calster B., Celi L.A., et al. Critical appraisal of fairness metrics in clinical predictive AI. arXiv. 2025 [Google Scholar]
- 148.Friedman C.P., Wong A.K., Blumenthal D. Achieving a nationwide learning health system. Sci Transl Med. 2010;2 doi: 10.1126/scitranslmed.3001456. [DOI] [PubMed] [Google Scholar]
- 149.Sutton R. The bitter lesson. https://www.cs.utexas.edu/∼eunsol/courses/data/bitter_lesson.pdf
- 150.Kaplan J., McCandlish S., Henighan T., et al. Scaling laws for neural language models. arXiv. 2020 [Google Scholar]
- 151.Waxler S., Blazek P., White D., et al. Generative medical event models improve with scale. arXiv. 2025 [Google Scholar]
- 152.Callaway E. Medical AI trained on whopping 57 million health records. Nature. 2025 doi: 10.1038/d41586-025-01422-3. [DOI] [PubMed] [Google Scholar]
- 153.Na L., Yang C., Lo C.-C., Zhao F., Fukuoka Y., Aswani A. Feasibility of reidentifying individuals in large national physical activity data sets from which protected health information has been removed with use of machine learning. JAMA Netw Open. 2018;1 doi: 10.1001/jamanetworkopen.2018.6040. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 154.Erlich Y., Shor T., Pe’er I., Carmi S. Identity inference of genomic data using long-range familial searches. Science. 2018;362:690–694. doi: 10.1126/science.aau4832. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 155.Goto S., Solanki D., John J.E., et al. Multinational federated learning approach to train ECG and echocardiogram models for hypertrophic cardiomyopathy detection. Circulation. 2022;146:755–769. doi: 10.1161/CIRCULATIONAHA.121.058696. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 156.Otoum Y., Nayak A. Differential privacy-driven framework for enhancing heart disease prediction. arXiv [csAI] 2025 [Google Scholar]
- 157.Yoon J., Drumright L.N., van der Schaar M. Anonymization through data synthesis using generative adversarial networks (ADS-GAN) IEEE J Biomed Health Inform. 2020;24:2378–2388. doi: 10.1109/JBHI.2020.2980262. [DOI] [PubMed] [Google Scholar]
- 158.Tucker A., Wang Z., Rotalinti Y., Myles P. Generating high-fidelity synthetic patient data for assessing machine learning healthcare software. NPJ Digit Med. 2020;3:147. doi: 10.1038/s41746-020-00353-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 159.McGinnis J.M., Fineberg H.V., Dzau V.J. Advancing the learning health system. N Engl J Med. 2021;385:1–5. doi: 10.1056/NEJMp2103872. [DOI] [PubMed] [Google Scholar]
- 160.Ansari S., Baur B., Singh K., Admon A.J. Challenges in the postmarket surveillance of clinical prediction models. NEJM AI. 2025;2 doi: 10.1056/aip2401116. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 161.Rosenthal J.T., Beecy A., Sabuncu M.R. Rethinking clinical trials for medical AI with dynamic deployments of adaptive systems. NPJ Digit Med. 2025;8:252. doi: 10.1038/s41746-025-01674-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 162.Vaid A., Sawant A., Suarez-Farinas M., et al. Implications of the use of artificial intelligence predictive models in health care settings : a simulation study. Ann Intern Med. 2023;176:1358–1369. doi: 10.7326/M23-0949. [DOI] [PubMed] [Google Scholar]
- 163.Longhurst C.A., Singh K., Chopra A., Atreja A., Brownstein J.S. A call for artificial intelligence implementation science centers to evaluate clinical effectiveness. NEJM AI. 2024;1 [Google Scholar]
- 164.Croon P.M., Boonstra M.J., Allaart C.P., et al. AI-ECG for LVSD detection: a systematic review and first-in-kind multinational head-to-head comparison. medRxiv. 2025 [Google Scholar]
- 165.Hourmozdi J., Easton N., Benigeri S., et al. Evaluating the performance and potential bias of predictive models for detection of transthyretin cardiac amyloidosis. JACC: Adv. 2025;4 doi: 10.1016/j.jacadv.2025.101901. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 166.Winstén A.K., Langén V., Airaksinen K.E.J., Teppo K. Net benefit of anticoagulation in subclinical device-detected atrial fibrillation. JAMA Netw Open. 2025;8 doi: 10.1001/jamanetworkopen.2025.8461. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 167.De Fauw J., Ledsam J.R., Romera-Paredes B., et al. Clinically applicable deep learning for diagnosis and referral in retinal disease. Nat Med. 2018;24:1342–1350. doi: 10.1038/s41591-018-0107-6. [DOI] [PubMed] [Google Scholar]
- 168.Huber F.A., Chaitanya K., Gross N., et al. Whole-body composition profiling using a deep learning algorithm: influence of different acquisition parameters on algorithm performance and robustness. Invest Radiol. 2022;57:33–43. doi: 10.1097/RLI.0000000000000799. [DOI] [PubMed] [Google Scholar]
- 169.Su J., Vargas D.V., Sakurai K. One pixel attack for fooling deep neural networks. IEEE Trans Evol Comput. 2019;23:828–841. [Google Scholar]
- 170.Lampert J., Bhatt D.L., Vaid A., et al. Calibration of ECG-based deep-learning algorithm scores for patients flagged as high risk for hypertrophic cardiomyopathy. NEJM AI. 2025;2 [Google Scholar]
- 171.Attia I.Z., Tseng A.S., Benavente E.D., et al. External validation of a deep learning electrocardiogram algorithm to detect ventricular dysfunction. Int J Cardiol. 2021;329:130–135. doi: 10.1016/j.ijcard.2020.12.065. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 172.Poterucha T.J., Jing L., Ricart R.P., et al. Detecting structural heart disease from electrocardiograms using AI. Nature. 2025 doi: 10.1038/s41586-025-09227-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 173.Ahmad F.S., Rasmussen L.V., Persell S.D., et al. Challenges to electronic clinical quality measurement using third-party platforms in primary care practices: the healthy hearts in the heartland experience. JAMIA Open. 2019;2:423–428. doi: 10.1093/jamiaopen/ooz038. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 174.Anon Index - FHIR v5.0.0. https://hl7.org/fhir/
- 175.Renc P., Jia Y., Samir A.E., et al. Zero shot health trajectory prediction using transformer. NPJ Digit Med. 2024;7:256. doi: 10.1038/s41746-024-01235-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 176.Horwitz L.I., Kuznetsova M., Jones S.A. Creating a learning health system through rapid-cycle, randomized testing. N Engl J Med. 2019;381:1175–1179. doi: 10.1056/NEJMsb1900856. [DOI] [PubMed] [Google Scholar]
- 177.Foley T., Vale L. A framework for understanding, designing, developing and evaluating learning health systems. Learn Health Syst. 2023;7 doi: 10.1002/lrh2.10315. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 178.Dixon-Woods M., Campbell A., Chang T., et al. A qualitative study of design stakeholders’ views of developing and implementing a registry-based learning health system. Implement Sci. 2020;15:16. doi: 10.1186/s13012-020-0976-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 179.Johnson A.E., Routh S., Taylor C.N., et al. Developing and implementing an mHealth heart failure self-care program to reduce readmissions: randomized controlled trial. JMIR Cardio. 2022;6 doi: 10.2196/33286. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 180.Allen L.A., Venechuk G., McIlvennan C.K., et al. An electronically delivered patient-activation tool for intensification of medications for chronic heart failure with reduced ejection fraction: the EPIC-HF trial. Circulation. 2021;143:427–437. doi: 10.1161/CIRCULATIONAHA.120.051863. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 181.Jaffe M.G., Lee G.A., Young J.D., Sidney S., Go A.S. Improved blood pressure control associated with a large-scale hypertension program. JAMA. 2013;310:699–705. doi: 10.1001/jama.2013.108769. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 182.Taylor R.A., Chmura C., Hinson J., et al. Impact of artificial intelligence–based triage decision support on emergency department care. NEJM AI. 2025;2 [Google Scholar]
- 183.Henry K.E., Kornfield R., Sridharan A., et al. Human-machine teaming is key to AI adoption: clinicians’ experiences with a deployed machine learning system. NPJ Digit Med. 2022;5:97. doi: 10.1038/s41746-022-00597-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 184.Al-Zaiti S.S., Martin-Gill C., Zègre-Hemsey J.K., et al. Machine learning for ECG diagnosis and risk stratification of occlusion myocardial infarction. Nat Med. 2023;29:1804–1813. doi: 10.1038/s41591-023-02396-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 185.Ghassemi M., Oakden-Rayner L., Beam A.L. The false hope of current approaches to explainable artificial intelligence in health care. Lancet Digit Health. 2021;3:e745–e750. doi: 10.1016/S2589-7500(21)00208-9. [DOI] [PubMed] [Google Scholar]
- 186.Raghunath S., Ulloa Cerna A.E., Jing L., et al. Prediction of mortality from 12-lead electrocardiogram voltage data using a deep neural network. Nat Med. 2020;26:886–891. doi: 10.1038/s41591-020-0870-z. [DOI] [PubMed] [Google Scholar]
- 187.Katsushika S., Kodera S., Sawano S., et al. An explainable artificial intelligence-enabled electrocardiogram analysis model for the classification of reduced left ventricular function. Eur Heart J Digit Health. 2023;4:254–264. doi: 10.1093/ehjdh/ztad027. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 188.Jang J.-H., Jo Y.-Y., Kang S., et al. A novel XAI framework for explainable AI-ECG using generative counterfactual XAI (GCX) Sci Rep. 2025;15:1–11. doi: 10.1038/s41598-025-08080-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 189.Morley J., Murphy L., Mishra A., Joshi I., Karpathakis K. Governing data and artificial intelligence for health care: developing an international understanding. JMIR Form Res. 2022;6 doi: 10.2196/31623. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 190.Palaniappan K., Lin E.Y.T., Vogel S., Lim J.C.W. Gaps in the global regulatory frameworks for the use of artificial intelligence (AI) in the healthcare services sector and key recommendations. Healthcare (Basel) 2024;12:1730. doi: 10.3390/healthcare12171730. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 191.Anon Software as a medical device (SaMD). US Food and Drug Administration. 2024. https://www.fda.gov/medical-devices/digital-health-center-excellence/software-medical-device-samd
- 192.Hogg H.D.J., Al-Zubaidy M., Technology Enhanced Macular Services Study Reference Group, et al. Stakeholder perspectives of clinical artificial intelligence implementation: systematic review of qualitative evidence. J Med Internet Res. 2023;25 doi: 10.2196/39742. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 193.Vo V., Chen G., Aquino Y.S.J., Carter S.M., Do Q.N., Woode M.E. Multi-stakeholder preferences for the use of artificial intelligence in healthcare: a systematic review and thematic analysis. Soc Sci Med. 2023;338 doi: 10.1016/j.socscimed.2023.116357. [DOI] [PubMed] [Google Scholar]
- 194.Robu D., Lazar J.B. European conference on knowledge management; Kidmore end. Academic Conferences International Limited; 2020. Shaping successful stakeholder engagement by design: digital transformation in healthcare; pp. 677–686. [Google Scholar]
- 195.Rozenblit L., Price A., Solomonides A., et al. Towards a multi-stakeholder process for developing responsible AI governance in consumer health. Int J Med Inform. 2025;195 doi: 10.1016/j.ijmedinf.2024.105713. [DOI] [PubMed] [Google Scholar]
- 196.Joshi M., Mecklai K., Rozenblum R., Samal L. Implementation approaches and barriers for rule-based and machine learning-based sepsis risk prediction tools: a qualitative study. JAMIA Open. 2022;5 doi: 10.1093/jamiaopen/ooac022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 197.Khera R., Simon M.A., Ross J.S. Automation bias and assistive AI: risk of harm from AI-driven clinical decision support. JAMA. 2023;330:2255–2257. doi: 10.1001/jama.2023.22557. [DOI] [PubMed] [Google Scholar]
- 198.Budd S., Robinson E.C., Kainz B. A survey on active learning and human-in-the-loop deep learning for medical image analysis. Med Image Anal. 2021;71 doi: 10.1016/j.media.2021.102062. [DOI] [PubMed] [Google Scholar]
- 199.Robertson C., Woods A., Bergstrand K., Findley J., Balser C., Slepian M.J. Diverse patients’ attitudes towards Artificial intelligence (AI) in diagnosis. PLoS Digit Health. 2023;2 doi: 10.1371/journal.pdig.0000237. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 200.Sangers T.E., Wakkee M., Kramer-Noels E.C., Nijsten T., Lugtenberg M. Views on mobile health apps for skin cancer screening in the general population: an in-depth qualitative exploration of perceived barriers and facilitators. Br J Dermatol. 2021;185:961–969. doi: 10.1111/bjd.20441. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 201.Young A.T., Amara D., Bhattacharya A., Wei M.L. Patient and general public attitudes towards clinical artificial intelligence: a mixed methods systematic review. Lancet Digit Health. 2021;3:e599–e611. doi: 10.1016/S2589-7500(21)00132-1. [DOI] [PubMed] [Google Scholar]
- 202.Moy S., Irannejad M., Manning S.J., et al. Patient perspectives on the use of artificial intelligence in health care: a scoping review. J Patient Cent Res Rev. 2024;11:51–62. doi: 10.17294/2330-0698.2029. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 203.Kim J., Chen M.L., Rezaei S.J., et al. Patient-centered research through artificial intelligence to identify priorities in cancer care. JAMA Oncol. 2025;24:2025. doi: 10.1001/jamaoncol.2025.0694. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 204.Satterfield K., Rubin J.C., Yang D., Friedman C.P. Understanding the roles of three academic communities in a prospective learning health ecosystem for diagnostic excellence. Learn Health Syst. 2020;4 doi: 10.1002/lrh2.10204. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 205.Sibbald M., Abdulla B., Keuhl A., Norman G., Monteiro S., Sherbino J. Electronic diagnostic support in emergency physician triage: qualitative study with thematic analysis of interviews. JMIR Hum Factors. 2022;9 doi: 10.2196/39234. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 206.Kohane I.S. Compared with what? Measuring AI against the health care we have. N Engl J Med. 2024;391:1564–1566. doi: 10.1056/NEJMp2404691. [DOI] [PubMed] [Google Scholar]
- 207.Hanneman K., Playford D., Dey D., et al. Value creation through artificial intelligence and cardiovascular imaging: a scientific statement from the American heart association. Circulation. 2024;149:e296–e311. doi: 10.1161/CIR.0000000000001202. [DOI] [PubMed] [Google Scholar]
- 208.Ling Kuo R.Y., Freethy A., Smith J., et al. Stakeholder perspectives towards diagnostic artificial intelligence: a co-produced qualitative evidence synthesis. EClinicalMedicine. 2024;71 doi: 10.1016/j.eclinm.2024.102555. [DOI] [PMC free article] [PubMed] [Google Scholar]





