Skip to main content
Lippincott Open Access logoLink to Lippincott Open Access
. 2025 Apr 24;38(5):21–30. doi: 10.1097/01.JAA.0000000000000204

Artificial intelligence in cardiovascular practice

Marci Farquhar-Snow 1, Amy E Simone 1, Sheel V Singh 1, Reamer L Bushardt 1
PMCID: PMC11984544  PMID: 40198000

ABSTRACT

Artificial intelligence (AI) is everywhere, but how is this expansive technology being used in cardiovascular care? This article explores common AI models, how they are transforming healthcare delivery, and important roles for clinicians, including advanced practice providers, in the development, adoption, evaluation, and ethical use of AI in cardiovascular care.

Keywords: advanced practice providers, AI, artificial intelligence, cardiovascular disease, machine learning, robotics


Figure.

Figure

No caption available.

Box 1.

Box 1

No caption available.

Artificial intelligence (AI) is everywhere these days, but how is this expansive technology being used in the healthcare of individuals with cardiovascular (CV) disease? What do advanced practice providers (APPs), including physician associates and advanced practice registered nurses, need to know about AI to ensure it advances safe, effective, and equitable care?

AI is a type of technology associated with computers and other machines that enables these devices to imitate or mimic human capabilities.1 Through adequate training, machines can identify recurring patterns and autonomously learn rules that facilitate their use in various clinical situations and tasks throughout the care continuum, from screening to diagnosis to management. Will AI replace healthcare professionals, such as APPs or physicians? Though this is a provocative question, AI is not fundamentally designed or developed to replace clinicians. It is designed, however, to repurpose roles, support complex analyses, and improve efficiency.2

This article describes and explores common AI models and how they are transforming healthcare delivery. It also outlines important roles for APPs in the development, adoption, evaluation, and ethical use of these AI models in CV care, pointing them to recent reports that highlight the utility and application of AI technologies in the care of individuals and populations with CV disease.3-6 Following its exploration of AI models, this article presents a novel schema for incorporating AI-infused technologies into the care continuum that considers patient perspectives, shared decision-making, and interprofessional collaborative practice. Two cases from clinical practice highlight how AI is currently being used in CV care, with considerations for how APPs are or can be involved in the medical, procedural, and surgical aspects of care. As AI-driven systems become more integrated into various workflows across healthcare organizations, it is crucial for APPs to understand these systems' capabilities, limitations, and challenges. Finally, we offer a call to action for APPs to be leaders and change agents in the ethical and responsible adoption of AI in clinical practice, where it can promote safe, effective, equitable, and high-quality care.

Box 2.

Box 2

No caption available.

Box 1 contains Case study #1, which provides a starting point for this article's discussion of AI models, how they can be incorporated into clinical practice, and ways in which they can promote enhanced patient care.

AI MODELS

Machine learning (ML)

ML is a subfield of AI that enables computers to analyze large sets of data autonomously, uncover patterns within datasets, apply patterns to new data, and execute computational tasks with greater efficiency than humans.7 ML can be used to reveal underlying disease mechanisms and enhance the accuracy of diagnosis, treatment, and risk prediction through its identification of clinically significant patterns or phenotypes that may not be readily discernible.8 For example, the management of heart failure (HF) often involves complex diagnostic and treatment options that require clinicians to sort through extensive, voluminous clinical information. Using traditional analytical methods, an APP encountering a patient with a new shortness of breath symptom may struggle to apply multiple algorithms and evidence-based research findings regarding specific CV diagnoses and non-CV comorbidities, such as hematologic or maternal health conditions, for precision patient management. Alternatively, by using ML to analyze large datasets, the APP could increase efficiency and accuracy in HF diagnosis and optimize the condition's management.9

ML algorithms can assist APPs in selecting an optimal drug therapy regimen. Jing and colleagues developed a population management strategy for patients with HF (N = 26,971) by harnessing large clinical datasets.10 Using Geisinger electronic health record (EHR) data, ML models were trained to predict 1-year all-cause mortality. The model identified actionable care gaps, meaning that the gaps could potentially be addressed or improved: these included influenza vaccination, BP of less than 130/80 mm Hg, A1C of less than 8%, cardiac resynchronization therapy, and active medications (angiotensin-converting enzyme inhibitor/angiotensin II receptor blocker/angiotensin receptor-neprilysin inhibitor, aldosterone receptor antagonist, hydralazine, and evidence-based beta-blocker).10 To determine if the identified care gaps were associated with higher mortality, ML methods were used to close these gaps (assuming the gaps were fixed in their model), while keeping all other variables unchanged. The change in risk score was calculated for the baseline study population and compared with the ML-adjusted model to estimate the reduction in estimated mortality and the number of lives that could be saved.10 Out of the 13,238 currently living patients in the study, 2,844 were predicted to die within 1 year; however, by applying ML to close the evidence-based care gaps, more than 230 of these lives were predicted to be saved. This study showed that ML has promising predictive performance that may outperform traditional logistic regression models.

There is an art to building reliable ML models. The process involves training algorithms to identify patterns in the data, extract relevant insights, and produce accurate predictions or classifications. This critical training step enables ML models to learn from data and generalize new knowledge that can be used in practice.

Deep learning (DL)

DL is a subset of ML that employs multiple layers of artificial neural networks to uncover or forecast patterns.11 DL is particularly advantageous in managing extensive data within the EHR with multiple data sources, since it can adapt with little to no human intervention. Multiple platforms that interpret echocardiograms, identify ECG heart rhythms, and detect cardiac diseases from surface ECGs can be used in risk prediction. Using these features, DL can generate new variables that are not in the training set to enhance efficiency for data analysis and transformation.9 For example, Golas and colleagues demonstrated that DL methods can surpass conventional techniques when predicting 30-day readmission rates among patients with HF.12 Kwon and colleagues also demonstrated that a DL algorithm could outperform existing scoring systems, such as the Get with the Guidelines-Heart Failure (GWTG-HF) score and Meta-Analysis Global Group in Chronic (MAGGIC) Heart Failure score, by accurately predicting both in-hospital and long-term mortality.13 Notably, clinicians may be hesitant to adopt an AI-recommended diagnosis if the rationale and evidence are unclear.14 To illustrate, Han and colleagues showed that imperceptible interference on ECG waveforms, known as adversarial examples, reduced the DL ECG rhythm classification accuracy from 88% to 26%.15

Model comprehension and clinical reasoning can improve with scale and prompt tuning, suggesting the potential utility of large language models (LLMs), such as Med-PaLM, for advanced reasoning capabilities in clinical practice. Some clinicians argue that these models remain inferior to more traditional methods of incorporating human interaction, which highlights the need for better evaluation frameworks and method development to ensure safe and effective use of LLMs in clinical settings.16

Limitations of ML

  • Clinical validation of ML methods would necessitate close collaboration with clinicians who understand the system and methods well enough to identify false claims/results. For example, a result may be technically accurate but not clinically useful or meaningful.

  • Poor inputs result in inaccurate data, which affect the quality of analysis and reliability of ML techniques.8 Models need to be developed using diverse populations, and the methods should be validated in an independent dataset with the goal of generating an unbiased, generalizable model.

  • ML and DL algorithms may encounter domain shift issues, where the diagnostic or predictive accuracy is compromised due to variations in data being used to train the computer, data collection methods, or data-capturing devices utilized by the organization. Therefore, external validation by expert clinicians and subject-matter experts plays a key role in ensuring reliability of AI systems.17 APPs managing patients across the spectrum of disease processes can offer important perspectives when these AI systems are being developed.

  • “Black box” models generated by AI may not be readily understandable by clinicians and/or the engineers who create them.18

  • Predictions generated by “black box” systems may be inaccurate due to flawed training data, which poses significant challenges.9 Repeatability and reproducibility of results by independent research teams are key.

Natural language processing (NLP)

NLP models aim to equip computers with the ability to comprehend and process human language in spoken or written format.19 NLP holds significant promise for streamlining data capture from the EHR that does not exist in discrete fields. NLP is also useful in generating patient education and optimizing provider-patient communication. Building upon this process, LLMs have the capability to produce language that is also contextually relevant to a specific population. LLMs can be used for streamlining administrative tasks and can, for example, enhance data collection methods for quality improvement projects or research. LLMs can also be used with population-specific, evidence-based clinical guidelines to help guide clinical decision-making.

Generative AI versus NLP

Generative AI models are more dynamic than NLP models and can create original content, such as distinctive and high-caliber images or unique educational materials that are focused on an individual's specific CV diagnosis. This capability is important for the digitization of EHRs, as generative AI models utilize NLP to create and condense patient notes, extract data from unstructured clinical text, and automate EHR data entry, thereby saving significant time for clinicians.9,19,20 Berman and colleagues used the open-source Canary NLP system to create a highly precise model and study its ability to detect five major CV comorbidities—hypertension, dyslipidemia, diabetes, coronary artery disease, and stroke/transient ischemic attack—within a vast EHR database against physicians' ability to extract such data manually.21 Utilizing the NLP models, the sentence- and note-level sensitivity, specificity, and positive predictive value for the modules were greater than 85%. Accuracy was highest for conditions for which diagnostic criteria are straightforward, such as diabetes and hypertension, and slightly lower for conditions involving more diagnostic complexity, such as myocardial infarction and embolic stroke.

Many APPs are involved in their organizations' data management for quality improvement and clinical research projects. Idnay and colleagues found that NLP systems are effective by impacting the time efficiency, workload demands, and recruitment process for clinical research.22 The systematic review of 11 studies showed that NLP systems can effectively exclude patients who do not meet eligibility criteria, achieving a 100% negative predictive value; although some ineligible patients were misclassified as eligible, eligible patients were not excluded.22 Seven of these studies noted that using NLP significantly reduced the time and workload required for eligibility prescreening with potential for additional reduction as the number of protocols and/or patients increase. Overall, the findings suggest that NLP systems are helpful in clinical research eligibility prescreening but require further investigation to determine their real-world applicability and reliability to produce meaningful data.

Limitations of NLP

  • Thorough responses to specialized diagnosis-specific queries are difficult to generate with some AI models due to their restricted grasp of domain-specific knowledge on which they have not been trained.19 For example, LLMs are generally trained on publicly accessible data that include minimal clinical documentation.

  • Training LLMs via the input of textbook-based instruction often fails to capture the intricacies of real-life patient scenarios (that is, the human element in the art of medicine).

  • NLP-generated suggestions for practice may fall short of clinical decisions generated through a team's interprofessional collaboration.19,21 Clinician accountability and well-being may be affected.

Rule-based expert systems (RBSs)

RBSs mimic human decision-making processes by using knowledge bases and predefined rules, generally defined as “if-then” statements, to diagnose diseases and recommend treatment. RBSs are best suited for straightforward tasks related to basic guidelines or standard practices. Chatbots often use RBSs. Chatbots are computer systems that detect keywords and language identifiers to trigger prewritten responses. If chatbots are equipped to use AI technologies such as ML, NLP, and RBSs, they can, for example, analyze symptoms from collected data, understand a clinician's question, and then automate a response. This combination of AI is well established in clinical practice.

RBSs offer transparency and interpretability, allowing clinicians to understand the RBS's decision-making process to validate its recommendations against established evidence-based clinical guidelines. Applying this concept, Brown and colleagues developed a straightforward RBS application that used a structured tool to integrate relevant established CV disease prevention guidelines and expert consensus statements to make related recommendations to clinicians in CV care of cancer survivors.23 Seto and colleagues adopted a comparable strategy in a distinct context.24 They developed an RBS for a HF mobile phone-based telemonitoring system that generated alerts and instructions based on the patient's weight, BP, heart rate, and symptoms.24 Findings of the trial suggested that the implementation of the rule set was linked to a notable improvement in self-care practices and the clinical management of HF.24

RBSs have the potential to detect HF among high-risk patients earlier using an AI-Clinical Decision Support System (AI-CDSS).8,25,26 For example, Choi and colleagues developed an AI-CDSS that achieved an accuracy rate of 98% for HF diagnosis, surpassing that of non-HF specialists (76%).25 Similarly, Attia and colleagues applied AI-CDSS to ECG data and found that it could serve as an effective screening approach for detecting left ventricular dysfunction, even in asymptomatic individuals.27 Boehmer and colleagues conducted the MultiSENSE study to create and validate a device-based diagnostic algorithm, HeartLogic, for forecasting HF events.28 Persons with implanted cardiac resynchronization therapy defibrillators were enrolled for chronic ambulatory data collection.28 The alert algorithm, which incorporated heart sounds, respiratory rate, tidal volume, heart rate, and physical activity, surpassed both primary endpoints, exhibiting 70% sensitivity and an unexplained alert rate of 1.47 per patient-year.28 Overall, the algorithm emerged as a sensitive and timely predictor of impending HF decompensation.28

Limitations of RBSs

  • Unlike ML that can learn from experiences, RBSs do not have a learning component; therefore, novel or complex scenarios would require manual intervention to incorporate new knowledge into the system.

  • RBSs lack the adaptability and learning elements of generative AI systems. Their functionality is thus limited by their predefined rules and database. This limitation would be problematic in complex CV care, which must adapt to clinical scenarios that do not consistently follow clear rules and patterns.

  • For reliability, in training chatbots, the rules need to be derived from data, books, research studies, and webpages that are vetted by subject-matter experts.

  • Hybrid approaches combining RBSs with ML aim to leverage transparency with adaptability for improved decision support.

Physical robots using AI

Integrating robotic and AI technologies offers numerous advantages in cardiology. Examples of applications include the early detection of CV risk, precise diagnosis of complex congenital structural heart abnormalities, tailored HF treatment, surgical precision during valve replacement, continuous monitoring for abnormal heart rhythms, and the ability to procure samples for predictive and prognostic analysis of CV disease.29-32 However, ethical considerations exist, such as the absence of human interaction, risk of errors and malfunctions, restricted accessibility, reliance on data accuracy, and regulatory and legal hurdles.29

The integration of AI and robotics can improve the speed and quality of care. AI use in robotic surgery is primarily performed by extensive data collection from sensors and images and can be categorized into two main areas: robotic autonomy and surgical assessment/feedback. For example, AI contributes to intelligent assistance and robotic autonomy during cardiac valve repair by enhancing the surgical field, recognizing native tissues, delineating instruments to use, providing tactile feedback, and automating steps progressively. ML aids in robotic surgical assessment and feedback by recognizing workflow patterns and gestures, assessing intraoperative conditions, and measuring surgical complexity; NLP systems can enhance realism in simulation environments when training APPs novice to CV practice.32,33

Robotic process automation (RPA) is defined as a software robot that mimics human actions, whereas AI is the simulation of human intelligence by machines. RPAs handle tedious, repetitive tasks such as billing, patient and staff scheduling, and patient records management to improve efficiency and accuracy. RPA applications include remote patient monitoring using implanted or wearable devices as well as simplification of insurance eligibility checks and claims processing.34,35 Automation assists organizations in predicting staffing and supply/equipment needs. RPA can be trained to compose clinical documentation from various sources such as EHRs, clinical assessments, or past correspondence and to utilize speech recognition NLP during consultations to ease the administrative workload for clinicians.36,37 Additionally, AI techniques have been suggested for extracting data from handwritten medical forms.38

Overall, robotic technology can enhance efficiency, accuracy, and consistency in healthcare delivery. The blending of AI and RPA technologies has the potential to enhance job satisfaction among APPs by allowing them to maintain focus on direct care tasks.39 While academic hospitals are exploring these innovations, it is necessary to perform trials that evaluate AI's impact on quality of care and utility to clinicians and care teams.40,41

Limitations of robots

  • Frontline workers have expressed concerns over robots and other technological advances, fearing potential drawbacks such as inaccurate diagnoses, compromised patient data privacy, and staff reductions under the guise of innovation and efficiency.

  • A commonly held belief is that AI technology cannot substitute for the expertise and care provided by human medical professionals. In a survey, Duke University physicians voiced concerns that AI models could worsen existing care issues, particularly relating to bias.41 One respondent emphasized the lack of understanding regarding how to assess algorithm performance, especially concerning certain racial and ethnic groups.

APPLICATION OF AI TO CLINICAL PRACTICE

Table 1 provides a summary of the AI models described throughout this article, and Figure 1 illustrates some of the ways in which these AI technologies can be incorporated into clinical practice for the benefit of the APP, the patient, and the overall healthcare system. Box 2 bookends the article's discussion of AI models and their application to APPs' clinical practice through its presentation of Case study #2, an additional exemplar examining use of AI in the care of a patient with CV disease.

TABLE 1. Summary of AI terminology2.

Type of AI Description Examples
ML AI subfield that enables computers to analyze data autonomously, uncover patterns within datasets, apply patterns to new data, and execute computational tasks with greater efficiency than humans Prediction models (readmission, CV risk, mortality, clinical trial screening eligibility)
DL ML subset that teaches computers to process data like the human brain by recognizing complex patterns in images, text, sounds, and other data to forecast patterns and make predictions Algorithm detection devices for atrial fibrillation or sudden cardiac death, prediction of clinical trial research outcomes
Generative AI Dynamic system capable of analyzing data, making predictions, and generating original content Distinctive, high-caliber images or educational materials that are unique to patient's diagnosis
NLP Computers with the ability to comprehend human language in spoken or written format Digitization of EHRs, comorbidity detection
LLM Artificial neural networks that learn language patterns, grammar, and context Translation of educational materials into different languages, compilation of multiple journal articles into a summary
RBSs Programs that mimic human decision-making processes to diagnose diseases and provide treatment recommendations using predefined rules and knowledge bases Treatment recommendations, early disease detection
Physical robots Robots used for surgical precision and continuous monitoring to enhance the ability for precise diagnosis and predictive analysis Surgical autonomy: automated dispensing cabinets, barcode medication administration, and closed-loop electronic medication management systems
RPA Enables AI to handle tedious, rule-based repetitive tasks Remote patient monitoring, insurance reimbursement systems, clinical note transcription

FIGURE 1.

FIGURE 1.

Application of AI in APP clinical practice

© Massachusetts General Hospital Institute of Health Professions 2025. The content in this original figure was conceived by Amy Simone with assistance from the other authors; content then underwent graphic design by Megan Wilcox (MGH Institute of Health Professions).

CHALLENGES IN APPLICATION OF AI ACROSS THE CARE CONTINUUM

The goals of AI use in healthcare are to predict or diagnose disease based on data input, customize treatment strategies, monitor individual outcomes, prognosticate trends across populations or disease-specific cohorts, and enhance provider efficiency, among other aims. Although use of AI is becoming more mainstream in healthcare delivery, APPs should consider several important, real-world issues when interacting with AI and integrating these technologies into clinical and administrative activities along the care continuum. Balagurunathan and colleagues have described several challenges in applying AI to patient care, such as the availability of curated datasets for model building, the ability of clinicians to interpret or comprehend the information produced by AI systems, and reliability and reproducibility of AI methods for routine clinical use.42 To expand upon common challenges and concerns involved with the application of AI, we offer a series of questions for APPs to consider.

Data acquisition

  • What resources are used to train the AI computer? Are they human- or computer-generated? How are training resources validated for reliability and accuracy?

  • How are data collected? Are the used data collection methods secure? Is the AI developer transparent regarding how the data are collected?

  • Does the technology use human- or machine-linked input? What is the potential margin for error/subjectivity?

  • What personnel/professions are involved in data collection/development of AI? Is the AI method appropriate for the topic being addressed or use case?

  • What is the potential for AI-related job displacements, organizational restructuring, or personnel retraining?

Techniques used by AI to acquire data may lack transparency, posing difficulties in understanding their reliability and accuracy. It may be unclear, for example, which data an AI tool is using, how representative those data are, and if biases are present. EHR data, for instance, may not be applicable to the general population, and their use as a data source in an AI tool could result in unjust treatment or erroneous decisions.43-45 Transparency can be achieved by explicit model definition followed by model validation.

AI may be limited in its ability to adopt a comprehensive approach to clinical decision-making, potentially overlooking the input of both psychological and social determinants of health. Therefore, it is important that clinicians know how AI models are developed to verify credibility of the outputs. Inaccurate predictions resulting from flawed training data used to develop the AI model pose significant reliability concerns. Overcoming these challenges necessitates collaboration among a variety of clinical and administrative professionals working with specific patient populations. Further, sustainable solutions will require collaboration among technology developers, researchers, policymakers, and clinicians to ensure the design and implementation of responsible, ethical, and safe AI methods and tools for clinical practice. Clinicians should investigate whether the AI technology will apply an algorithm that could compound existing inequities in their patient population.

Data recognition

  • What platforms are being used? ML, DL, NLP, RBSs, and/or RPA?

  • In which groups is the use of AI recognition models appropriate or inappropriate? Which cohorts/trends will be used in the data recognition?

  • Who has access to data? Patients, providers, healthcare organizations, insurance companies, researchers, patient advocacy organizations, etc.?

A nationwide call has been made for companies to set standards around the development of AI, but no definitive standards have been universally adopted. There is growing concern among healthcare organizations that, in the absence of regulation, some technology companies may prioritize profit over the privacy and safety of patients. Additional concern has been reported in the media over how AI can be weaponized by hackers and other bad actors.46 Health systems will need to consider the cost, complexity, and function of new AI technology and its implementation. Recently, comprehensive return on investment calculators have emerged that can guide health systems in assessing the financial and non-financial benefits of AI for patient management.47

Diagnosis

  • What algorithms were used to generate probable clinical diagnoses? Are those algorithms comprehensive and representative of the population being assessed?

  • What is the repeatability and reproducibility of the AI model's results or data output?

  • What cohorts/trends were used in the development of models to generate diagnoses or inform management decisions? Are they current?

AI can significantly enhance the speed and accuracy of CV risk prediction and disease detection. ML techniques currently used by many CV tests and procedures have a greater impact on an APP's ability to practice when DL methods are applied. For example, echocardiograms assess parameters like pressures and volumes that lead to various cardiomyopathies, ECGs detect life-threatening heart rhythms or left ventricular dysfunction, and cardiac implantable devices recognize impending acute HF exacerbation.27,48,49 Weng and colleagues demonstrated the usefulness of ML models in predicting CV disease development, with the most highly performing examined ML model correctly predicting an additional 7.6% of patients who ultimately developed CV disease over 10 years when compared with established risk prediction algorithms.50 In medical decision-making, APPs can bring together information gathered from patients, such as preferences and concerns, with clinical data to develop individualized, evidence-based treatment plans.

Management

  • How is information disseminated regarding management?

  • Is the AI system secure? Is it linked to an organization or a proprietary, national, or global database?

  • What algorithms are used in treatment recommendations? Are they proprietary or based on national or global agency guidelines? Do they incorporate insurance plan coverage considerations?

  • What actions does the tool take regarding dissemination and implementation?

  • How are outsourcing recommendations generated (for example, recommendations for surgical consult, case management, or palliative care)?

  • Will information or guidance be transmitted directly to patients or will input/validation from the provider be required first?

AI algorithms are intended to streamline administrative processes, optimize resource allocations, and perform general administrative operations such as billing, staffing, and composing clinical notes. Development of robust AI in these healthcare processes relies on the availability of precise, high-quality healthcare data. To achieve broad adoption, data will need to be shared widely, requiring deidentification and increased considerations for patient confidentiality and privacy.51

Patient privacy, engagement, and adherence

  • How are patient data integrated into AI algorithms and rules to generate patient-specific educational materials?

  • Will patient data be used to build future AI algorithms? How will patient identifiers be kept secure?

  • Does the AI method used create biases based on demographics or social determinants of health?

AI applications provide new avenues for personalized healthcare management or precision medicine. AI-driven systems have the capacity to produce tailored suggestions across a range of domains such as nutrition, mental health, and fitness; however, the validity and efficacy of these methods still require more thorough study.52-56 For example, Dergaa and colleagues investigated the efficacy of exercise prescriptions generated by OpenAI's Generative Pre-trained Transformer 4 (GPT-4) model and found that although the model was capable of generating a range of safety-conscious programs, it lacked precision in addressing the patient's personal health conditions and goals.57 AI strives to increase patient engagement and adherence by enhancing access to exercise prescriptions, diet prescriptions, and surgical and rehabilitation monitoring, but it is not yet globally recommended for replacing personalized and health condition-specific prescriptions provided by trained professionals.57,58 AI-based regimens tend to be conservative and lack specificity in recommendations that can impact a patient's motivation and adherence. Without real-time feedback and responsive program adjustments, which an APP can provide through virtual or in-person follow-up visits, these challenges are further compounded.57

Surveillance

  • How will trending be done regarding specific cohort-related Diagnosis-Related Group (DRG)/International Classification of Diseases (ICD)?

  • How will trending be used to monitor at the individual, provider, and practice group levels?

  • How will trending be used by the Centers for Medicare and Medicaid Services (CMS) and other insurance providers?

Ethical, social, and technical concerns often arise when applying AI techniques to clinical practice. Patient data security concerns often arise when AI uses EHR data. Ethically, AI systems need to preserve patient confidentiality while providing data to build artificial neural networks.59 There are perceived reliability and trust issues in AI-driven decision-making. Therefore, APPs must learn to use AI responsibly across all clinical settings to protect patient privacy while monitoring individual patient outcomes among population-specific pooled data.

Evaluation

  • What quality control measures are in place? Who will evaluate and perform quality control (for example, the AI proprietor or a national/governmental agency)?

  • Does the AI method used need adjustments to reflect current APP practices?

  • How will new AI methods be developed based on quality control recommendations?

The evaluation process is important for any new AI system to determine if the methods or algorithms work well for the intended purpose and if they are based on the most current evidence or studies available. For example, the inclusion of APPs on insurance group panels was delayed due to CMS being slow to recognize APPs as independent and/or as contributing providers of healthcare management. AI systems must continuously learn or be updated to reflect current national standards. Balagurunathan and colleagues offer a checklist of criteria, adapted below, that can be used to improve the reliability and transparency of AI methods:

  • ✓ Diverse cohort of patient records for model training, achieved either through use of centralized or federated/distributed learning models that use silos of different data sources

  • ✓ Use of independent data cohort for testing, preferably in a distributed setting with diverse patient types

  • ✓ Transparency of deep network model architecture with confidence levels in its decisions

  • ✓ Ethically appropriate use of AI methods with some level of oversight

  • ✓ Assessment of reproducibility of AI models with test-retest type studies

  • ✓ Model transparency that discloses the architecture, datasets, and trained weights for the network

  • ✓ Quality assurance program for implementation and continuous performance monitoring.42

CONCLUSION

AI systems can be used by APPs to augment their knowledge base for optimal clinical decision-making and to enhance communication between patients and the entire healthcare team. AI methods hold enormous promise in CV care and are already demonstrating value in driving efficiency, supporting complex analyses, generating predictive models, and empowering clinicians with more tools to screen, diagnose, and manage patients with increasingly personalized approaches.

AI methods are here to stay, so it is imperative for APPs to gain an understanding of AI technologies used in healthcare, recognize their potential benefits and limitations, and support the development and evaluation of reliable and reproducible models that add meaningful clinical value and efficiency to the care continuum. Fortunately, frameworks and protocols exist to help integrate equity, diversity, and inclusion principles and practices in the design, development, and implementation of AI technologies in healthcare.60 APPs have the opportunity to embrace and utilize AI in their clinical practice to optimize quality and safety of care. This article has provided examples and resources of how to incorporate AI into clinical practice. We recommend accessing “Navigating the Responsible and Ethical Incorporation of Artificial Intelligence into Clinical Practice,” a resource from the Federation of State Medical Boards, for further reading.61

BOX 1. Case study #1

This case study illustrates the aspects of AI models that can be incorporated into clinical practice for patients to navigate between facilities and providers. Below the case study, a list of ways in which AI can augment patient care is provided; the example (EX) codes cited throughout the case text refer to this list.

Case

Mrs. Smith is a 78-year-old Black woman who recently presented to an ED with chest pain and shortness of breath. She underwent an echocardiogram, which revealed significant aortic stenosis, and received supportive care. Further testing ruled out myocardial infarction or significant coronary artery disease. She was discharged home with instructions to follow up with her primary care provider (PCP).

Mrs. Smith resides in a rural community with limited access to advanced heart and vascular services. Her PCP referred her to a comprehensive valve center approximately 100 miles away that offers an echo surveillance application (EX1, EX2) as a quality initiative to ensure patients with significant valvular heart disease are referred for evaluation. The application assessed her aortic valve disease and, utilizing predictive analytics (EX3), indicated that she was likely to progress rapidly to critical aortic stenosis.

Mrs. Smith's PCP was notified (EX4) that she met criteria for significant aortic stenosis based on the latest American College of Cardiology (ACC)/American Heart Association (AHA) guidelines. Mrs. Smith received a phone call from an APP at the valve center explaining the circumstances and that her PCP had approved a referral (EX4) for further evaluation. Education materials were generated that were appropriate to Mrs. Smith's health literacy and education levels (EX5) and sent to her prior to the appointment. Mrs. Smith underwent a CT scan, which applied a predictive modeling application (EX3) and identified significant risks (EX6, EX7) for transcatheter aortic valve replacement. The patient was engaged in a shared decision-making conversation by the care team using tools personalized for Mrs. Smith's unique anatomical considerations and health literacy levels (EX8). Mrs. Smith ultimately underwent surgical aortic valve replacement with an excellent result.

Ways AI can augment care

  • EX1: Mitigate disparities

  • EX2: Improve access to care

  • EX3: Apply predictive analytics to diagnostic tests

  • EX4: Ensure swift referral and operational efficiencies

  • EX5: Provide personalized education resources

  • EX6: Strive to mitigate complications

  • EX7: Tailor treatment plans for optimal clinical quality and outcomes

  • EX8: Generate an individualized shared decision-making toolkit

BOX 2. Case study #2

This case study illustrates the aspects of AI models that can be incorporated into clinical practice. Below the case study, a list of ways in which AI can augment patient care is provided; the example (EX) codes cited throughout the case text refer to this list.

Case

Mr. Garcia is a 52-year-old Latino man referred to an advanced HF clinic for newly diagnosed nonischemic cardiomyopathy following a recent hospitalization for decompensated systolic congestive HF. AI flagged Mr. Garcia's chart for APP review (EX1) prior to his visit because the medical regimen Mr. Garcia was prescribed was identified as suboptimal. Prior to his appointment, a buccal swab collection kit and educational materials in his preferred language (Spanish) were sent to him so that a pharmacogenetic test could be performed (EX1, EX2, EX3, EX4). During the appointment, AI was implemented to translate Spanish to English in real time (EX4) and help construct the clinic note (EX5), allowing the APP to focus on building rapport and ensuring an ideal patient experience despite the language barrier. AI provided a personalized medical regimen of guideline-driven medical therapy (EX1), in view of Mr. Garcia's genetic profile and predicted levels of response. At the visit, his echocardiogram, ECG, and biomarkers were assessed by AI (EX2), which determined him to be at elevated risk for sudden cardiac death via ventricular arrythmia; therefore, an internal cardiac defibrillator (ICD) was recommended. A shared decision-making (EX6) toolkit was utilized by the HF APP, and Mr. Garcia agreed to ICD implantation and placement of an invasive pulmonary artery pressure device. Mr. Garcia did well clinically and underwent ongoing remote monitoring. The care team was alerted several months later that he was becoming fluid overloaded (EX7, EX8) and would benefit from outpatient IV diuretics to avoid hospitalization and overt decompensation. After 12 months, an AI application automatically reviewed Mr. Garcia's clinical progress and alerted the APP (EX1) that Mr. Garcia would be a reasonable candidate for a left ventricular assist device as a bridge to heart transplantation.

Ways AI can augment care

  • EX1: Tailor treatment plans for optimal clinical quality and outcomes

  • EX2: Apply predictive analytics to diagnostic tests

  • EX3: Strive to mitigate complications

  • EX4: Provide personalized education resources

  • EX5: Utilize generative AI in clinic note construction

  • EX6: Generate an individualized shared decision-making toolkit

  • EX7: Decrease risk for inpatient hospitalization

  • EX8: Prevent complications

REFERENCES

  • 1.Mintz Y, Brodie R. Introduction to artificial intelligence in medicine. Minim Invasive Ther Allied Technol. 2019;28(2):73–81. doi:10.1080/13645706.2019.1575882. [DOI] [PubMed] [Google Scholar]
  • 2.Sezgin E. Artificial intelligence in healthcare: complementing, not replacing, doctors and healthcare providers. Digit Health. 2023;9:20552076231186520. doi:10.1177/20552076231186520. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Elias P, Jain SS, Poterucha T, et al. Artificial intelligence for cardiovascular care—Part 1: Advances: JACC review topic of the week. J Am Coll Cardiol. 2024;83(24):2472–2486. doi:10.1016/j.jacc.2024.03.400. [DOI] [PubMed] [Google Scholar]
  • 4.Lüscher TF, Wenzl FA, D'Ascenzo F, et al. Artificial intelligence in cardiovascular medicine: clinical applications. Eur Heart J. 2024;45(40):4291–4304. doi:10.1093/eurheartj/ehae465. [DOI] [PubMed] [Google Scholar]
  • 5.Makimoto H, Kohro T. Adopting artificial intelligence in cardiovascular medicine: a scoping review. Hypertens Res. 2024;47(3):685–699. doi:10.1038/s41440-023-01469-7. [DOI] [PubMed] [Google Scholar]
  • 6.Wieneke H, Voigt I. Principles of artificial intelligence and its application in cardiovascular medicine. Clin Cardiol. 2024;47(1):e24148. doi:10.1002/clc.24148. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Rajkomar A, Dean J, Kohane I. Machine learning in medicine. N Engl J Med. 2019;380(14):1347–1358. doi:10.1056/NEJMra1814259. [DOI] [PubMed] [Google Scholar]
  • 8.Johnson KW, Torres Soto J, Glicksberg BS, et al. Artificial intelligence in cardiology. J Am Coll Cardiol. 2018;71(23):2668–2679. doi:10.1016/j.jacc.2018.03.521. [DOI] [PubMed] [Google Scholar]
  • 9.Yoon M, Park JJ, Hur T, et al. Application and potential of artificial intelligence in heart failure: past, present, and future. Int J Heart Fail. 2024;6(1):11–19. doi:10.36628/ijhf.2023.0050. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Jing L, Ulloa Cerna AE, Good CW, et al. A machine learning approach to management of heart failure populations. JACC Heart Fail. 2020;8(7):578–587. doi:10.1016/j.jchf.2020.01.012. [DOI] [PubMed] [Google Scholar]
  • 11.Noorbakhsh-Sabet N, Zand R, Zhang Y, Abedi V. Artificial intelligence transforms the future of health care. Am J Med. 2019;132(7):795–801. doi:10.1016/j.amjmed.2019.01.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Golas SB, Shibahara T, Agboola S, et al. A machine learning model to predict the risk of 30-day readmissions in patients with heart failure: a retrospective analysis of electronic medical records data. BMC Med Inform Decis Mak. 2018;18(1):44. doi:10.1186/s12911-018-0620-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Kwon JM, Kim KH, Jeon KH, et al. Artificial intelligence algorithm for predicting mortality of patients with acute heart failure. PLoS One. 2019;14(7):e0219302. doi:10.1371/journal.pone.0219302. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Feeny AK, Chung MK, Madabhushi A, et al. Artificial intelligence and machine learning in arrhythmias and cardiac electrophysiology. Circ Arrhythm Electrophysiol. 2020;13(8):e007952. doi:10.1161/circep.119.007952. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Han X, Hu Y, Foschini L, et al. Deep learning models for electrocardiograms are susceptible to adversarial attack. Nat Med. 2020;26(3):360–363. doi:10.1038/s41591-020-0791-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Singhal K, Azizi S, Tu T, et al. Large language models encode clinical knowledge. arXiv preprint arXiv:221213138. 2022. [Google Scholar]
  • 17.Choi MH, Yang JH, Seo JS, et al. Prevalence and diagnosis experience of osteoporosis in postmenopausal women over 50: focusing on socioeconomic factors. PLoS One. 2021;16(3):e0248020. doi:10.1371/journal.pone.0248020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Cabitza F, Rasoini R, Gensini GF. Unintended consequences of machine learning in medicine. JAMA. 2017;318(6):517–518. doi:10.1001/jama.2017.7797. [DOI] [PubMed] [Google Scholar]
  • 19.Boonstra MJ, Weissenbacher D, Moore JH, et al. Artificial intelligence: revolutionizing cardiology with large language models. Eur Heart J. 2024;45(5):332–345. doi:10.1093/eurheartj/ehad838. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Harvey D, Lobban F, Rayson P, et al. Natural language processing methods and bipolar disorder: scoping review. JMIR Ment Health. 2022;9(4):e35928. doi:10.2196/35928. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Berman AN, Biery DW, Ginder C, et al. Natural language processing for the assessment of cardiovascular disease comorbidities: the cardio-Canary comorbidity project. Clin Cardiol. 2021;44(9):1296–1304. doi:10.1002/clc.23687. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Idnay B, Dreisbach C, Weng C, Schnall R. A systematic review on natural language processing systems for eligibility prescreening in clinical research. J Am Med Inform Assoc. 2021;29(1):197–206. doi:10.1093/jamia/ocab228. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Brown SA, Hamid A, Pederson E, et al. Simplified rules-based tool to facilitate the application of up-to-date management recommendations in cardio-oncology. Cardiooncology. 2023;9(1):37. doi:10.1186/s40959-023-00179-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Seto E, Leonard KJ, Cafazzo JA, et al. Developing healthcare rule-based expert systems: case study of a heart failure telemonitoring system. Int J Med Inform. 2012;81(8):556–565. doi:10.1016/j.ijmedinf.2012.03.001. [DOI] [PubMed] [Google Scholar]
  • 25.Choi DJ, Park JJ, Ali T, Lee S. Artificial intelligence for the diagnosis of heart failure. NPJ Digit Med. 2020;3:54. doi:10.1038/s41746-020-0261-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Guo A, Pasque M, Loh F, et al. Heart failure diagnosis, readmission, and mortality prediction using machine learning and artificial intelligence models. Curr Epidemiol Rep. 2020;7(4):212–219. doi:10.1007/s40471-020-00259-w. [Google Scholar]
  • 27.Attia ZI, Kapa S, Lopez-Jimenez F. et al. Screening for cardiac contractile dysfunction using an artificial intelligence-enabled electrocardiogram. Nat Med. 2019;25(1):70–74. doi:10.1038/s41591-018-0240-2. [DOI] [PubMed] [Google Scholar]
  • 28.Boehmer JP, Hariharan R, Devecchi FG, et al. A multisensor algorithm predicts heart failure events in patients with implanted devices: results from the MultiSENSE study. JACC Heart Fail. 2017;5(3):216–225. doi:10.1016/j.jchf.2016.12.011. [DOI] [PubMed] [Google Scholar]
  • 29.Saeed A, AlShafea A, Bin Saeed A, et al. Robotics and artificial intelligence and their impact on the diagnosis and treatment of cardiovascular diseases. Cureus. 2023;15(7):e42252. doi:10.7759/cureus.42252. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Amabile A, Torregrossa G, Balkhy HH. Robotic-assisted coronary artery bypass grafting: current knowledge and future perspectives. Minerva Cardioangiol. 2020;68(5):497–510. doi:10.23736/s0026-4725.20.05302-5. [DOI] [PubMed] [Google Scholar]
  • 31.Amabile A, Komlo C, Van Praet KM, et al. Techniques for robotic-assisted surgical myocardial revascularization. Surg Technol Int. 2021;39:251–259. doi:10.52198/21.Sti.39.Cv1467. [PubMed] [Google Scholar]
  • 32.Knudsen JE, Ghaffar U, Ma R, Hung AJ. Clinical applications of artificial intelligence in robotic surgery. J Robot Surg. 2024;18(1):102. doi:10.1007/s11701-024-01867-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Hashimoto DA, Rosman G, Rus D, Meireles OR. Artificial intelligence in surgery: promises and perils. Ann Surg. 2018;268(1):70–76. doi:10.1097/sla.0000000000002693. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Shwetha R, Kirubanand VB. Remote monitoring of heart patients using robotic process automation (RPA). ITM Web Conf. 2021;37:01002. [Google Scholar]
  • 35.Murray NM, Phan P, Hager G, et al. Insurance payment for artificial intelligence technology: methods used by a stroke artificial intelligence system and strategies to qualify for the new technology add-on payment. Neuroradiol J. 2022;35(3):284–289. doi:10.1177/19714009211067408. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Luo R, Sun L, Xia Y, et al. BioGPT: generative pre-trained transformer for biomedical text generation and mining. Brief Bioinform. 2022;23(6). doi:10.1093/bib/bbac409. [DOI] [PubMed] [Google Scholar]
  • 37.Searle T, Ibrahim Z, Teo J, Dobson RJB. Discharge summary hospital course summarisation of in patient Electronic Health Record text with clinical concept guided deep pre-trained Transformer models. J Biomed Inform. 2023;141:104358. doi:10.1016/j.jbi.2023.104358. [DOI] [PubMed] [Google Scholar]
  • 38.Gal-Nadasan N, Stoicu-Tivadar V, Gal-Nadasan E, Dinu AR. Robotic process automation based data extraction from handwritten medical forms. Stud Health Technol Inform. 2023;309:68–72. doi:10.3233/shti230741. [DOI] [PubMed] [Google Scholar]
  • 39.Gaffar S, Gearhart AS, Chang AC. The next frontier in pediatric cardiology: artificial intelligence. Pediatr Clin North Am. 2020;67(5):995–1009. doi:10.1016/j.pcl.2020.06.010. [DOI] [PubMed] [Google Scholar]
  • 40.Shermock SB, Shermock KM, Schepel LL. Closed-loop medication management with an electronic health record system in U.S. and Finnish hospitals. Int J Environ Res Public Health. 2023;20(17). doi:10.3390/ijerph20176680. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Verma P. Hospital bosses love AI. Doctors and nurses are worried. The Washington Post. https://www.washingtonpost.com/technology/2023/08/10/ai-chatbots-hospital-technology/. [Google Scholar]
  • 42.Balagurunathan Y, Mitchell R, El Naqa I. Requirements and reliability of AI in the medical context. Phys Med. 2021;83:72–78. doi:10.1016/j.ejmp.2021.02.024. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Poalelungi DG, Musat CL, Fulga A, et al. Advancing patient care: how artificial intelligence is transforming healthcare. J Pers Med. 2023;13(8). doi:10.3390/jpm13081214. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Hoff T. Deskilling and adaptation among primary care physicians using two work innovations. Health Care Manage Rev. 2011;36(4):338–48. doi:10.1097/HMR.0b013e31821826a1. [DOI] [PubMed] [Google Scholar]
  • 45.Povyakalo AA, Alberdi E, Strigini L, Ayton P. How to discriminate between computer-aided and computer-hindered decisions: a case study in mammography. Med Decis Making. 2013;33(1):98–107. doi:10.1177/0272989x12465490. [DOI] [PubMed] [Google Scholar]
  • 46.Sahota N. The dark side of AI is how bad actors manipulate minds. Forbes. https://www.forbes.com/sites/neilsahota/2024/07/29/the-dark-side-of-ai-is-how-bad-actors-manipulate-minds/. [Google Scholar]
  • 47.Bharadwaj P, Nicola L, Breau-Brunel M, et al. Unlocking the value: quantifying the return on investment of hospital artificial intelligence. J Am Coll Radiol. 2024;21(10):1677–1685. doi:10.1016/j.jacr.2024.02.034. [DOI] [PubMed] [Google Scholar]
  • 48.Attia ZI, Kapa S, Yao X, et al. Prospective validation of a deep learning electrocardiogram algorithm for the detection of left ventricular systolic dysfunction. J Cardiovasc Electrophysiol. 2019;30(5):668–674. doi:10.1111/jce.13889. [DOI] [PubMed] [Google Scholar]
  • 49.Alsharqi M, Woodward WJ, Mumith JA, et al. Artificial intelligence and echocardiography. Echo Res Pract. 2018;5(4):R115–R125. doi:10.1530/erp-18-0056. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Weng SF, Reps J, Kai J, et al. Can machine-learning improve cardiovascular risk prediction using routine clinical data. PLoS One. 2017;12(4):e0174944. doi:10.1371/journal.pone.0174944. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.He J, Baxter SL, Xu J, et al. The practical implementation of artificial intelligence technologies in medicine. Nat Med. 2019;25(1):30–36. doi:10.1038/s41591-018-0307-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Oliveira Chaves L, Gomes Domingos AL, Louzada Fernandes D, et al. Applicability of machine learning techniques in food intake assessment: a systematic review. Crit Rev Food Sci Nutr. 2023;63(7):902–919. doi:10.1080/10408398.2021.1956425. [DOI] [PubMed] [Google Scholar]
  • 53.Le Glaz A, Haralambous Y, Kim-Dufor DH, et al. Machine learning and natural language processing in mental health: systematic review. J Med Internet Res. 2021;23(5):e15708. doi:10.2196/15708. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Lee EE, Torous J, De Choudhury M, et al. Artificial intelligence for mental health care: clinical applications, barriers, facilitators, and artificial wisdom. Biol Psychiatry Cogn Neurosci Neuroimaging. 2021;6(9):856–864. doi:10.1016/j.bpsc.2021.02.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Washif J, Pagaduan J, James C, et al. Artificial intelligence in sport: exploring the potential of using ChatGPT in resistance training prescription. Biol Sport. 2023;41(2):209–220. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Singh B, Olds T, Brinsley J, et al. Systematic review and meta-analysis of the effectiveness of chatbots on lifestyle behaviours. NPJ Digit Med. 2023;6(1):118. doi:10.1038/s41746-023-00856-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Dergaa I, Saad HB, El Omri A, et al. Using artificial intelligence for exercise prescription in personalised health promotion: a critical evaluation of OpenAI's GPT-4 model. Biol Sport. 2024;41(2):221–241. doi:10.5114/biolsport.2024.133661. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Zaleski AL, Berkowsky R, Craig KJT, Pescatello LS. Comprehensiveness, accuracy, and readability of exercise recommendations provided by an AI-based chatbot: mixed methods study. JMIR Med Educ. 2024;10:e51308. doi:10.2196/51308. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Association of Clinical Research Professionals (ACRP). Code of Ethics. https://acrpnet.org/about-acrp/code-of-ethics.
  • 60.Nyariro M, Emami E, Caidor P, Abbasgholizadeh Rahimi S. Integrating equity, diversity and inclusion throughout the lifecycle of AI within healthcare: a scoping review protocol. BMJ Open. 2023;13(9):e072069. doi:10.1136/bmjopen-2023-072069. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Navigating the Responsible and Ethical Incorporation of Artificial Intelligence into Clinical Practice. Federation of State Medical Boards; 2024. https://www.fsmb.org/siteassets/advocacy/policies/incorporation-of-ai-into-practice.pdf.

Articles from Jaapa are provided here courtesy of Wolters Kluwer Health

RESOURCES