Abstract
Randomized clinical trials are the gold standard for establishing the efficacy and safety of cardiovascular therapies. However, current pivotal trials are expensive, lengthy, and insufficiently diverse. Emerging artificial intelligence (AI) technologies can potentially automate and streamline clinical trial operations. This review describes opportunities to integrate AI throughout a trial’s life cycle, including designing the trial, identifying eligible patients, obtaining informed consent, ascertaining physiological and clinical event outcomes, interpreting imaging, and analyzing or disseminating the results. Nevertheless, AI poses risks, including generating inaccurate results, amplifying biases against underrepresented groups, and violating patient privacy. Medical journals and regulators are developing new frameworks to evaluate AI research tools and the data they generate. Given the high-stakes role of randomized trials in medical decision making, AI must be integrated carefully and transparently to protect the validity of trial results.
Keywords: artificial intelligence, automated, large language model, randomized
CENTRAL ILLUSTRATION
Opportunities to Improve Clinical Trials With Artificial Intelligence
Artificial intelligence (AI) research tools have the potential to improve clinical trials at multiple stages including design, recruitment, follow-up, and interpretation. EHR = electronic health record; NLP = natural language processing.

Randomized clinical trials are the gold standard for establishing the efficacy and safety of medical therapies.1–3 Evidence from randomized trials supports the regulatory approval or clearance of novel drugs and medical devices, as well as the clinical practice guidelines and insurance coverage determinations that govern whether, when, and in whom such therapies are used. Unfortunately, pivotal randomized trials are expensive and lengthy, and they frequently fail to include a group of patients representative of those who will ultimately receive therapy.4 These limitations are particularly salient in cardiovascular (CV) medicine, where trials are designed to detect relatively modest reductions (eg, 20%) in outcomes that accrue slowly, such as myocardial infarctions, heart failure events, or CV deaths.
The cost of clinical trials supporting regulatory approval of new agents in CV medicine has been reported to be $35,000 per participant or more.5,6 The several-year duration required for pivotal trials may delay potential benefits for patients and also reduces the time between regulatory approval and patent expiration, the window in which pharmaceutical companies must generate a return on research and development expenses that often exceed $1 billion.7 Despite the cost, marginalized racial and ethnic groups and women are underrepresented in clinical trials, a factor that undermines the equity, generalizability, fairness, and credibility of trial results.8–11 Thus, clinical trials must evolve to be less expensive, faster, and more representative of diverse patient groups.
Artificial intelligence (AI) has affected many areas of society, spurred by technical advances in deep learning and large language models. The release of public-facing tools such as ChatGPT (OpenAI) demonstrated the utility of AI to individual users. In biologic science, the potential value of AI has been demonstrated by, for example, deep learning models that accurately predict protein structure and have rationally designed a new class of antibiotics.12,13 AI has begun to be tested (although rarely applied) in clinical CV care, particularly for automated imaging interpretation and early disease diagnosis.14–19 Although there is much promise, greater interpretability, validation, and monitoring of AI performance in a clinical environment are needed to support AI uptake.20,21
AI has the potential to accelerate and automate clinical trials throughout their life cycle, from initial planning to patient recruitment, informed consent, ascertainment of endpoints, and dissemination of the results (Central Illustration).22 To date, AI tools have infrequently been applied within CV trials. Given the high stakes of clinical trials in evaluating new therapies, there is understandable concern that AI could introduce bias, reinforce existing inequities, or instill inaccuracy or inconsistency compared with traditional approaches. Regulators and journals increasingly receive submissions describing AI tools and AI-generated results, which require new evaluation frameworks.23
On March 15, 2024, the Heart Failure Collaboratory convened a special focus meeting to discuss the role of AI in CV clinical trials and therapeutic development. Key stakeholders from academia, industry, medical journals, and the U.S. Food and Drug Administration (FDA) addressed opportunities for AI to improve trial design, conduct, and interpretation, as well as challenges and risks. This review summarizes the discussion from this meeting. It highlights ongoing work and future directions for applying AI in various aspects of clinical trial design and execution, with a focus on CV trials.
TRIAL DESIGN
AI may be able to assist investigators in trial design, including selection of inclusion criteria and pre-specified subgroups. In oncology, the Trial Pathfinder AI tool was developed to emulate trial results using electronic health record data and inverse probability weighting.24 Applying this tool with various sets of inclusion criteria suggested that oncology trials could achieve similar treatment effect HRs with broader inclusion criteria that would in turn facilitate faster enrollment and greater generalizability of results. However, in silico trials are not a substitute for testing an intervention in real patients, in a prospective randomized fashion. Emulation of previously completed trials has not always found the same results as the randomized study.25 Moreover, trial emulation from real-world data requires that patients are treated with the investigational therapeutic in routine clinical care, and therefore this approach is not possible for novel therapeutics that are not yet approved.
SCREENING POTENTIAL PARTICIPANTS AT SCALE
Recruiting participants is frequently the rate-limiting step in trial progress. Screening potential participants for complex eligibility criteria is time-consuming and often delegated to research assistants. Cohort identification using electronic health record data is common, as are alerts that notify investigators in real time of patients who potentially meet eligibility criteria. However, these methods are typically limited to the evaluation of discrete data elements and require timely data availability. An automated tool able to access free text, imaging, and laboratory data could improve accurate participant identification, with the ability to screen hundreds of thousands of patients in the electronic health record. Such a tool could uncover eligible subjects not initially considered by a human reviewer, and it could more quickly exclude those patients found to be ineligible from the cohort lists. Natural language processing (NLP) and generative AI may help to move beyond discrete data (ie, left ventricular ejection fraction, estimated glomerular filtration rate) to more subjective and nuanced trial criteria (symptomatic heart failure or NYHA functional class).
Several models have emerged for this task. Rules-based language models that are founded on specific words or phrases have successfully extracted inclusion and exclusion criteria data from unstructured notes on the basis of specific words or phrases.26–30 More recently, the RECTIFIER (Retrieval-Augmented Generation–Enabled Clinical Trial Infrastructure for Inclusion Exclusion Review) tool was developed and tested in the COPILOT-HF (Co-Operative Program for Implementation of Optimal Therapy in Heart Failure) trial, a randomized trial for patients with symptomatic heart failure.31 RECTIFIER used Generative Pretrained Transformer Version 4 and Retrieval-Augmented Generation to assess 6 inclusion criteria and 17 exclusion criteria in potential participants’ electronic health record data. RECTIFIER was inexpensive ($0.10 per patient screened) and accurate (its eligibility assessment agreed with expert clinicians in 98%-100% of cases). As AI increasingly standardizes documentation, automated annotation tasks may become easier and more accurate. Eligibility assessment at scale may represent a relatively low-risk application of AI because the final decision for enrollment resides with an investigator who can prevent the enrollment of inappropriate participants.
OBTAINING INFORMED CONSENT
Current informed consent conversations between investigators and patients are often perfunctory and may bias the interaction toward enrollment. Long consent forms written in obtuse legal or scientific language do not help participants understand the planned research.32,33 Generative large language models have been successful in reducing the complexity and reading time of surgical consent forms.34 Interactive chatbots could play a future role in improving the efficiency and quality of informed consent for clinical trial participation. Unlike human investigators, chatbots have unlimited time to interact with participants and answer their questions, as well as the capacity to assess the participant’s understanding objectively and adjust language, readability, and mode of interaction to fit the participant’s needs. Moreover, AI-based informed consent may reduce participant burden by negating the need for in-person study visits during business hours.
The Pediatric Mendelian Genomic Research Center observational study implemented an optional chatbot-based consent process on participant smartphones that used a predetermined script.35 Patients electing chatbot consent had shorter consent interactions, often completed the consent outside regular business hours, and scored as well as patients who received traditional in-person consent information on a quiz assessing their understanding of the study.35 Participant satisfaction with the chatbot consent process was high (86%). In a study of consent for surgical care (not research), ChatGPT-generated text describing the risks and benefits of common surgical procedures was more readable, complete, and accurate than text written by surgeons.36
The use of AI to obtain informed consent raises ethical concerns. Chatbots may behave in a coercive or biased manner. The human connection between investigator and participant plays a vital role in building trust and maintaining respect for the participant’s autonomy and dignity.37 Given these concerns, AI should only augment rather than replace the investigator’s role for randomized clinical trials.
CLINICAL ENDPOINT ADJUDICATION
Automated adjudication of clinical outcomes by NLP has the potential to improve the trial cost, speed, and reproducibility. In current pivotal trials, outcomes such as heart failure hospitalization or myocardial infarction are commonly adjudicated by a central clinical events committee (CEC) of physicians who review participant medical records on the basis of established criteria or alternatively by site investigators.38 CEC adjudication is labor-intensive, expensive, and not easily scalable, and individual site investigator decisions may not be uniform.
Early studies using NLP for clinical outcome adjudication focused on observational electronic health record data.39–42 In the INVESTED (Influenza Vaccine to Effectively Stop Cardio Thoracic Events and Decompensated Heart Failure) trial, investigators externally validated an NLP model developed at a single center, Mass General Brigham (Boston, Massachusetts, USA), compared with human CEC adjudication. The NLP adjudication of heart failure agreed with the CEC in 87% of cases, thus demonstrating the model’s generalizability from the single center in a multicenter setting within the United States and Canada (Figure 1). Fine-tuning the model within INVESTED improved performance up to a reproducibility level equal to that of human reviewers.43
FIGURE 1. NLP for Automated Endpoint Adjudication in the INVESTED Trial.

A natural language processing (NLP) model developed to identify heart failure (HF) hospitalizations in a single-center electronic health record (EHR) cohort was externally validated in the multicenter INVESTED (Influenza Vaccine to Effectively Stop Cardio Thoracic Events and Decompensated Heart Failure) clinical trial. CEC = clinical event committee.
AI has the potential to enhance outcome ascertainment in large-scale pragmatic trials in which human adjudication is not feasible. AI approaches are rapid, avoid delays in detecting treatment benefits or harms, and may provide more consistency than site-level adjudication.44,45 Top priorities for research and development in outcomes ascertainment include developing more accurate models, assessing generalizability to international trials or CV endpoints beyond heart failure, and evaluating strategies that combine human and AI chart review.46 Moreover, in addition to efficacy outcome adjudication, AI could categorize adverse events, ascertain relevant and possible underdiagnosed comorbidities, or identify patients who are at high risk for dropping out and who may benefit from additional education.
DIGITAL BIOMARKERS
Mobile health and wearable technologies now enable the collection of vast amounts of physiological data from trial participants without the burden of in-person specific study visits. Digital biomarkers can be obtained from such devices, including metrics derived from vital sign measurements, skin temperature, skin conductance, and accelerometry, among others (Figure 2).47–49 Consumer wearables measure heart rate more consistently and precisely than in-clinic measurements, and they can capture diurnal variation, thereby offering even more information than in-clinic measurements.50 Emerging large language models refined to analyze data from wearables, such as the Personal Health Large Language Model from Google, may help extract meaningful insights and recommendations for patients.51 Global positioning systems on smartphones can assist investigators in determining when a patient visits a clinic or hospital by using geofencing.52 The HearO speech analysis tool (Cordio) has been developed to identify congested compared with euvolemic patients with heart failure from recordings of their speech.53,54 Noninvasive tools to estimate pulmonary congestion from handheld devices are also in development.55 These tools may help identify worsening heart failure events in clinical trial participants remotely.
FIGURE 2. Applications of Digital Health Technologies for Remote Data Collection.

Digital health technologies may aid in the collection of physiological and patient-reported outcomes. GPS = global positioning system
Translating digital biomarkers into clinical practice or research applications requires rigorous verification and validation, which, in the most straightforward implementation, involve correlation with traditional surrogate markers. With AI, there is an opportunity to discover novel digital biomarkers that may not have a comparable traditional metric. Both the comparative and novel paths of digital biomarker discovery require establishing a link to outcomes that are meaningful to clinicians and patients.56
Digital biomarker data collection often depends on patients’ owning and being comfortable using digital technology. This requirement may lead to exclusion of patients who are poor or who lack technological literacy, and it may thereby exacerbate the underrepresentation of some socioeconomic groups. One solution is to provide a parallel option for patients that does not require technology.
AI-ENABLED INTERPRETATION OF CARDIOVASCULAR IMAGING
CV imaging may be included in clinical trials as an inclusion criterion or outcome, or for safety monitoring. Central core laboratories often provide standardized reviews of imaging studies within trials, but these reviews are labor-intensive, costly, and time-consuming. Advances in deep learning applied to images have catalyzed the development of numerous models for automated interpretation of CV imaging studies.14
Within clinical trials, automated imaging interpretation can potentially reduce the cost and improve the reproducibility of imaging-based outcomes. For example, recent clinical trials of cardiac myosin inhibitors for hypertrophic cardiomyopathy required frequent echocardiograms to assess the drug’s effect on left ventricular systolic function and left ventricular outflow tract obstruction and determine dose adjustments.57 Instantaneous interpretation of these echocardiograms by using validated AI tools could have eliminated the time required for the echocardiograms to be read, or even needed, and, therefore, could have shortened the trial’s duration and reduced costs. In echocardiography, deep learning models accurately measure left ventricular ejection fraction and other parameters.58–61 Deep learning models for interpreting electrocardiograms,17 cardiac magnetic resonance,62,63 and computed tomography64 are available and could be similarly applied in clinical trials. AI interpretation of an electrocardiogram may be a proxy for the echocardiogram and may even obviate the need for an echocardiogram.65,66 Deploying AI for image interpretation at the point of care is an important hurdle to meaningful patient impact. Open-source tools such as the AI-integrated Picture Archives Communication System (PACS-AI) platform can help integrate AI into existing imaging systems and thereby facilitate safe deployment.67
Automated assessment of cardiac catheterization images could enable rapid screening of patients with coronary artery disease for trial eligibility. CathAI and DeepCoro are deep learning tools that accurately measure stenosis severity on coronary angiograms.68,69 Several companies, such as HeartFlow and Cleerly, have algorithms that automate assessments of stenoses and flow. CathEF measures left ventricular ejection fraction from standard angiographic videos with a mean absolute error in ejection fraction of 7% to 9% compared with echocardiographic measurements simultaneously.70 These tools, and many others that will be developed, will help overcome the challenges of recruiting and randomizing research patients in the catheterization laboratory immediately after initial angiography without delaying treatment.
CONTINUOUS PARTICIPANT MONITORING OUTSIDE INDIVIDUAL SITES
Explanatory clinical trials typically collect participant follow-up data (vital signs, laboratory testing, and patient-reported outcomes) from clinical electronic health records and intermittent in-person visits. For clinical event outcomes, medical records are printed, scanned, and manually submitted for adjudication by research core laboratories or committees, such as the CEC. The move to decentralize clinical trials seeks to shift the evaluation and monitoring of patients from the research site to the patient’s home or other convenient setting.71–74 Decentralization can reduce the considerable site-related and socioeconomic burdens for patient participation, such as transportation, time off work, and child care. Decentralization expands the geographic reach of clinical trial participation to patients who do not live near academic medical centers. Lower participant burden may lead to faster and more diverse recruitment, including of groups traditionally underrepresented in clinical trials.
AI is expected to facilitate the transition of trials from in-person site-based study-specific research visits to in-home or other ambulatory assessments. For example, an AI-enabled chatbot that obtains high-quality consent by smartphone obviates the study visit where traditional consent would be signed. Continuous at-home vital sign monitoring by wearables could eliminate in-person postrandomization study visits.49,50 Direct access to a participant’s electronic health record and automated endpoint adjudication eliminates the burden of working with the participant to identify clinical events and obtain clinical documents and is being integrated into trials.75 AI can improve the data processing and incorporate a vast corpus of records to distill key information vital to the trial.76 In addition to lowering costs and reducing participant burden, these methods generate a continuously updated and richer data set. Continuous and blinded evaluation of trials by AI algorithms may provide faster feedback to the sponsor and regulators for signals of potential treatment benefit or harm, facilitate adaptive trial design, and complement the current practice of episodic reviews by a human data safety monitoring board.
ANALYZING WHICH PATIENTS BENEFIT MOST FROM THERAPIES
Identifying subgroups of trial participants who may have benefited more from the investigational therapy is a key question in clinical trial analysis and is critical for personalized medicine. A traditional approach is to prespecify a handful of biologically plausible subgroups and test whether these baseline variables modify the effect of treatment. However, trials are rarely powered to test for such interactions.
AI has the potential to identify subgroups of patients who respond to therapy by using all pre-randomization features. One such methodology creates a multidimensional representation of the trial group across all baseline characteristics, identifies “neighborhoods” of similar participants, and quantifies a participant’s likely treatment response on the basis of the response of similar participants. This approach has demonstrated accuracy in predicting treatment responses in external validation data sets for aggressive blood pressure management trials, anatomic vs functional testing for coronary disease, and sodium-glucose cotransporter 2 inhibitors.77–79 Such an analysis of expected treatment effect could be applied not only for post hoc interpretation and personalized medicine, but also during ongoing trials to set and adapt eligibility criteria to enrich for participants likely to benefit and to limit harm in patients less likely to benefit and thus more fully power analyses in the selected subgroup. Simulation analyses suggest an adaptive eligibility strategy could reduce the necessary sample size by 15% to 20%.80
PUBLICATION AND DISSEMINATION OF RESULTS
Generative AI is already used to hasten the preparation of academic manuscripts. After the release of ChatGPT, the academic publishing community was forced to grapple with whether using generative AI to draft, edit, or review papers is ethical and safe.81 Many journals, including the New England Journal of Medicine and JACC, have coalesced in permitting generative AI in manuscript preparation as long as the authors disclose its use and take full responsibility for the final manuscript.82,83 Science, which initially forbade any use of generative AI, relaxed its policy in line with this consensus.84
Applying generative AI to automate additional post-trial activities from data analysis to preparation of regulatory submissions could meaningfully reduce the time from trial completion to availability of effective drugs to patients.22 Moreover, AI could assist with communicating trial results back to trial participants in nonmedical language, a responsibility that has often been neglected. Each of these steps presents, however, a potential risk for biased, fraudulent, or simply incorrect interpretation of the trial results. The gradual integration of AI with careful human oversight remains the most prudent path to improving efficiency while maintaining safety and refining best practices.
ROLE OF MEDICAL JOURNALS IN EVALUATING AI METHODOLOGY
Academic journals also have a key role to play in vetting and popularizing AI tools.85,86 Studies assessing AI’s technical accuracy and clinical impact should be subjected to peer review. Papers on AI tools pose several new challenges for journal editors. First, because AI technology is progressing quickly, journals must offer prompt review and publication, given that the field is evolving rapidly. Second, journal editors must be open-minded to novel research methods that challenge the status quo and recognize the need to find qualified reviewers. Third, although some journals contend that authors must release underlying models publicly at publication, this may not be feasible when the software is proprietary intellectual property. An alternative to making the models public is to require validation by a qualified, independent party with full access to the software model and data sets. This is particularly important because developers of AI tools often have a financial interest in them; journals should continue to insist that all authors disclose financial ties.85 Independent “health AI assurance laboratories,” federal agencies, or academic groups may be positioned to evaluate AI tools credibly.20
REGULATORY PRIORITIES AND GUIDANCE ON AI IN CLINICAL TRIALS
Since 1995, the FDA has received >300 submissions for drugs and biologic products with AI components and >700 submissions for AI-enabled medical devices.23,87,88 The drug and biologic application submissions using AI traverse the landscape of drug development from drug discovery to postmarket safety surveillance and cut across a range of therapeutic areas, including CV disease. The diverse uses of AI in these submissions highlight the need for careful regulatory assessment of both benefits and risks and underscore the importance of adopting a risk-based approach commensurate with the level of risk posed by the specific context of use. For any specific AI application in drug development, model risk calculations will be determined by model influence and the decision consequence on the basis of the context of use. For example, high-risk models may require more evidence of credibility than low-risk models, and the regulatory approach may differ accordingly.
As with any innovation, AI creates opportunities and new and unique challenges. To meet these challenges, the FDA has accelerated its efforts to create an agile regulatory ecosystem that can facilitate innovation and adoption while ensuring public safety and guarding against potential risks.89 For example, in 2021, the FDA, together with Health Canada and the United Kingdom’s Medicines and Healthcare Products Regulatory Agency, jointly identified 10 guiding principles informing Good Machine Learning Practice for medical devices that are AI enabled.90 These principles include the importance of bias mitigation by ensuring that clinical study participants and data sets are representative of the intended patient group using the device, the performance of the human AI team, and postapproval monitoring of performance. In March 2024, the FDA’s medical products centers published a joint report describing 4 major areas of focus in regulating AI (Table 1).87 These areas are collaboration with key stakeholders, support for innovation, development of standards and best practices, and research related to monitoring AI performance for bias and inequity.
TABLE 1.
U.S. Food and Drug Administration Areas of Focus Regarding the Development and Use of AI Across the Medical Product Lifecycle
| 1. Foster collaboration to safeguard public health. |
| 2. Advance the development of regulatory approaches that support innovation. |
| 3. Promote the development of harmonized standards, guidelines, best practices, and tools. |
| 4. Support research related to the evaluation and monitoring of AI performance. |
Reprinted from the U.S. Food and Drug Administration.87
AI = artificial intelligence.
The regulatory process for AI tools targeted to research rather than patient care, or a combination of the 2, is distinct from the 510(k), de novo, or premarket approval medical device review but follows similar principles. The FDA Drug Development Tool (DDT) and Medical Device Development Tool (MDDT) programs specifically assess AI tools for use in research rather than clinical care. The DDT program from the Center for Drug Evaluation and Research (CDER) and the Center for Biological Evaluation and Research (CBER) qualifies methods, materials, and measures that have the potential to facilitate drug development.91 Examples of non-AI DDTs include biomarkers for clinical trial enrichment, clinical outcome assessments to evaluate clinical benefit, and animal models. DDT qualification is not required for applying a tool in clinical trials. Nonetheless, it may streamline the review of subsequent regulatory submissions by avoiding the need for the FDA to reconsider and reconfirm the tool’s validity in each drug program to which it is applied. DDTs are evaluated for a specific context of use, including recognition of a tool’s limitations and the contexts in which its application is inappropriate. The Innovative Science and Technology Approaches for New Drugs (ISTAND) Pilot Program accepts submissions for DDTs focused on AI and digital health technology (DHT).92 The MDDT program at the Center for Devices and Radiological Health (CDRH) is similar to the DDT program but focuses on medical device research and regulation.93 These programs do not replace the traditional dialogue between investigators or sponsors and regulators to review clinical trial methodology in advance.
Additional FDA programs provide opportunities for early engagement with the agency regarding applying AI tools in clinical trials. For example, the Critical Path Innovation Meetings (CPIM) facilitate dialogue among industry, academia, patients, and government regarding new tools and technologies to improve the efficiency of drug development. These nonbinding discussions seek to encourage innovations, including potential submissions to the DDT program. Another example of a pathway for engagement is the DHT Program, which enables developers of AI-enabled DHT used in a drug development program (eg, wearables that collect data remotely and therefore reduce the burden on trial participants) to interact with the agency.94 Further dialogue on approaches to qualifying such technologies is planned. The recently formed CDER Center for Clinical Trial Innovation seeks to improve the efficiency and effectiveness of clinical trials, participant diversity, and the pace of drug development, and it may provide an additional forum to discuss best practices for integrating AI. In these programs, the FDA emphasizes the importance of early discussions between stakeholders and the agency regarding using novel AI technologies in clinical trials.
LIMITATIONS AND POTENTIAL PITFALLS OF AI IN CV TRIALS
Despite the tremendous promise of AI in improving CV trials, several potential risks must be acknowledged and managed carefully (Table 2). First, AI may be less accurate than time-tested data collection and interpretation methods by humans. Given the importance of clinical trial results to patient care, using “quick and dirty” methods to save cost or time carries even greater risk. Second, AI is susceptible to data set shift, in which differences between the data used to train and validate the model and the data on which the model is applied lead to poor performance.95 For example, an event adjudication model developed in one electronic health record system may perform poorly after a site transition to a new system. Stable models may become quickly outdated, and continuously learning models may accumulate new biases or experience degraded performance. Approved AI technologies require a plan for monitoring accuracy and bias at a routine cadence for as long as they are in use.96 Third, AI tools that learn from completed trials may encode or amplify biases against women or underserved groups.97 For example, a model seeking to identify eligible participants that was trained on trials with very low representation of Black patients may systematically ignore eligible Black patients and thus perpetuate inequity. Mitigating bias requires ensuring that the data used to train and evaluate the AI tool are representative of the diverse groups of patients to which it will ultimately be applied. Large, publicly available data sets such as CheXpert for chest radiographs,98 MIMIC-IV for electronic health records,99 and EchoNet-Dynamic for echocardiograms,58 can help mitigate bias and promote transparency. Evaluation for and mitigation of bias must occur through the life cycle of the tool, including the postmarket phase.100 Fourth, the confidentiality of participant medical data must be protected. For example, generative AI models automatically use all submitted data for future training unless specific restrictions or safeguards are in place, such as limitations on the basis of patient privacy or consent. Emerging open-source large language models that can be hosted locally within health care institution servers and tools such as LM Studio may help overcome privacy concerns.101 Fifth, human trial researchers relying on automated AI tools may not learn core competencies such as obtaining informed consent or adjudicating events, thus leaving the trial vulnerable if AI becomes unavailable or is flawed. For competencies critical to participant safety or the integrity of results, trials must ensure appropriate human oversight and competency.
TABLE 2.
Mitigating Key Risks of AI in Clinical Trials
| Risk | Mitigation Strategy | |
|---|---|---|
|
| ||
| Poor generalizability | AI model is inaccurate when applied in a novel context of a prospective trial. | Validate AI models in the context in which they will be used. |
| Data set shift | AI model performance erodes with changes in medical practice or data organization. | Update the AI model with contemporary data guided by a predetermined change control plan. |
| Algorithmic bias | AI models may perpetuate or amplify biases against marginalized groups. | Train or validate the AI model on representative data, and evaluate for bias. |
| Lack of clinical interpretability | Digital biomarkers may be unproven surrogates for clinical outcomes. | Insist on proven relationship between biomarkers and clinical outcomes. |
| Patient data privacy | Sharing patient data with third-party AI providers risks breach of confidentiality. | Require strict privacy agreements and encryption. |
| Patient access to technology | Requiring participants to use digital technology may exacerbate biases in enrollment. | Provide backup options by which patients can enroll and be followed up without digital technology. |
| Loss of human competency | Future researchers relying on AI may not learn critical skills. | Maintain human oversight of tasks required for participant safety and integrity of results. |
AI = artificial intelligence.
CONCLUSIONS
Randomized clinical trials of sufficient size are necessary to evaluate novel CV therapeutics and inform regulatory judgments, but traditionally they have been costly, long, and insufficiently diverse. Emerging AI technologies have the potential to address these limitations by automating and streamlining clinical trial operations. Opportunities to integrate AI exist throughout a trial’s life cycle, including identifying and obtaining consent from eligible patients, ascertaining physiological and clinical event outcomes, and analyzing or disseminating the results. Medical journals and regulators are developing new frameworks to evaluate AI research tools and the data they generate. Given the high-stakes role of randomized trials in medical decision making, AI methods must be integrated cautiously and thoughtfully to protect the validity of trial results.
FUNDING SUPPORT AND AUTHOR DISCLOSURES
Dr Cunningham has received support from the KL2/Harvard Catalyst Medical Research Investigator Training program and the American Heart Association (23CDA1052151); and has received consulting fees from Roche Diagnosis, Edgewise Therapeutics, KCK, and Occlutech. Dr Abraham has served as a consultant to Boehringer Ingelheim, CVRx, Impulse Dynamics, Sensible Medical, Vectorious, V-Wave, and Zoll Respicardia. Dr Bhatt has received research grant support to his institution from the National Institutes of Health (NIH) National Heart, Lung, and Blood Institute (NHLBI) and National Institute on Aging, the American College of Cardiology Foundation, and the Centers for Disease Control and Prevention (CDC); and has received consulting fees from the Kinetix Group, Merck, Sanofi Pasteur, and Novo Nordisk. Dr Dunn has served as a scientific advisor to Veri. Dr Felker has received research grants from NIH, Bayer, Bristol Myers Squibb, Novartis, Daxor, Merck, Cytokinetics, and CSL-Behring; has acted as a consultant to Novartis, Bristol Myers Squibb, Cytokinetics, Innolife, Boehringer Ingelheim, Abbott, Sanofi, Regeneron, Myovant, Sequana, Windtree Therapeutics, and Whiteswell; and has served on clinical endpoint committees or data safety monitoring boards for Merck, Medtronic, EBR Systems, Rocket Pharma, V-Wave, and LivaNova. Dr Jain has received consulting fees from Bristol Myers Squibb, ARTIS Ventures, and Broadview Ventures outside of the submitted work. Dr Lindsell has received grants and contracts from the NIH, the U.S. Department of Defense, CDC, Biomeme, Novartis, bioMérieux, Astra-Zeneca, AbbVie, Entegrion, and Endpoint Health, all outside of the submitted work; has obtained patents for risk stratification in sepsis and septic shock issued to Cincinnati Children’s Hospital Medical Center; has served on data safety monitoring boards unrelated to the current work; has held stock options in Bioscape Digital unrelated to the current work; and has served as Editor-in-Chief of the Journal of Clinical and Translational Science. Mr Mace is an employee of Acorai AB; and has held stock interest in Abbott Laboratories. Dr Martyn has served as an advisor to or has received consulting fees from Fire1, Cleveland Clinic/American Well Joint Venture, Boehringer Ingelheim/Eli Lilly, NIRSense, Novo Nordisk, AstraZeneca, and Apricity Robotics; and has received grant support from Ionis Therapeutics, AstraZeneca, and the Heart Failure Society of America. Dr Shah is an employee of Meta, which had no role in this work or providing financial support. Dr Tison has received research grants from MyoKardia, a wholly owned subsidiary of Bristol Myers Squibb, and Janssen Pharmaceuticals; and is an advisor to Viz.ai and Prolaio. Dr Fakhouri is an employee of the Office of Medical Policy, Center for Drug Evaluation and Research, U.S. Food and Drug Administration; the views expressed in this article are those of the authors and do not necessarily represent the views or policies of the U.S. Food and Drug Administration. Dr Krumholz has received options from Element Science and Identifeye; has received payments from F-Prime for advisory roles; has co-founded and held equity in Hugo Health, Refactor Health, and ENSIGHT-AI; and has been associated with research contracts through Yale University from Janssen, Kenvue, and Pfizer. Dr O’Connor has received consulting fees from Merck, Abiomed, and Zealcare. Dr Solomon has received research grants from Alexion, Alnylam, AstraZeneca, Bellerophon, Bayer, Bristol Myers Squibb, Boston Scientific, Cytokinetics, Edgewise, Eidos, Gossamer, GSK, Ionis, Lilly, MyoKardia, NIH NHLBI, Novartis, Novo Nordisk, Respicardia, Sanofi Pasteur, Theracos, and US2.AI; and has consulted for Abbott, Action, Akros, Alexion, Alnylam, Amgen, Arena, AstraZeneca, Bayer, Boehringer Ingelheim, Bristol Myers Squibb, Cardior, Cardurion, Corvia, Cytokinetics, Daiichi-Sankyo, GSK, Lilly, Merck, MyoKardia, Novartis, Roche, Theracos, Quantum Genomics, Janssen, Cardiac Dimensions, Tenaya, Sanofi-Pasteur, Dinaqor, Tremeau, CellProThera, Moderna, American Regent, Sarepta, Lexicon, AnaCardio, Akros, and Valo. Drs Psotka and Fiuzat have reported that they have no relationships relevant to the contents of this paper to disclose.
ABBREVIATIONS AND ACRONYMS
- AI
artificial intelligence
- CDER
Center for Drug Evaluation and Research
- CEC
clinical events committee
- CV
cardiovascular
- DDT
Drug Development Tool
- DHT
Digital Health Technology
- FDA
U.S. Food and Drug Administration
- MDDT
Medical Device Development Tool
- NLP
natural language processing
Footnotes
The authors attest they are in compliance with human studies committees and animal welfare regulations of the authors’ institutions and Food and Drug Administration guidelines, including patient consent where appropriate. For more information, visit the Author Center.
REFERENCES
- 1.Collins R, Bowman L, Landray M, Peto R. The magic of randomization versus the myth of real-world evidence. N Engl J Med. 2020;382(7):674–678. 10.1056/NEJMsb1901642 [DOI] [PubMed] [Google Scholar]
- 2.Freemantle N, Marston L, Walters K, Wood J, Reynolds MR, Petersen I. Making inferences on treatment effects from real world data: propensity scores, confounding by indication, and other perils for the unwary in observational research. BMJ. 2013;347:f6409. 10.1136/bmj.f6409 [DOI] [PubMed] [Google Scholar]
- 3.McMurray JJV. Only trials tell the truth about treatment effects. J Am Coll Cardiol. 2018;71(23):2640–2642. 10.1016/j.jacc.2018.04.019 [DOI] [PubMed] [Google Scholar]
- 4.O’Connor CM, Psotka MA, Fiuzat M, et al. improving heart failure therapeutics development in the United States. J Am Coll Cardiol. 2018;71(4):443–453. 10.1016/j.jacc.2017.11.048 [DOI] [PubMed] [Google Scholar]
- 5.Moore TJ, Zhang H, Anderson G, Alexander GC. Estimated costs of pivotal trials for novel therapeutic agents approved by the US Food and Drug Administration, 2015–2016. JAMA Intern Med. 2018;178(11):1451–1457. 10.1001/jamainternmed.2018.3931 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Moore TJ, Heyward J, Anderson G, Alexander GC. Variation in the estimated costs of pivotal clinical benefit trials supporting the US approval of new therapeutic agents, 2015–2017: a cross-sectional study. BMJ Open. 2020;10(6):e038863. 10.1136/bmjopen-2020-038863 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Wouters OJ, McKee M, Luyten J. Estimated research and development investment needed to bring a new medicine to market, 2009–2018. JAMA. 2020;323(9):844–853. 10.1001/jama.2020.1166 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Schwartz AL, Alsan M, Morris AA, Halpern SD. Why diverse clinical trial participation matters. N Engl J Med. 2023;388(14):1252–1254. 10.1056/NEJMp2215609 [DOI] [PubMed] [Google Scholar]
- 9.Ortega RF, Yancy CW, Mehran R, Batchelor W. Overcoming lack of diversity in cardiovascular clinical trials. Circulation. 2019;140(21):1690–1692. 10.1161/CIRCULATIONAHA.119.041728 [DOI] [PubMed] [Google Scholar]
- 10.Farb A, Viviano CJ, Tarver ME. Diversity in clinical trial enrollment and reporting—where we are and the road ahead. JAMA Cardiol. 2023;8(9):803–805. 10.1001/jamacardio.2023.2106 [DOI] [PubMed] [Google Scholar]
- 11.Lau ES, Braunwald E, Morrow DA, et al. Sex, permanent drug discontinuation, and study retention in clinical trials. Circulation. 2021;143(7):685–695. 10.1161/CIRCULATIONAHA.120.052339 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Jumper J, Evans R, Pritzel A, et al. Highly accurate protein structure prediction with AlphaFold. Nature. 2021;596(7873):583–589. 10.1038/s41586-021-03819-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Wong F, Zheng EJ, Valeri JA, et al. Discovery of a structural class of antibiotics with explainable deep learning. Nature. 2024;626(7997):177–185. 10.1038/s41586-023-06887-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Elias P, Jain S, Poterucha T, et al. Artificial intelligence for cardiovascular care - part 1: advances: JACC review topic of the week. J Am Coll Cardiol. 2024;83(24):2472–2486. 10.1016/j.jacc.2024.03.400 [DOI] [PubMed] [Google Scholar]
- 15.Jain S, Elias P, Poterucha T, et al. Artificial intelligence in cardiovascular care — part 2: applications: JACC review topic of the week. J Am Coll Cardiol. 2024;83(24):2487–2496. 10.1016/j.jacc.2024.03.401 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Khera R, Oikonomou EK, Nadkarni GN, et al. Transforming cardiovascular care with artificial intelligence: from discovery to practice. J Am Coll Cardiol. 2024;84(1):97–114. 10.1016/j.jacc.2024.05.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Lin CS, Liu WT, Tsai DJ, et al. AI-enabled electrocardiography alert intervention and all-cause mortality: a pragmatic randomized clinical trial. Nat Med. 2024;30(5):1461–1470. 10.1038/s41591-024-02961-4 [DOI] [PubMed] [Google Scholar]
- 18.Yao X, Rushlow DR, Inselman JW, et al. Artificial intelligence–enabled electrocardiograms for identification of patients with low ejection fraction: a pragmatic, randomized clinical trial. Nat Med. 2021;27(5):815–819. 10.1038/s41591-021-01335-4 [DOI] [PubMed] [Google Scholar]
- 19.Avram R, Fearon WF. AI-RISE to the challenge — artificial intelligence reduces time to treatment in STEMI. NEJM AI. 2024;1(7):AIe2400472. 10.1056/AIe2400472 [DOI] [Google Scholar]
- 20.Shah NH, Halamka JD, Saria S, et al. A nationwide network of health AI assurance laboratories. JAMA. 2024;331(3):245–249. 10.1001/jama.2023.26930 [DOI] [PubMed] [Google Scholar]
- 21.Blueprint for trustworthy AI implementation guidance and assurance for healthcare. Coalition for Health AI. 2023. Accessed May 20, 2024. https://www.coalitionforhealthai.org/papers/blueprint-for-trustworthy-ai_V1.0.pdf [Google Scholar]
- 22.Hernandez AF, Lindsell CJ. The future of clinical trials: artificial to augmented to applied intelligence. JAMA. 2023;330(21):2061–2063. 10.1001/jama.2023.23822 [DOI] [PubMed] [Google Scholar]
- 23.Liu Q, Huang R, Hsieh J, et al. Landscape analysis of the application of artificial intelligence and machine learning in regulatory submissions for drug development from 2016 to 2021. Clin Pharmacol Ther. 2023;113(4):771–774. 10.1002/cpt.2668 [DOI] [PubMed] [Google Scholar]
- 24.Liu R, Rizzo S, Whipple S, et al. Evaluating eligibility criteria of oncology trials using real-world data and AI. Nature. 2021;592(7855):629–633. 10.1038/s41586-021-03430-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Wang SV, Schneeweiss S, RCT-DUPLICATE Initiative. Emulation of randomized clinical trials with nonrandomized database analyses: results of 32 clinical trials. JAMA. 2023;329(16):1376–1385. 10.1001/jama.2023.4221 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Zhang K, Demner-Fushman D. Automated classification of eligibility criteria in clinical trials to facilitate patient-trial matching for specific patient populations. J Am Med Inform Assoc. 2017;24(4):781–787. 10.1093/jamia/ocw176 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Yuan C, Ryan PB, Ta C, et al. Criteria2Query: a natural language interface to clinical databases for cohort definition. J Am Med Inform Assoc. 2019;26(4):294–305. 10.1093/jamia/ocy178 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Kang T, Zhang S, Tang Y, et al. EliIE: an open-source information extraction system for clinical trial eligibility criteria. J Am Med Inform Assoc. 2017;24(6):1062–1071. 10.1093/jamia/ocx019 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Jonnalagadda SR, Adupa AK, Garg RP, Corona-Cox J, Shah SJ. Text mining of the electronic health record: an information extraction approach for automated identification and subphenotyping of HFpEF patients for clinical trials. J Cardiovasc Transl Res. 2017;10(3):313–321. 10.1007/s12265-017-9752-2 [DOI] [PubMed] [Google Scholar]
- 30.Idnay B, Dreisbach C, Weng C, Schnall R. A systematic review on natural language processing systems for eligibility prescreening in clinical research. J Am Med Inform Assoc. 2022;29(1):197–206. 10.1093/jamia/ocab228 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Unlu O, Shin J, Mailly CJ, et al. Retrieval augmented generation enabled generative pretrained transformer 4 (GPT-4) performance for clinical trial screening. medRxiv. Preprint. Posted online February 8, 2024. 10.1101/2024.02.08.24302376, 2024.02.08.24302376. [DOI] [Google Scholar]
- 32.Fortun P, West J, Chalkley L, Shonde A, Hawkey C. Recall of informed consent information by healthy volunteers in clinical trials. QJM. 2008;101(8):625–629. 10.1093/qjmed/hcn067 [DOI] [PubMed] [Google Scholar]
- 33.Grant SC. Informed consent—we can and should do better. JAMA Netw Open. 2021;4(4):e2110848. 10.1001/jamanetworkopen.2021.10848 [DOI] [PubMed] [Google Scholar]
- 34.Mirza FN, Tang OY, Connolly ID, et al. Using ChatGPT to facilitate truly informed medical consent. NEJM AI. 2024;1(2):AIcs2300145. 10.1056/AIcs2300145 [DOI] [Google Scholar]
- 35.Savage SK, LoTempio J, Smith ED, et al. Using a chat-based informed consent tool in large-scale genomic research. J Am Med Inform Assoc. 2024;31(2):472–478. 10.1093/jamia/ocad181 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Decker H, Trang K, Ramirez J, et al. Large language model–based chatbot vs surgeon-generated informed consent documentation for common procedures. JAMA Netw Open. 2023;6(10):e2336997. 10.1001/jamanetworkopen.2023.36997 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Rothstein MA. Should chatbots be used to obtain informed consent for research? Ethics Hum Res. 2023;45(6):46–50. 10.1002/eahr.500190 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Hicks KA, Mahaffey KW, Mehran R, et al. 2017 cardiovascular and stroke endpoint definitions for clinical trials. Circulation. 2018;137(9):961–972. 10.1161/CIRCULATIONAHA.117.033502 [DOI] [PubMed] [Google Scholar]
- 39.Ambrosy AP, Parikh RV, Sung SH, et al. A natural language processing–based approach for identifying hospitalizations for worsening heart failure within an integrated health care delivery system. JAMA Netw Open. 2021;4(11):e2135152. 10.1001/jamanetworkopen.2021.35152 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Ambrosy AP, Parikh RV, Sung SH, et al. Analysis of worsening heart failure events in an integrated health care system. J Am Coll Cardiol. 2022;80(2):111–122. 10.1016/j.jacc.2022.04.045 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Goto S, Homilius M, John JE, et al. Artificial intelligence-enabled event adjudication: estimating delayed cardiovascular effects of respiratory viruses. medRxiv. Preprint. Posted online November 16, 2020. 10.1101/2020.11.12.20230706, 2020.11.12.20230706. [DOI] [Google Scholar]
- 42.Cunningham JW, Singh P, Reeder C, et al. Natural language processing for adjudication of heart failure in the electronic health record. JACC Heart Fail. 2023;11(7):852–854. 10.1016/j.jchf.2023.02.012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Cunningham JW, Singh P, Reeder C, et al. Natural language processing for adjudication of heart failure in a multicenter clinical trial: a secondary analysis of a randomized clinical trial. JAMA Cardiol. 2024;9(2):174–181. 10.1001/jamacardio.2023.4859 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Carson P, Fiuzat M, O’Connor C, et al. Determination of hospitalization type by investigator case report form or adjudication committee in a large heart failure clinical trial (b-Blocker Evaluation of Survival Trial [BEST]). Am Heart J. 2010;160(4):649–654. 10.1016/j.ahj.2010.07.004 [DOI] [PubMed] [Google Scholar]
- 45.Carson P, Teerlink JR, Komajda M, et al. Comparison of investigator-reported and centrally adjudicated heart failure outcomes in the EMPEROR-Reduced trial. JACC Heart Fail. 2023;11(4):407–417. 10.1016/j.jchf.2022.11.017 [DOI] [PubMed] [Google Scholar]
- 46.Mahaffey KW, Gibson CM, Lopes RD. Innovation in event adjudication—human vs machine. JAMA Cardiol. 2024;9(2):101–102. 10.1001/jamacardio.2023.4900 [DOI] [PubMed] [Google Scholar]
- 47.Dunn J, Runge R, Snyder M. Wearables and the medical revolution. Pers Med. 2018;15(5):429–448. 10.2217/pme-2018-0044 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Coravos A, Khozin S, Mandl KD. Developing and adopting safe and effective digital biomarkers to improve patient outcomes. NPJ Digit Med. 2019;2(1):14. 10.1038/s41746-019-0090-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Bent B, Wang K, Grzesiak E, et al. The digital biomarker discovery pipeline: an open-source software platform for the development of digital biomarkers using mHealth and wearables data. J Clin Transl Sci. 2021;5(1):e19. 10.1017/cts.2020.511 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Dunn J, Kidzinski L, Runge R, et al. Wearable sensors enable personalized predictions of clinical laboratory measurements. Nat Med. 2021;27(6):1105–1112. 10.1038/s41591-021-01339-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Cosentino J, Belyaeva A, Liu X, et al. Towards a personal health large language model. 2024. ArXiv.org. 2024. Accessed June 1, 2024 https://arxiv.org/abs/2406.06474 [Google Scholar]
- 52.Nguyen KT, Olgin JE, Pletcher MJ, et al. Smartphone-based geofencing to ascertain hospitalizations. Circ Cardiovasc Qual Outcomes. 2017;10(3):e003326. 10.1161/CIRCOUTCOMES.116.003326 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Amir O, Anker SD, Gork I, et al. Feasibility of remote speech analysis in evaluation of dynamic fluid overload in heart failure patients undergoing haemodialysis treatment. ESC Heart Fail. 2021;8(4):2467–2472. 10.1002/ehf2.13367 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Amir O, Abraham WT, Azzam ZS, et al. Remote speech analysis in the evaluation of hospitalized patients with acute decompensated heart failure. JACC Heart Fail. 2022;10(1):41–49. 10.1016/j.jchf.2021.08.008 [DOI] [PubMed] [Google Scholar]
- 55.Mace MI. A novel multisensor device for absolute intracardiac pressure measurement, detection, and management of heart failure. JACC Basic Transl Sci. 2023;8(4):377–379. 10.1016/j.jacbts.2023.02.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Goldsack JC, Coravos A, Bakker JP, et al. Verification, analytical validation, and clinical validation (V3): the foundation of determining fit-for-purpose for biometric monitoring technologies (BioMeTs). NPJ Digit Med. 2020;3(1):55. 10.1038/s41746-020-0260-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Olivotto I, Oreziak A, Barriales-Villa R, et al. Mavacamten for treatment of symptomatic obstructive hypertrophic cardiomyopathy (EXPLORER-HCM): a randomised, double-blind, placebo-controlled, phase 3 trial. Lancet. 2020;396(10253):759–769. 10.1016/S0140-6736(20)31792-X [DOI] [PubMed] [Google Scholar]
- 58.Ouyang D, He B, Ghorbani A, et al. Video-based AI for beat-to-beat assessment of cardiac function. Nature. 2020;580(7802):252–256. 10.1038/s41586-020-2145-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.He B, Kwan AC, Cho JH, et al. Blinded, randomized trial of sonographer versus AI cardiac function assessment. Nature. 2023;616(7957):520–524. 10.1038/s41586-023-05947-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Tromp J, Bauer D, Claggett BL, et al. A formal validation of a deep learning-based automated workflow for the interpretation of the echocardiogram. Nat Commun. 2022;13(1):6776. 10.1038/s41467-022-34245-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Lau ES, Di Achille P, Kopparapu K, et al. Deep learning–enabled assessment of left heart structure and function predicts cardiovascular outcomes. J Am Coll Cardiol. 2023;82(20):1936–1948. 10.1016/j.jacc.2023.09.800 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Pirruccello JP, Bick A, Wang M, et al. Analysis of cardiac magnetic resonance imaging in 36,000 individuals yields genetic insights into dilated cardiomyopathy. Nat Commun. 2020;11(1):2254. 10.1038/s41467-020-15823-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Pirruccello JP, Di Achille P, Nauffal V, et al. Genetic analysis of right heart structure and function in 40,000 people. Nat Genet. 2022;54(6):792–803. 10.1038/s41588-022-01090-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Lin A, Manral N, McElhinney P, et al. Deep learning-enabled coronary CT angiography for plaque and stenosis quantification and cardiac risk prediction: an international multicentre study. Lancet Digit Health. 2022;4(4):e256–e265. 10.1016/S2589-7500(22)00022-X [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Sangha V, Mortazavi BJ, Haimovich AD, et al. Automated multilabel diagnosis on electrocardiographic images and signals. Nat Commun. 2022;13(1):1583. 10.1038/s41467-022-29153-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Sangha V, Nargesi AA, Dhingra LS, et al. Detection of left ventricular systolic dysfunction from electrocardiographic images. Circulation. 2023;148(9):765–777. 10.1161/CIRCULATIONAHA.122.062646 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Theriault-Lauzier P, Cobin D, Tastet O, et al. A responsible framework for applying artificial intelligence on medical images and signals at the point of care: the PACS-AI platform. Can J Cardiol. 2024;40(10):1828–1840. 10.1016/j.cjca.2024.05.025 [DOI] [PubMed] [Google Scholar]
- 68.Avram R, Olgin JE, Ahmed Z, et al. CathAI: fully automated coronary angiography interpretation and stenosis estimation. NPJ Digit Med. 2023;6(1):142. 10.1038/s41746-023-00880-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Labrecque Langlais É, Corbin D, Tastet O, et al. Evaluation of stenoses using AI video models applied to coronary angiography. NPJ Digit Med. 2024;7(1):138. 10.1038/s41746-024-01134-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70.Avram R, Barrios JP, Abreau S, et al. Automated assessment of cardiac systolic function from coronary angiograms with video-based artificial intelligence algorithms. JAMA Cardiol. 2023;8(6):586–594. 10.1001/jamacardio.2023.0968 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Jones WS, Mulder H, Wruck LM, et al. Comparative effectiveness of aspirin dosing in cardiovascular disease. N Engl J Med. 2021;384(21):1981–1990. 10.1056/NEJMoa2102137 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Decentralized clinical trials for drugs, biological products, and devices: guidance for industry, investigators, and other stakeholders. U.S. Food and Drug Administration; 2023. Accessed April 18, 2024. https://www.fda.gov/media/167696/download [Google Scholar]
- 73.Mentz RJ, Anstrom KJ, Eisenstein EL, et al. Effect of torsemide vs furosemide after discharge on all-cause mortality in patients hospitalized with heart failure: the TRANSFORM-HF randomized clinical trial. JAMA. 2023;329(3):214–223. 10.1001/jama.2022.23924 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.Van Norman GA. Decentralized clinical trials. JACC Basic Transl Sci. 2021;6(4):384–387. 10.1016/j.jacbts.2021.01.011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75.Cowie MR, Blomster JI, Curtis LH, et al. Electronic health records to facilitate clinical research. Clin Res Cardiol. 2017;106(1):1–9. 10.1007/s00392-016-1025-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.Krumholz HM, Sawano M, Bhattacharjee B, et al. The PAX LC trial: a decentralized, phase 2, randomized, double-blind study of nirmatrelvir/ritonavir compared with placebo/ritonavir for long COVID. Am J Med. Published online May 10, 2024. 10.1016/j.amjmed.2024.04.030 [DOI] [PubMed] [Google Scholar]
- 77.Oikonomou EK, Spatz ES, Suchard MA, Khera R. Individualising intensive systolic blood pressure reduction in hypertension using computational trial phenomaps and machine learning: a post-hoc analysis of randomised clinical trials. Lancet Digit Health. 2022;4(11):e796–e805. 10.1016/S2589-7500(22)00170-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78.Oikonomou EK, Suchard MA, McGuire DK, Khera R. Phenomapping-derived tool to individualize the effect of canagliflozin on cardiovascular risk in type 2 diabetes. Diabetes Care. 2022;45(4):965–974. 10.2337/dc21-1765 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Oikonomou EK, Van Dijk D, Parise H, et al. A phenomapping-derived tool to personalize the selection of anatomical vs. functional testing in evaluating chest pain (ASSIST). Eur Heart J. 2021;42(26):2536–2548. 10.1093/eurheartj/ehab223 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80.Oikonomou EK, Thangaraj PM, Bhatt DL, et al. An explainable machine learning-based phenomapping strategy for adaptive predictive enrichment in randomized clinical trials. NPJ Digit Med. 2023;6(1):217. 10.1038/s41746-023-00963-z [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Lund BD, Wang T, Mannuru NR, Nie B, Shimray S, Wang Z. ChatGPT and a new academic reality: artificial intelligence-written research papers and the ethics of the large language models in scholarly publishing. J Assoc Inf Sci Technol. 2023;74(5):570–581. 10.1002/asi.24750 [DOI] [Google Scholar]
- 82.Koller D, Beam A, Manrai A, et al. Why we support and encourage the use of large language models in NEJM AI submissions. NEJM AI. 2023;1(1):AIe2300128. 10.1056/AIe2300128 [DOI] [Google Scholar]
- 83.Guide for authors. Journal of the American College of Cardiology. 2024. Accessed April 22, 2024. https://www.sciencedirect.com/journal/journal-of-the-american-college-of-cardiology/publish/guide-for-authors
- 84.Thorp HH, Vinson V. Change to policy on the use of generative AI and large language models. Editor’s blog. Science. 2023. Accessed April 22, 2024. https://www.science.org/content/blog-post/change-policy-use-generative-ai-and-large-language-models
- 85.Beam AL, Drazen JM, Kohane IS, Leong TY, Manrai AK, Rubin EJ. Artificial intelligence in medicine. N Engl J Med. 2023;388(13):1220–1221. 10.1056/NEJMe2206291 [DOI] [PubMed] [Google Scholar]
- 86.Khera R, Butte AJ, Berkwits M, et al. AI in medicine—JAMA’s focus on clinical outcomes, patient-centered care, quality, and equity. JAMA. 2023;330(9):818–820. 10.1001/jama.2023.15481 [DOI] [PubMed] [Google Scholar]
- 87.Artificial intelligence & medical products: how CBER, CDER, CDRH, and OCP are working together. U.S. Food and Drug Administration. 2024. Accessed April 17, 2024. https://www.fda.gov/media/177030/download?attachment [Google Scholar]
- 88.Artificial intelligence and machine learning (AI/ML)-enabled medical devices. U.S. Food and Drug Administration. Accessed May 20, 2024. https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-aiml-enabled-medical-devices [Google Scholar]
- 89.ElZarrad MK, Lee AY, Purcell R, Steele SJ. Advancing an agile regulatory ecosystem to respond to the rapid development of innovative technologies. Clin Transl Sci. 2022;15(6):1332–1339. 10.1111/cts.13267 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 90.Good machine learning practice for medical device development: guiding principles. U.S. Food and Drug Administration; 2021. Accessed April 17, 2024. https://www.fda.gov/medical-devices/software-medical-device-samd/good-machine-learning-practice-medical-device-development-guiding-principles [Google Scholar]
- 91.Qualification process for drug development tools: guidance for industry and FDA staff U.S. Food and Drug Administration; 2020. Accessed April 17, 2024. https://www.fda.gov/media/133511/download [Google Scholar]
- 92.The Innovative Science and Technology Approaches for New Drugs (ISTAND) pilot program. U.S. Food and Drug Administration; 2024. Accessed April 17, 2024. https://www.fda.gov/drugs/drug-development-tool-ddt-qualification-programs/innovative-science-and-technology-approaches-new-drugs-istand-pilot-program [Google Scholar]
- 93.Qualification of medical device development tools: guidance for industry, tool developers, and food and drug administration staff. U.S. Food and Drug Administration; 2023. Accessed April 17, 2024. https://www.fda.gov/media/87134/download [Google Scholar]
- 94.Framework for the use of digital health technologies in drug and biological product development. U.S. Food and Drug Administration; 2023. Accessed April 17, 2024. https://www.fda.gov/media/166396/download?attachment [Google Scholar]
- 95.Finlayson SG, Subbaswamy A, Singh K, et al. The clinician and dataset shift in artificial intelligence. N Engl J Med. 2021;385(3):283–286. 10.1056/NEJMc2104626 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 96.Predetermined change control plans for machine learning-enabled medical devices: guiding principles. U.S. Food and Drug Administration; 2023. Accessed May 20, 2024. https://www.fda.gov/media/173206/download?attachment [Google Scholar]
- 97.Obermeyer Z, Powers B, Vogeli C, Mullainathan S. Dissecting racial bias in an algorithm used to manage the health of populations. Science. 2019;366(6464):447–453. 10.1126/science.aax2342 [DOI] [PubMed] [Google Scholar]
- 98.Irvin J, Rajpurkar P, Ko M, et al. CheXpert: a large chest radiograph dataset with uncertainty labels and expert comparison. 2019. arXiv.org. 2019. Accessed May 15, 2024. https://arxiv.org/abs/1901.07031 [Google Scholar]
- 99.Johnson AEW, Bulgarelli L, Shen L, et al. MIMIC-IV, a freely accessible electronic health record dataset. Sci Data. 2023;10(1):1. 10.1038/s41597-022-01899-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 100.Abràmoff MD, Tarver ME, Loyo-Berrios N, et al. Considerations for addressing bias in artificial intelligence for health equity. NPJ Digit Med. 2023;6(1):170. 10.1038/s41746-023-00913-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101.LM Studio. 2024. Accessed July 15, 2024. https://lmstudio.ai/
