Skip to main content
JAMIA Open logoLink to JAMIA Open
. 2025 Oct 8;8(5):ooaf104. doi: 10.1093/jamiaopen/ooaf104

Assessing the quality of reporting in artificial intelligence/machine learning research for cardiac amyloidosis

Asiful Arefeen 1,2,#, Simar Singh 3,4,#, Crystal Razavi 5, Hassan Ghasemzadeh 6, Sandesh Dev 7,8,
PMCID: PMC12507469  PMID: 41069357

Abstract

Objectives

Despite the rapid development of AI in clinical medicine, reproducibility and methodological limitations hinder its clinical utility. In response, MINimum Information for Medical AI Reporting (MINIMAR) standards were introduced to enhance publication standards and reduce bias, but their application remains unexplored. In this review, we sought to assesses the quality of reporting in AI/ML studies of cardiac amyloidosis (CA) an increasingly important cause of heart failure.

Materials and Methods

Using PRISMA-ScR guidelines, we performed a scoping review of English-language articles published through May 2023 which applied AI/ML techniques to diagnose or predict CA. Non-CA studies and those with selective feature sets were excluded. Two researchers independently screened and extracted data. In all, 20 studies met criteria and were assessed for adherence to MINIMAR standards.

Results

The studies showed variable compliance with MINIMAR. Most reported participant age (90%) and gender (85%), but only 25% included ethnic or racial data, and none provided socioeconomic details. The majority (95%) developed diagnostic models, yet only 85% clearly described training features, and 20% addressed missing data. Model evaluation revealed gaps; 80% reported internal validation, but only 20% conducted external validation.

Discussion and Conclusion

This study, one of the first to apply MINIMAR criteria to ML research in CA, reveals significant variability and deficiencies in reporting, particularly in patient demographics, model architecture, and evaluation. These findings underscore the need for stricter adherence to standardized reporting guidelines to enhance the reliability, generalizability, and clinical applicability of ML/AI models in CA.

Keywords: cardiac amyloidosis, ATTR, artificial intelligence, machine learning

Introduction

Cardiac amyloidosis (CA) is a disease of heart muscle caused by the extracellular deposition of amyloid fibrils that is responsible for up to 15% of heart failure cases. The most common subtypes include transthyretin (ATTR) and immunoglobulin light chain (AL) amyloidosis, and ATTR represents the majority of cases.1,2 ATTR CA disproportionately impacts older, male, and Black persons. In addition, 3%-4% of US Black person are carriers for TTR mutation and carrier status is associated with an average 2 years of reduced life expectancy.3 Early diagnosis of CA is critical to prevent progression and initiate disease-modifying therapies, yet clinical recognition remains delayed due to nonspecific presentations and reliance on expert-driven diagnostic pathways.4–6 As a result, there is growing interest in applying machine learning (ML) and artificial intelligence (AI) to accelerate CA detection using data from imaging, electronic health records (EHRs), and clinical biomarkers.7–9

Although ML methods show promise for improving diagnostic efficiency, substantial concerns persist regarding the reproducibility, generalizability, and transparency of these models.10,11 Several studies across multiple medical specialties, including obstetrics, orthopedics, oncology, rheumatology, and radiology have shown that critical details about data handling, model development including patient demographics, and model evaluation are often omitted, limiting the ability of clinicians and researchers to appraise, reproduce, and apply the results clinically.12–20 For example, Bozkurt et al found that only 36% of studies reported race/ethnicity while a mere 8% reported on the socioeconomic descriptors of populations used to generate ML models.13

To address these concerns, several reporting guidelines have been developed. Notably, the MINimum Information for Medical AI Reporting (MINIMAR) guideline was published in 2020 to establish baseline expectations for AI/ML studies in healthcare.21 Complementary reporting standards have been developed including CONSORT-AI (2020), CLAIM (2020), and TRIPOD-AI (2024), which provide guidelines for reporting predictive modeling, clinical trials, and medical imaging, respectively.22–24 Despite the availability of these tools, evidence suggests that adherence to reporting standards remains inconsistent across medical specialties, including a recent study noting that 93% of ML models in oncology failed to provide details on how to access model code and 92% did not report strategies for model pretraining.12,17–20,25

In this scoping review, we systematically evaluated the quality of reporting in ML studies focused on CA using the MINIMAR framework. We chose MINIMAR as it has broad applicability across data types and was one of the earliest proposed standards for AI in healthcare and was available when we began this work in 2022. Here, we aim to quantify adherence to MINIMAR criteria across critical reporting domains by ML studies focused on CA. Our findings provide a foundation for improving the rigor, reproducibility, and transparency of ML applications in this emerging field of cardiovascular medicine.

Materials and methods

We conducted a scoping review in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews (PRISMA-ScR) checklist.26 Our objective was to assess how well ML studies in CA adhere to reporting standards defined by the MINIMAR guideline, which outlines essential information for evaluating ML models in healthcare.

Eligibility criteria

We included peer-reviewed studies and preprints published in English through May 2023 that employed ML to develop diagnostic or prognostic models for CA using patient data, clinical imaging, or laboratory tests. Exclusion criteria included studies focused exclusively on non-cardiac forms of amyloidosis or those that used domain-specific knowledge to pre-select highly curated features.

Information sources and search strategy

We searched PubMed, MEDLINE, EMBASE, Elsevier, Google Scholar, and preprint servers (bioRxiv, arXiv) using a standardized query containing terms related to “cardiac amyloidosis,” “machine learning,” “deep learning,” and “artificial intelligence.” Boolean logic was applied to expand the search. Reference lists of included articles were reviewed to identify additional eligible studies.

Study selection

Two independent reviewers screened titles and abstracts using the Covidence systematic review software.27 Full-text articles were obtained for those meeting initial criteria, with duplicates identified and removed. The titles and abstracts of the remaining citations were then screened to determine their eligibility based on the predefined inclusion and exclusion criteria. Any discrepancies or disagreements between the reviewers were resolved through consensus. The full texts of the articles that met the initial screening criteria were obtained and thoroughly reviewed in detail. During the full-text review, any studies that did not meet the inclusion criteria were excluded, and the reasons for their exclusion were documented.

Data charting

Data were extracted into a standardized form reflecting the domains of the MINIMAR guideline. These domains included:

  1. General Study Information—authorship, country, publication year, and study setting.

  2. Patient Demographics—age, sex, race/ethnicity, and socioeconomic indicators.

  3. Model Architecture—type of ML algorithm used, input features, training and test split, handling of missing data, and model optimization techniques.

  4. Model Evaluation—optimization, internal and external validation, performance metrics, and transparency (code and data availability).

Results

We identified 20 studies that met the inclusion criteria for this scoping review (Figure 1; Tables S1 and S5). The majority were published after the release of MINIMAR in 2020 (n = 16). These studies applied ML approaches to the diagnosis or prediction of CA using structured clinical data, imaging, or electronic health records. Using the MINIMAR framework, we evaluated adherence to 4 reporting domains: general study information, patient demographics, model architecture, and model evaluation.

Figure 1.

PRISMA 2020 flow diagram summarizing study selection. Records were identified from medical databases (n = 57), brute force searching (n = 22), and open access searching (n = 9), totaling 88 studies imported. After removing duplicates (n = 59), 35 abstracts were screened, with 11 excluded. Twenty-four full-text studies were assessed, and 4 excluded for wrong design, population, or setting. Twenty studies were included in the final scoping review.

Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) 2020 flow diagram for study selection.

General study information

Across the category of general study information (Table 1 and Table S1), all 20 studies (n = 20, 100%) stated their objectives and described the data sources used. Most (n = 17, 85%) reported the study setting, such as academic medical centers or hospital networks. However, 15% (n = 3) of studies did not describe the institutional or clinical context in which the data were collected. Patient selection criteria were clearly specified in 75% (n = 15) of studies, while 25% (n = 5) did not provide sufficient detail to determine how patients were chosen.

Table 1.

Summary statistics of MINIMAR criteria-quality of reporting (Metadata).

Key attributes Reported Not reported
Well-defined objectives 20 (100%) 0 (0%)
Well-defined patient selection criteria 15 (75%) 5 (25%)
Study setting 17 (85%) 3 (15%)
Data sources 20 (100%) 0 (0%)
Dataset details
 Age 18 (90%) 2 (10%)
 Gender 17 (85%) 3 (15%)
 Race/Ethnicity 12 (60%) 8 (40%)
 Socioeconomic status (SES) 0 (0%) 20 (100%)
Gold standard 18 (90%) 2 (10%)
List of all features (N/A for articles which did not used features) 5 (45%) 6 (55%)
Model optimization (N/A for articles which did not develop models) 13 (65%) 7 (35%)
Validation
 Internal 17 (85%) 3 (15%)
 External 6 (30%) 14 (70%)
Dealing with missing data (N/A for image/video/ECG data) 4 (44%) 5 (56%)
Transparency
 Code availability 5 (25%) 15 (75%)
 Data availability 11 (55%) 9 (45%)
Model type
 Decision trees and variants28–35 n = 8
 Neural network36–42 n = 7
 Clustering35,43,44 n = 3
 Logistic regression29,45 n = 2
 Support vector machine46,47 n = 2
Data sources
 EHR28,30,45 n = 3
 ECG31,36,41,43 n = 4
 Echocardiogram36,38 n = 2
 Lab parameters and test results29,44 n = 2
 PET image39 n = 1
 CMR image37,40,42 n = 3
 Claims data34,45 n = 2

Patient demographic characteristics

In terms of reporting patient demographics (Table 1 and Table S2), 90% (n = 18) of studies included information on participant age, and 85% (n = 17) reported gender. However, only 25% (n = 5) of studies reported racial or ethnic data, and none included socioeconomic indicators such as insurance status, income, or education level.

Model architecture

Model architecture was generally well described (Table 1 and Table S3). Nearly all studies (n = 19, 95%) developed diagnostic models, while only one study focused on prognosis. Most (n = 17, 85%) detailed the features used for training, and 90% (n = 18) reported how the dataset was split for model development and testing. Nevertheless, only 20% (n = 4) of studies addressed how missing data were handled, an important component for reproducibility and model robustness.

Model evaluation

Model optimization was performed for more than half of studies (n = 13, 65%); however, not all studies reported the optimization method (n = 10, 50%) (Table 1 and Table S4). Validation methodology was reported for almost all studies (n = 19, 95%). Moreover, 16 studies performed internal validation (n = 16, 80%) whereas only 4 performed external validation (n = 4, 20%); 4 studies performed both internal and external validation (n = 4, 20%). Model code was made available for 5 studies (n = 5, 25%). Data were made available for 11 studies (n = 11, 55%).

Discussion

This scoping review assessed the quality of reporting in ML-based studies of CA using the MINIMAR framework. We found occasional deficiencies in general study information (ie, study setting, patient selection) but noted significant deficiencies in the reporting of demographic characteristics (ie, race/ethnicity, socioeconomic indicators), model architecture (ie, missing data), and model evaluation (ie, transparency). These limitations mirror similar findings from other medical fields, including radiology, oncology, and rheumatology, where studies have shown incomplete reporting across multiple domains.12,17–20,25 This study is the first to systematically evaluate ML reporting in CA and, to our knowledge, the first to apply the MINIMAR framework to assess ML literature within a specific clinical domain.

Demographic reporting was among the weakest areas. Only 1 in 4 studies included racial or ethnic information, and none reported socioeconomic data. These omissions are particularly concerning given the established racial disparities in the incidence of transthyretin amyloidosis48 and the socioeconomic barriers to diagnosis and care. Black Americans develop CA at twice the rate of CA compared to the White population.48 Furthermore, there is under detection in certain US regions such as the South despite over 50% of the Black population residing there, thus highlighting need for improved equity of care.48 Incomplete reporting in this domain limits the generalizability of ML models and perpetuates structural bias.49–51 The inclusion of social determinants of health has been shown to improve models predicting hospital readmission for sepsis.52 Our findings are also consistent with prior analyses of EHR-based ML studies, which reported similarly low rates of demographic disclosure13 and highlight the need to address potential sources of bias early in model development.53

Model architecture was generally well reported with the exception of handling of missing data. Most studies did not report how missingness was handled or whether imputation strategies were employed, potentially biasing model performance and reducing replicability.51 This omission is particularly concerning given the frequency of missing data in real-world clinical datasets, where improper handling can introduce systematic bias and compromise generalizability, especially in CA, where models are often developed on small, heterogeneous cohorts.

We found major deficiencies in the reporting of model evaluation parameters, particularly with transparency and reproducibility. Despite growing calls for enhancing reproducibility in AI/ML science, only 25% of the reviewed studies shared their model code, and 55% provided information on data availability. Without access to code, data, and a clear description of modeling pipelines, independent replication is impossible, resulting in decreased trustworthiness and clinical adoption.54–57 This aligns with other reports that found incomplete reporting related of transparency measures in AI studies.11,17,55,58

Model evaluation reporting was also often deficient with regard to validation practices. Internal validation was commonly reported, but only 20% of studies reported external validation using independent datasets, a vital step in demonstrating model generalizability. This rate is slightly higher than that reported in imaging-based AI literature but remains far below what is needed for clinical deployment.55,58,59 Given the rarity and clinical heterogeneity of CA, external validation using multi-institutional or prospective datasets is particularly important.

Our review has several strengths and additional considerations. We applied a standardized evaluation framework (MINIMAR) to assess the quality of studies and included all available studies of ML in CA, including both peer-reviewed and preprint literature. Second, most of the studies that were evaluated were reported after MINIMAR was published in 2020. Studies published after MINIMAR release trended toward improved reporting, particularly with regard to model evaluation. Our analysis is limited by the small number of eligible studies, precluding formal statistical comparisons particularly in comparing studies completed before and after the release of MINIMAR. It should be acknowledged that other reporting standards have recently been created or updated to incorporate ML studies, and MINIMAR is not a universally accepted reporting standard. However, the goal of this work is to evaluate the quality of ML reporting to inform this rapidly evolving specialty in cardiovascular medicine. In addition, given the explosive growth of ML studies in medicine, our study does not address studies published since 2023. Fourth, other reporting standards tools such as TRIPOD-AI and CLAIM might offer complementary insights not captured in our evaluation. For example, TRIPOD-AI, which was released in 2024, is intended to be a comprehensive guide to reporting multivariable predictive models.22 It shares the key reporting domains espoused in MINIMAR (with emphasis on fairness, openness and trustworthiness) and provides additional guidance and on how to report each aspect of model development. Nonetheless, because of the similarity of MINIMAR and TRIPOD-AI, and the availability of MINIMAR before TRIPOD-AI, our choice of MINIMAR was appropriate to provide a snapshot of the ML studies in this field. We believe there is concordance between MINIMAR and CLAIM standards, and furthermore our review was not exclusively focused on medical imaging.

Future directions

Given the recent introduction of several new reporting standards for AI studies in healthcare, it is possible that the quality of reporting of AI classification and prediction models may improve over time. However, current research suggests that their adoption remains low17,60 and is predicated on self-reporting, which has been shown to be prone to error.61,62 Widespread utilization of standardized reporting practices will require buy-in from researchers and enforcement from publishers and funding agencies63 to ensure systematic change. As new AI “solutions” come to market to improve diagnosis and treatment of CA, it will be critical to carefully examine the data and the population those data represent. Better reporting of ML studies will accelerate the safe, equitable, and effective integration of these technologies and lead to improvements in the health of patients with CA.

Conclusion

This study is one of the first to systematically assess the quality of reporting of AI/ML research within the CA field, a rapidly developing therapeutic area in medicine. By employing the MINIMAR reporting standards as a benchmark, we have adopted a rigorous and broadly applicable framework to evaluate the reporting quality of studies in this field. This systematic approach has enabled us to identify and document key areas where reporting deficiencies are prevalent, such as in demographic details, model architecture, and evaluation processes. Addressing these deficiencies is crucial for enhancing the transparency and reproducibility of future research in CA.

Supplementary Material

ooaf104_Supplementary_Data

Contributor Information

Asiful Arefeen, School of Computing and Augmented Intelligence, Arizona State University, Tempe, AZ 85281, United States; College of Health Solutions, Arizona State University, Phoenix, AZ 85004, United States.

Simar Singh, Sarver Heart Center, Department of Medicine, University of Arizona College of Medicine, Tucson, AZ, 85724, United States; Medicine Service, Southern Arizona Veterans Affairs Health Care System, Tucson, AZ, 85723, United States.

Crystal Razavi, Sarver Heart Center, Department of Medicine, University of Arizona College of Medicine, Tucson, AZ, 85724, United States.

Hassan Ghasemzadeh, College of Health Solutions, Arizona State University, Phoenix, AZ 85004, United States.

Sandesh Dev, College of Health Solutions, Arizona State University, Phoenix, AZ 85004, United States; Medicine Service, Southern Arizona Veterans Affairs Health Care System, Tucson, AZ, 85723, United States.

Author contributions

Asiful Arefeen (Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Validation, Visualization, Writing—original draft, Writing—review & editing), Simar Jeet Singh (Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Writing—original draft, Writing—review & editing), Crystal Razavi (Data curation, Formal analysis, Investigation, Writing—original draft), Hassan Ghasemzadeh (Conceptualization, Methodology, Project administration, Supervision, Writing—review & editing), and Sandesh Dev (Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Writing—original draft, Writing—review & editing)

Supplementary material

Supplementary material is available at JAMIA Open online.

Funding

This material is the result of work supported with resources and the use of facilities at the Southern Arizona Veterans Affairs Health Care System Tucson, AZ and Arizona State University, Complex Adaptive Systems Initiative. The contents do not represent the views of the U.S. Department of Veterans Affairs or the United States Government.

Conflicts of interest

The authors have no competing interests or disclosures to declare.

Data availability

The data underlying this article are available in the primary publications identified in the systematic search. No new datasets were generated or analyzed in this study.

References

  • 1. Kittleson MM, Maurer MS, Ambardekar AV, et al.  Cardiac amyloidosis: evolving diagnosis and management: a scientific statement from the American Heart Association. Circulation  2020;142:e7-e22. [DOI] [PubMed] [Google Scholar]
  • 2. Yilmaz A, Bauersachs J, Bengel FM, et al.  Diagnosis and treatment of cardiac amyloidosis: position statement of the German Cardiac Society (DGK). Clin Res Cardiol. 2021;110:479-506. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Selvaraj S, Claggett B, Shah SH, et al.  Cardiovascular burden of the V142I transthyretin variant. JAMA. 2024;331:1824-1833. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Lane T, Fontana M, Martinez-Naharro A, et al.  Natural history, quality of life, and outcome in cardiac transthyretin amyloidosis. Circulation. 2019;140:16-26. [DOI] [PubMed] [Google Scholar]
  • 5. Maurer MS, Schwartz JH, Gundapaneni B, et al. ; ATTR-ACT Study Investigators. Tafamidis treatment for patients with transthyretin amyloid cardiomyopathy. N Engl J Med. 2018;379:1007-1016. [DOI] [PubMed] [Google Scholar]
  • 6. Nativi-Nicolau J, Judge DP, Hoffman JE, et al.  Natural history and progression of transthyretin amyloid cardiomyopathy: insights from attract. ESC Heart Fail. 2021;8:3875-3884. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Haug CJ, Drazen JM.  Artificial intelligence and machine learning in clinical medicine, 2023. N Engl J Med. 2023;388:1201-1208. [DOI] [PubMed] [Google Scholar]
  • 8. Quer G, Arnaout R, Henne M, Arnaout R.  Machine learning and the future of cardiovascular care: JACC state-of-the-art review. J Am Coll Cardiol. 2021;77:300-313. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Johnson KW, Torres Soto J, Glicksberg BS, et al.  Artificial intelligence in cardiology. J Am Coll Cardiol. 2018;71:2668-2679. [DOI] [PubMed] [Google Scholar]
  • 10. Price WN.  Big data and black-box medical algorithms. Sci Transl Med. 2018;10:eaao5333. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Nagendran M, Chen Y, Lovejoy CA, et al.  Artificial intelligence versus clinicians: systematic review of design, reporting standards, and claims of deep learning studies. BMJ. 2020;368:m689. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Belue MJ, Harmon SA, Lay NS, et al.  The low rate of adherence to checklist for artificial intelligence in medical imaging criteria among published prostate MRI artificial intelligence algorithms. J Am Coll Radiol. 2023;20:134-145. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Bozkurt S, Cahan EM, Seneviratne MG, et al.  Reporting of demographic data and representativeness in machine learning models using electronic health records. J Am Med Inform Assoc. 2020;27:1878-1884. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Lu JH, Callahan A, Patel BS, et al.  Assessment of adherence to reporting guidelines by commonly used clinical prediction models from a single vendor: a systematic review. JAMA Netw Open. 2022;5:e2227779. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Navarro CLA, Damen JAA, Takada T, et al.  Completeness of reporting of clinical prediction models developed using supervised machine learning: a systematic review. BMC Med Res Methodol. 2022;22:12. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Corti C, Cobanaj M, Marian F, et al.  Artificial intelligence for prediction of treatment outcomes in breast cancer: systematic review of design, reporting standards, and bias. Cancer Treat Rev. 2022;108:102410. [DOI] [PubMed] [Google Scholar]
  • 17. Smiley A, Villarreal-Zegarra D, Reategui-Rivera CM, Escobar-Agreda S, Finkelstein J.  Methodological and reporting quality of machine learning studies on cancer diagnosis, treatment, and prognosis. Front Oncol. 2025;15:1555247. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Salybekov AA, Wolfien M, Hahn W, Hidaka S, Kobayashi S.  Artificial intelligence reporting guidelines’ adherence in nephrology for improved research and clinical outcomes. Biomedicines. 2024;12:606. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Munguía-Realpozo P, Etchegaray-Morales I, Mendoza-Pinto C, et al.  Current state and completeness of reporting clinical prediction models using machine learning in systemic lupus erythematosus: a systematic review. Autoimmun Rev. 2023;22:103294. [DOI] [PubMed] [Google Scholar]
  • 20. Yang Q, Fan X, Cao X, et al.  Reporting and risk of bias of prediction models based on machine learning methods in preterm birth: a systematic review. Acta Obstet Gynecol Scand. 2023;102:7-14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Hernandez-Boussard T, Bozkurt S, Ioannidis JPA, Shah NH.  MINIMAR (Minimum Information for Medical AI Reporting): developing reporting standards for artificial intelligence in health care. J Am Med Inform Assoc. 2020;27:2011-2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Collins GS, Moons KGM, Dhiman P, et al.  Tripod+AI statement: updated guidance for reporting clinical prediction models that use regression or machine learning methods. BMJ. 2024;385:e078378. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Liu X, Rivera SC, Moher D, Calvert MJ, Denniston AK; SPIRIT-AI and CONSORT-AI Working Group. Reporting guidelines for clinical trial reports for interventions involving artificial intelligence: the CONSORT-AI extension. BMJ. 2020;370:m3164. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Tejani AS, Klontzas ME, Gatti AA, et al. ; July CLAIM 2024 Update Panel. Checklist for Artificial Intelligence in Medical Imaging (CLAIM): 2024 update. Radiol Artif Intell. 2024;6:e240300. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Feller D, Wingbermuhle R, Berg B, et al.  Improvements are needed in the adherence to the tripod statement for clinical prediction models for patients with spinal pain or osteoarthritis: a metaresearch study. J Pain. 2024;25:104624. [DOI] [PubMed] [Google Scholar]
  • 26. Tricco AC, Lillie E, Zarin W, et al.  Prisma Extension for Scoping Reviews (PRISMA-ScR): checklist and explanation. Ann Intern Med. 2018;169:467-473. [DOI] [PubMed] [Google Scholar]
  • 27. Babineau J.  Product Review: Covidence (Systematic Review Software). J Can Health Libr Assoc. 2014;35:68-71. [Google Scholar]
  • 28. Huda A, Castaño A, Niyogi A, et al.  A machine learning model for identifying patients at risk for wild-type transthyretin amyloid cardiomyopathy. Nat Commun. 2021;12:2725. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Agibetov A, Seirer B, Dachs TM, et al.  Machine learning enables prediction of cardiac amyloidosis by routine laboratory parameters: a proof-of-concept study. J Clin Med. 2020;9:1334. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Garg RP, Dong S, Shah SJ, Jonnalagadda SR.  2016. A bootstrap machine learning approach to identify rare disease patients from electronic health records. arXiv, arXiv:609.01586, preprint: not peer reviewed. [Google Scholar]
  • 31. Gnanadurai GJ, Raaza A, Velayutham R, Palani SK, Bramwell EA.  Detection of cardiac amyloidosis on electrocardiogram images using machine learning and deep learning techniques. Comput Intell. 2022;9:554-576. [Google Scholar]
  • 32. Sammani A, Jansen M, de Vries NM, et al.  Automatic identification of patients with unexplained left ventricular hypertrophy in electronic health record data to improve targeted treatment and family screening. Front Cardiovasc Med. 2022;9:768847. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Wu Z-W, Zheng J-L, Kuang L, Yan H.  Machine learning algorithms to automate differentiating cardiac amyloidosis from hypertrophic cardiomyopathy. Int J Cardiovasc Imaging. 2023;39:339-348. [DOI] [PubMed] [Google Scholar]
  • 34. Bruno M, Sheer RL, Reed C, et al.  Clinical characteristics and health care resource use of patients at risk for wild-type transthyretin amyloid cardiomyopathy identified by machine learning model. J Manag Care Spec Pharm. 2023;29:530-540. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Antonopoulos AS, Boutsikou M, Simantiris S, et al.  Machine learning of native T1 mapping radiomics for classification of hypertrophic cardiomyopathy phenotypes. Sci Rep. 2021;11:23596. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Goto S, Mahara K, Beussink‐Nelson L, et al.  Artificial intelligence-enabled fully automated detection of cardiac amyloidosis using electrocardiograms and echocardiograms. Nat Commun. 2021;12:2726. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Martini N, Aimo A, Barison A, et al.  Deep learning to diagnose cardiac amyloidosis from cardiovascular magnetic resonance. J Cardiovasc Magn Reson. 2020;22:84. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Duffy G, Cheng P, Yuan N, et al.  High-throughput precision phenotyping of left ventricular hypertrophy with cardiovascular deep learning. JAMA Cardiol. 2022;7:386-395. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Santarelli MF, Genovesi D, Positano V, et al.  Deep-learning-based cardiac amyloidosis classification from early acquired pet images. Int J Cardiovasc Imaging. 2021;37:2327-2335. [DOI] [PubMed] [Google Scholar]
  • 40. Germain P, Vardazaryan A, Padoy N, et al.  Deep learning supplants visual analysis by experienced operators for the diagnosis of cardiac amyloidosis by CINE-CMR. Diagnostics. 2021;12:69. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41. Grogan M, Lopez‐Jimenez F, Cohen-Shelly M, et al.  Artificial intelligence-enhanced electrocardiogram for the early detection of cardiac amyloidosis. Mayo Clin Proc  2021;96:2768-2778. [DOI] [PubMed] [Google Scholar]
  • 42. Agibetov A, Kammerlander AA, Duca F, et al.  Convolutional neural networks for fully automated diagnosis of cardiac amyloidosis by cardiac magnetic resonance imaging. J Pers Med. 2021;11:1268. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Schrutka L, Anner P, Agibetov A, et al.  Machine learning-derived electrocardiographic algorithm for the detection of cardiac amyloidosis. Heart. 2022;108:1137-1147. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44. Bonnefous L, Kharoubi M, Bézard M, et al.  Assessing cardiac amyloidosis subtypes by unsupervised phenotype clustering analysis. J Am Coll Cardiol. 2021;78:2177-2192. [DOI] [PubMed] [Google Scholar]
  • 45. García-García E, González-Romero GM, Martín-Pérez EM, Zapata Cornejo EdD, Escobar-Aguilar G, Cárdenas Bonnet MF.  Real-world data and machine learning to predict cardiac amyloidosis. Int J Environ Res Public Health. 2021;18:908. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Eckstein J, Moghadasi N, Körperich H, et al.  A machine learning challenge: detection of cardiac amyloidosis based on bi-atrial and right ventricular strain and cardiac function. Diagnostics. 2022;12:2693. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Jiang S, Zhang L, Wang J, et al.  Differentiating between cardiac amyloidosis and hypertrophic cardiomyopathy on non-contrast cine-magnetic resonance images using machine learning-based radiomics. Front Cardiovasc Med. 2022;9:1001269. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48. Gilstrap LG, Dominici F, Wang Y, et al.  Epidemiology of cardiac amyloidosis-associated heart failure hospitalizations among fee-for-service Medicare beneficiaries in the United States. Circ Heart Fail. 2019;12:e005407. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49. Arora A, Alderman JE, Palmer J, et al.  The value of standards for health datasets in artificial intelligence-based applications. Nat Med. 2023;29:2929-2938. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. Tipton K, Leas BF, Flores E, et al.  Impact of Healthcare Algorithms on Racial and Ethnic Disparities in Health and Healthcare. Technical Report 24-EHC004. Agency for Healthcare Research and Quality (US); December 2023. [PubMed] [Google Scholar]
  • 51. Gianfrancesco MA, Tamang S, Yazdany J, Schmajuk G.  Potential biases in machine learning algorithms using electronic health record data. JAMA Intern Med. 2018;178:1544-1547. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52. Amrollahi F, Shashikumar SP, Meier A, Ohno-Machado L, Nemati S, Wardi G.  Inclusion of social determinants of health improves sepsis readmission prediction models. J Am Med Inform Assoc. 2022;29:1263-1270. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53. Moons KGM, Damen JAA, Kaul T, et al.  PROBAST+AI: an updated quality, risk of bias, and applicability assessment tool for prediction models using regression or artificial intelligence methods. BMJ. 2025;388:e082505. Mar. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54. Char DS, Shah NH, Magnus D.  Implementing machine learning in health care – addressing ethical challenges. N Engl J Med. 2018;378:981-983. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55. Mooghali M, Stroud AM, Yoo DW, et al.  Trustworthy and ethical AI-enabled cardiovascular care: a rapid review. BMC Med Inform Decis Mak. 2024;24:247. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56. Abgrall G, Holder AL, Dagdia ZC, Zeitouni K, Monnet X.  Should AI models be explainable to clinicians?  Crit Care. 2024;28:301. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57. Alderman JE, Palmer J, Laws E, et al.  Tackling algorithmic bias and promoting transparency in health datasets: the standing together consensus recommendations. Lancet Digit Health. 2025;7:e64-e88. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58. Kim DW, Jang HY, Kim KW, Shin Y, Park SH.  Design characteristics of studies reporting the performance of artificial intelligence algorithms for diagnostic analysis of medical images: results from recently published papers. Korean J Radiol. 2019;20:405-410. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59. Spaanderman DJ, Marzetti M, Wan X, et al.  AI in radiological imaging of soft-tissue and bone tumours: a systematic review evaluating against claim and future-AI guidelines. EBioMedicine. 2025;114:105642. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60. Critelli B, Hassan A, Lahooti I, et al.  A systematic review of machine learning-based prognostic models for acute pancreatitis: towards improving methods and reporting quality. PLoS Med. 2025;22:e1004432. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61. Koçak B, Keleş A, D’Antonoli TA.  Self-reporting with checklists in artificial intelligence research on medical imaging: a systematic review based on citations of claim. Eur Radiol. 2024;34:2805-2815. [DOI] [PubMed] [Google Scholar]
  • 62. Koçak B, Köse F, Keleş A, Şendur A, Meşe İ, Karagülle M.  Adherence to the Checklist for Artificial Intelligence in Medical Imaging (CLAIM): an umbrella review with a comprehensive two-level analysis. Diagn Interv Radiol. 2025;31:440-455. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63. Klontzas ME.  Reporting checklists as compulsory supplements to artificial intelligence manuscript submissions. Diagn Interv Radiol. 2025;31:17-18. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

ooaf104_Supplementary_Data

Data Availability Statement

The data underlying this article are available in the primary publications identified in the systematic search. No new datasets were generated or analyzed in this study.


Articles from JAMIA Open are provided here courtesy of Oxford University Press

RESOURCES