Skip to main content
. 2021 Sep 17;10(9):e27799. doi: 10.2196/27799

Table 2.

Data extraction elements.

Category Elements appraised
Reviewer information
  • Reviewer name

  • Reviewer comments

Bibliometrics
  • First and last name of the first author

  • Title

  • Source

  • Year of publication

  • Country

  • Status of publication

Primary care function (adapted from Kueper et al [15])
  • Diagnostic decision support: artificial intelligence–assisted diagnostics

  • Treatment decision support: artificial intelligence–assisted treatment, including remote management of care

  • Referral support: artificial intelligence–assisted support for any portion of the referral process, especially for direct referrals of patients to specialist services

  • Scheduling assistance: models for optimizing clinic schedules and overbooking

  • Future state prediction: artificial intelligence offering predictions about the future state, such as consult service utilization or prognosis of existing conditions. (this excludes predictions of one’s chances of developing a health condition in the near term, which falls under diagnostic decision support)

  • Health care utilization analyses: artificial intelligence extracts information retrospectively to understand more about the current processes or interactions within a health care system

  • Knowledge base and ontology construction or use

  • Information extraction: artificial intelligence extracts knowledge from structured or unstructured data sources

  • Descriptive information provision: Artificial intelligence summarizes existing data in interpretable or useful ways

  • Other: function not represented above, but specifics of function will still be recorded in case a new category emerges


Author-reported intended end-users
  • The intended user of the artificial intelligence product, including but not limited to patients, physicians, nurses, nurse practitioners, administrators, researchers, others, and unknown (if an end-user is not specified as the tool was still in development, a researcher was designated)

Target health condition (adapted from Kueper et al [15])
  • General

  • Diabetes

  • Cancer, non-skin

  • Heart valves, murmurs

  • Musculoskeletal/joint

  • Dementia, cognitive impairment

  • Lung apnea, chronic obstructive pulmonary disease

  • Chronic disease, frailty

  • Skin cancer

  • Stroke, neurological

  • Psychiatric

  • Coronary artery disease

  • Heart failure

  • Hypertension

  • Other cardiovascular disease

  • Gastrointestinal/liver

  • Ear, nose, and throat

  • Eye and retina

  • Trauma, emergency surgery

  • Infection

  • Metabolic

  • Kidney and urinary tract

  • Immunization, reactions

  • Skin disorders

  • Obesity

  • Pediatric/developmental

  • Other

Data set
  • Size: number of unique patients

  • Time period if applicable

  • Source of data:

    • Electronic health record

    • National registry

    • Claims

    • Remote monitoring devices (ie, smart watch or mobile phone)

    • Other (specified)

    • Unknown

  • Number of institutions: single or multiple

  • Setting (urban, rural, both, or unknown): We use the United States Census’ County Classification Lookup Table [34] to determine whether a certain area was urban or rural. If there were multiple locations, we selected both.

Compliance with “Ethics Guidelines for Trustworthy AI” [35]: which of the 7 elements were addressed (yes/no)?
  • Human agency and oversight: how well does the algorithm support human decision-making and permit oversight on its predictions?

  • Technical robustness and safety: how well-suited is the algorithm for its intended use? How well does it mitigate harm?

  • Privacy and data governance: how well does the algorithm’s data ingestion and analysis pipeline respect patient privacy (eg, HIPAA compliance) and enforce safeguards against unpermitted access?

  • Transparency: does the artificial intelligence algorithm explain reasons for its outputs in a traceable and interpretable way?

  • Diversity, nondiscrimination, and fairness: how biased is the algorithm with regard to its performance? How easy is it for stakeholders to provide feedback on the algorithm’s performance for its continuous development?

  • Societal and environmental well-being: what are the societal (eg, dehumanizing relationships) and ecological (eg, energy consumption) impacts of the algorithm?

  • Accountability: who is held responsible to ensure the algorithm’s development, outcomes, harm, and regulation?

Model fairness and focus on health equity: is the main purpose of the study specifically outlined to improve health for a vulnerable population (yes/no)? Must be explicitly stated in the introduction or abstract as motivation for the paper to focus on at least 1 vulnerable population (though there may be other populations studied as well) defined by any of the following categories which are largely based off of the NIMHD Research Framework [36]:
  • Place of residence (eg, rural)

  • Race, ethnicity (eg, Black African American or Latinx)

  • Occupation (eg, coal miners)

  • Gender, sex (eg, transgender)

  • Religion (eg, Amish)

  • Education (eg, low)

  • Socioeconomic status (eg, low income)

  • Social capital (eg, isolation)

  • Does the study include key variables that could reflect disparities across protected classes (eg, age, sex, or race/ethnicity)?

  • If reported, do they include these variables in their evaluation (eg, subgroup analysis to demonstrate equal performance)?

  • Existing biases: does the study discuss biases or potential repercussions related to vulnerable populations? [9]

  • Historical bias (ie, data retrieval)

  • Representation bias (ie, population representation)

  • Measurement bias

  • Aggregation bias

  • Evaluation bias

  • Deployment bias

  • Bias mitigation: does the study attempt to reduce existing biases, either explicitly or implicitly? If so, what methodology do they employ?

  • Preoutput (changes to the algorithm or input data)

  • Postoutput (user education, transparency, and specifying the use case)

  • Other

Stage of the study
  • Methodological development: generation of novel artificial intelligence methods or modification of existing artificial intelligence methods to accomplish a task relevant to primary care.

  • Retrospective data analysis or model development: developed an artificial intelligence model trained on retrospectively collected data to identify trends or perform a task that awaits prospective validation.

  • Evaluation: artificial intelligence implemented in the intended setting as part of a pilot study, such as a prospective cohort study or randomized controlled trial.

  • Postimplementation: assessing the impact of an artificial intelligence implementation after officially deployed in its intended setting.