Abstract
Background. It is unclear how to optimize the user interface and user experience of cancer screening artificial intelligence (AI) tools for clinical decision-making in primary care. Methods. We developed an electronic survey for US primary care clinicians to assess 1) general attitudes toward AI in cancer screening and 2) preferences for various aspects of AI model deployment in the context of colorectal, breast, and lung cancer screening. We descriptively analyzed the responses. Results. Ninety-nine surveys met criteria for analysis out of 733 potential respondents (response rate 14%). Ninety (>90%) somewhat or strongly agreed that their medical education did not provide adequate AI training. A plurality (52%, 39%, and 37% for colon, breast, and lung cancers, respectively) preferred that AI tools recommend the interval to the next screening as compared with the 5-y probability of future cancer diagnosis, a binary recommendation of “screen now,” or identification of suspicious imaging findings. In terms of workflow, respondents preferred generating a flag in the electronic health record to communicate an AI prediction versus an interactive smartphone application or the delegation of findings to another healthcare professional. No majority preference emerged for an explainability method for breast cancer screening. Limitations. The sample was primarily obtained from a single health care system in the Northeast. Conclusions. Providers indicated that AI models can be most helpful in cancer screening by providing prescriptive outputs, such as recommended intervals until next screening, and by integrating with the electronic health record. Implications. A preliminary framework for AI model development in cancer screening may help ensure effective integration into clinical workflow. These findings can better inform how healthcare systems govern and receive reimbursement for services that use AI.
Highlights
Clinicians do not feel their undergraduate or graduate medical education has properly prepared them to engage with AI in patient care.
We provide a preliminary framework for deploying AI models in primary care–based cancer screening.
This framework may have implications for health system governance and provider reimbursement in the age of AI.
Keywords: artificial intelligence, primary care, cancer screening
Artificial intelligence (AI) and machine learning (ML) have shown promise in cancer screening. Several models have demonstrated strong predictive performance in identifying patients at high risk for colorectal,1–4 breast,5–8 and lung9–12 cancers. Previous attempts to understand physician perceptions of AI-based tools have revealed generally positive sentiments, with respondents optimistic that AI can improve diagnostic capabilities and efficiency in clinical workflows.13–16 Common concerns raised included the lack of explainability in recommendations from ML models and poor interpretability of model uncertainty.17–21 Amann et al. 22 describe explainability as “a characteristic of an AI-driven system allowing a person to reconstruct why a certain AI came up with the presented predictions.”
Given the contribution of primary care providers (PCPs) toward decreased cancer morbidity and mortality,23–25 there is an opportunity to optimize the use of AI in cancer screening in the primary care setting. We developed a conceptual model to guide this study. Before the advent of AI, providers used a shared decision-making framework with their patients26–29 with multiple information sources such as 1) evidence-based population-based guidelines, such as those from the United States Preventative Services Task Force (USPSTF) 30 ; 2) the patient’s individual risk factors for cancer; and 3) patient values and preferences31–34 (Figure 1). Using these 3 criteria, clinicians could risk stratify their patients, order additional diagnostic workup, and/or schedule their patient’s next cancer screening. AI-based clinical decision tools offer a new source of information but it is unclear how software developers and health systems should deploy these tools to meaningfully inform clinical decision making. 24
Figure 1.
Conceptual model depicting the interplay between individual patient risk factors, general population-based guidelines, previous cancer screening, patient preferences, and a hypothetical AI model when making a personalized cancer screening recommendation in the primary care setting. The AI model outputs a personalized prediction regarding cancer screening, the result of which could be communicated to the primary care provider (PCP) in a variety of ways (“notification system in clinical workflow”). This study aims to determine which specific prediction or functionality will best inform personalized cancer screening and how to implement AI technology in the clinical workflow. These aims are labeled with a gold star. Arrows represent the conduit by which information is communicated to the PCP so a clinical decision can be made.
Knowledge gaps that must be addressed for software developers include whether to create AI models that generate neutral statements, such as an individual patient’s absolute 10-y probability of colorectal cancer, versus models that generate prescriptive outputs, such as the recommended time until a specific patient’s next colonoscopy. Given the “black box” nature of many AI tools, it is also unclear how models should justify their outputs, which has implications for clinical decision-making and clinician trust. 35 Lastly, due to clinicians’ busy patient schedules and documentation responsibilities, these tools cannot reach their potential if health systems do not appropriately deploy them in the clinical workflow.36,37 It remains an open question how to ensure these tools are an asset in clinical workflow, rather than a hindrance or distraction.
This survey study aims to describe the general attitudes and preferences of PCPs regarding personalized AI-based clinical decision tools for which the USPSTF has reported population-based cancer screening guidelines: colorectal, breast, and lung cancers. Answering these questions will inform a preliminary framework for the user interface and user experience (UI/UX) of such tools in the primary care setting. A standardized framework for AI tools not only increases the likelihood of better patient care but also carries consequences for healthcare policy and practice.
Methods
Eligibility Criteria and Study Design
We surveyed attending physicians, family medicine residents, and advanced practice providers (APPs) using an electronic survey instrument developed on the Qualtrics (Provo, UT, USA) platform. Participants were eligible to complete the form if they were attending physicians or APPs practicing in primary care or if they were family medicine residents. We distributed the survey to providers fitting this criterion at the Hospital of the University of Pennsylvania (HUP), an urban academic medical center, and Lancaster General Hospital (LGH), a regional hospital affiliated with the University of Pennsylvania. We also employed convenience sampling to distribute the survey to other PCPs within our professional network in the United States. All attending physicians or APPs had trained in internal medicine or family medicine. This study was exempt from approval of the institutional review board at the University of Pennsylvania.
Instrument Development
The survey was separated into 3 sections: 1) demographics and background of respondents; 2) general attitudes and preferences towards AI, including the expected impacts of AI on health care utilization related to cancer screening; 3) preferences for prediction outputs for hypothetical AI models, implementation of such models in clinical workflow, and methods of AI model explainability assessed with clinical vignettes (Appendix). Questions and clinical vignettes were informed by our conceptual model of clinical decision-making (Figure 1), expert input, and prior literature on attitudes toward AI. The survey was piloted and iteratively revised based on primary care and specialty clinician feedback before dissemination.
Terminology
Given the potential of care managers and nurse navigators to provide patient-centered care as a part of a large multidisciplinary health care team38–42 as well as the study team’s personal experience working with such individuals in oncology care, we wished to know if care managers could help deploy AI tools. We defined the term “licensed care manager” as a “non-physician licensed healthcare provider who can directly coordinate additional lab work or imaging studies with a patient.” We intentionally did not define AI or ML in the survey introduction because we wished to understand respondents’ preconceived notions about AI and ML. When referencing “AI” in the survey, “AI” encompassed both AI and ML tools.
General Demographics and Respondent Background
Section 1 of the survey included the participant’s date of birth, health care role, medical specialty, presence of fellowship training (if applicable), years in independent practice, whether the participant had spent at least 50% of their clinical time in a primary care context in the last year, and whether the participant had previously used AI tools in a clinical context.
Attitudes toward AI and ML
Section 2 of the survey contained 15 Likert scale–based questions regarding general attitudes toward AI and/or ML (henceforth referred to as “AI”) using a 5-point response scale of strongly disagree, somewhat disagree, neither agree nor disagree, somewhat agree, and strongly agree. The domains covered in this section were prior understanding of AI, patient safety and privacy, previous experience with AI in medical training, impact on the primary care workforce, and the predicted impact of AI on health care utilization related to cancer screening (in terms of costs and unnecessary diagnostic procedures). One question asked respondents to select all entities that should bear responsibility for regulating AI, including the federal government (such as the Food and Drug Administration), state governments, physician organizations, software developers, and individual providers. We surveyed respondents about their medical training and AI regulation since these perceptions may influence preferences toward AI in cancer screening.
Clinical Vignette Questions
Section 3 was composed of 11 clinical vignettes that asked participants to assume the role of a PCP and report various preferences regarding AI in the context of cancer screening decisions (see Appendix).
Vignettes 1 to 3 asked the respondent to select the preferred prediction metric or functionality for a hypothetical AI model to communicate a cancer screening recommendation. Options provided were a 5-y probability of a cancer diagnosis, recommended time until the next screening test should be ordered, whether to screen now or defer to the next USPSTF-recommended interval, or not using any of these tools. Participants were given the same scenario in each vignette but the type of cancer screening test in question (colorectal, breast, or lung cancer) varied. An example is seen below:
You are a primary care provider who has just learned that your health system is deciding between purchasing a license to one of four FDA-approved AI-based tools for colorectal cancer screening management. This tool only applies to patients who have received at least one colonoscopy in the past. Assume that all tools have comparable accuracy.
Vignettes 4 to 6 asked the respondents to choose how best to implement an AI tool in the clinical workflow for cancer screening. We sought to examine the clinical workflow because previous work has suggested that automated alert systems can improve patient follow-up and closed-loop communication in radiology contexts,36,37,43–45 although it is unclear if the electronic health record (EHR) is the preferred avenue for communicating AI recommendations in the primary care setting. The following scenarios were identical for each vignette but varied in cancer type (colorectal, breast, or lung). Response options included 1) generating a flag in the EHR for eligible patients, 2) having an accessible risk calculator online or in a smartphone application, 3) providing predictions to a practice’s licensed care manager once per year, or 4) none of the above. An example scenario is seen below:
You are a primary care provider who has just learned that your health system is purchasing a license for a new tool that estimates the five-year risk of colorectal cancer ([0-100%] with 95% CI) for each patient of screening age, regardless of previous screenings. The health system is deciding how best to implement the tool in clinical workflow. Please select how you would prefer to implement the tool’s predictions.
Vignettes 7 and 8 asked respondents which members of the healthcare team should be alerted if a hypothetical AI model detects suspicious findings on routine lung cancer or breast cancer screening imaging. Options included 1) alerting the attending radiologist, 2) alerting the patient’s PCP, 3) alerting a licensed care manager, or 4) none of the above. Participants could select multiple options if desired. The vignettes varied only in the cancer type. The order in which cancer types were presented was the same for all respondents (lung then breast). An example scenario is seen below:
Your hospital system has recently purchased a license to an AI-based tool that segments (outlines) regions that are suspicious for breast masses. The tool offers three options for how suspicious masses can be communicated to the healthcare team. Assume the patient has never been diagnosed with a cancer. Please select which members of the care team should be alerted to suspicious masses (select all that apply).
Vignettes 9 and 10 asked respondents when, in the clinical workflow, an attending radiologist should be able to view an AI-generated radiology report for screening mammograms or low-dose chest computed tomography (CT). Options included 1) the radiologist viewing the AI-generated report after viewing the imaging herself but before signing her final report, 2) the radiologist signing her final report before viewing the AI-generated report, 3) the radiologist viewing an AI-generated report before viewing the imaging, or 4) none of the above. The vignettes varied only in the cancer type; the order in which cancer types were presented was the same for all respondents (lung then breast). An example scenario is seen below:
Your hospital system has recently purchased a license to an AI-based tool that segments (outlines) regions of mammograms that are suspicious for pathologic breast masses. The tool offers three options for how suspicious breast masses can be communicated to the healthcare team. Assume the patient has never been diagnosed with a cancer. Please select which members of the care team should be alerted to suspicious breast masses (select all that apply).
Vignette 11 dealt with different methods for explainability pertaining to a breast cancer risk prediction tool that provides a binary recommendation (follow-up needed or no follow-up needed). Options included 1) model reporting the percentage probability of developing breast cancer within the next 5 years with delineation of the most relevant risk factors, 2) model estimating the volumetric breast density from the mammogram and assigning a Breast Imaging-Reporting and Data System (BI-RADS) density category 46 , 3) a model segmenting suspicious masses on previous mammograms, 4) no preference for any of these explainability methods, or 5) no preference for any of these methods. See the appendix for the graphics accompanying each response option; graphics for options 2 and 3 for this vignette were reproduced from He et al. 47 and Hassan et al., 48 respectively. The following background was given:
The “explainability” of AI models is an important factor in determining provider trust of these models’ recommendations. You are a primary care provider who is piloting a new tool to aid in the management of breast cancer screening for patients who recently received a standard mammogram. The tool provides a binary recommendation (follow-up needed or no follow-up needed). The developers of this tool want to know the best way in which the tool can explain its decision-making. They provide you three options for how to accomplish this. Visual accompaniments are below.
Cancer types were presented in the same order (colorectal, breast, lung) across all vignettes, unless otherwise specified.
Statistical Analysis
Responses were included in the analysis if at least 50% of the questions were answered. Age was calculated using the provided date of birth. Time in independent practice was categorized as ≤5 y, 6 to 10 y, 11 to 20 y, 21 to 30 y, or >30 y. Healthcare role was categorized as attending physician, advanced practice provider, or resident. We conducted a descriptive analysis of attitude items and clinical vignettes. Attitude questions were assessed on a 5-point Likert-type scale (strongly disagree = 1, somewhat disagree = 2, neither agree nor disagree = 3, somewhat agree = 4, strongly agree = 5). For each attitudes-related question from section 2, we reported the number of respondents who selected each Likert scale–based option and then calculated an average Likert scale value with a 95% confidence interval. For questions relating to AI regulation (section 2) and workflow (section 3), respondents could select more than 1 option, and we report the total affirmative responses for each option.
Results
Response Rate and Sample
We distributed the survey to 733 providers and received 99 responses from 9 different clinics that met criteria for analysis, yielding an overall response rate of 14%. The demographics of the respondents can be seen in Table 1. The median age of respondents who provided their age was 44 y (interquartile range 36–55 y). Seventy-two (73%) were attending physicians, 23 (23%) were APPs, and 2 (2%) were family medicine residents. Seventy-one (72%) were affiliated with the HUP and 18 (18%) with LGH. Nine (9%) were not affiliated with HUP or LGH, including 2 (2%) in California and 7 (7%) in Illinois. Seventy-four (75%) had been practicing independently in the primary care setting for more than 5 y. Fourteen (14%) reported using an AI-based tool in clinical practice in the past.
Table 1.
Demographics of the Respondents
| Characteristic | N = 99 |
|---|---|
| Median age (interquartile range), y | 44 (36–55) |
| Health care role | |
| Attending physician | 72 (73%) |
| Advanced practice provider | 23 (23%) |
| Family medicine resident | 2 (2%) |
| No response | 2 (2%) |
| Institutional affiliation | |
| Hospital of the University of Pennsylvania | 71 (72%) |
| Lancaster General Hospital | 18 (18%) |
| Other US-based community medical center | 9 (9%) |
| Unsure | 1 (1%) |
| Time spent in practice independently, y | |
| ≤5 | 21 (21%) |
| 6–10 | 20 (20%) |
| 11–20 | 17 (17%) |
| 21–30 | 24 (24%) |
| >30 | 13 (13%) |
| No response | 4 (4%) |
| Previous experience with AI in the clinical setting | |
| Yes | 14 (14%) |
| No | 56 (57%) |
| Unsure/no response | 29 (29%) |
| Completed fellowship training | |
| Yes | 11 (11%) |
| No | 61(62%) |
| No response | 27 (27%) |
| Spent at least 50% of clinical time in primary care during the last year | |
| Yes | 93 (94%) |
| No | 4 (4%) |
| Unsure | 1 (1%) |
| No response | 1 (1%) |
Results of the Survey of Attitudes
Of the respondents, 73 (74%) were somewhat or strongly confident that they could explain the term artificial intelligence to a layperson. However, 90 (91%) and 89 (90%) of respondents felt they did not receive adequate training regarding AI in their undergraduate and graduate medical training, respectively. According to the respondents (Figure 2), entities that should have responsibility for regulating healthcare AI in the United States were the federal government (76%), individual hospital systems (49%), and physician organizations (48%). Only 25% felt that individual providers should be responsible for the AI models they used. A minority of respondents expressed concern that AI would endanger patient safety (25%) or affect the privacy of protected health information (35%).
Figure 2.

Participant opinion regarding who bears responsibility for regulating artificial intelligence (AI) in the health care field. Participants could select multiple options if desired.
Most respondents somewhat agreed or strongly agreed that AI tools would improve personalized screening decisions for patients at risk for colorectal cancer (65%), breast cancer (67%), and lung cancer (70%). A plurality of respondents expected AI tools to somewhat or significantly decrease the number of unnecessary colonoscopies (39%), breast biopsies (39%), and lung nodule biopsies (39%). Only 14% reported concern that the advent of AI in healthcare would negatively affect job prospects in primary care in the United States (Table 2).
Table 2.
Experiences of Respondents Regarding AI in Their Medical Training as Well as Perceptions of Its Benefits and Risks in Health Care a
| Statement | Strongly Disagree | Somewhat Disagree | Neither Agree nor Disagree | Somewhat Agree | Strongly Agree | No Response | Mean Response (95% CI) |
|---|---|---|---|---|---|---|---|
| “I could explain to a friend or family member what ‘artificial intelligence’ means.” | 2 (2%) | 8 (8%) | 14 (14%) | 50 (51%) | 23 (23%) | 2 (2%) | 3.87 (3.67–4.06) |
| “My undergraduate medical training provided adequate training on AI.” | 78 (79%) | 12 (12%) | 6 (6%) | 0 (0%) | 1 (1%) | 2 (2%) | 1.29 (1.15–1.43) |
| “My graduate medical training provided adequate training on AI.” | 76 (77%) | 13 (13%) | 5 (5%) | 2 (2%) | 0 (0%) | 3 (3%) | 1.30 (1.17–1.44) |
| “I am concerned that AI will endanger patient safety.” | 8 (8%) | 28 (28%) | 36 (36%) | 20 (20%) | 5 (5%) | 2 (2%) | 2.86 (2.65–3.06) |
| “I am concerned that AI tools will negatively impact the confidentiality of protected health information (PHI).” | 7 (7%) | 28 (28%) | 27 (27%) | 28 (28%) | 7 (7%) | 2 (2%) | 3.00 (2.78–3.22) |
| “AI tools will improve personalized screening decisions for patients at risk for colorectal cancer.” | 0 (0%) | 5 (5%) | 27 (27%) | 46 (46%) | 19 (19%) | 2 (2%) | 3.81 (3.65–3.98) |
| “AI tools will improve personalized screening decisions for patients at risk for breast cancer.” | 0 (0%) | 3 (3%) | 27 (27%) | 48 (48%) | 19 (19%) | 2 (2%) | 3.86 (3.70–4.01) |
| “AI tools will improve personalized screening decisions for patients at risk for lung cancer.” | 0 (0%) | 3 (3%) | 25 (25%) | 50 (51%) | 19 (19%) | 2 (2%) | 3.88 (3.72–4.03) |
| “I am concerned that AI tools will decrease the number of primary care attending physician, NP, or PA job positions in the United States.” | 37 (37%) | 30 (30%) | 16 (16%) | 12 (12%) | 2 (2%) | 2 (2%) | 2.09 (1.87–2.32) |
AI, artificial intelligence; NP, nurse practitioner; PA, physician assistant.
Mean response = average of numerically coded Likert values excluding “no response.”
Clinical Vignettes
With regard to the preferred recommendation format for AI algorithms in cancer screening, a plurality of respondents opted for a recommended time interval until next screening for all 3 cancers (Tables 3 and 4). Fifty-two percent chose this option for the next colonoscopy, compared with 39% for the next mammogram and 37% for low-dose chest CT.
Table 3.
Perceptions Regarding the Future Impact of AI on the Costs of Cancer Screening and the Number of Unnecessary Diagnostic Procedures for Colorectal, Breast, and Lung Cancers. a
| Statement | Significantly Decrease | Somewhat Decrease | Have No Significant Effect On | Somewhat Increase | Significantly Increase | No Response |
|---|---|---|---|---|---|---|
| Overall costs of screening | ||||||
| Colorectal cancer | 6 (6%) | 33 (33%) | 36 (36%) | 22 (22%) | 1 (1%) | 1 (1%) |
| Breast cancer | 6 (6%) | 33 (33%) | 30 (30%) | 27 (27%) | 1 (1%) | 2 (2%) |
| Lung cancer | 7 (7%) | 32 (32%) | 24 (24%) | 32 (32%) | 2 (2%) | 2 (2%) |
| Number of unnecessary diagnostic procedures | ||||||
| Colonoscopies | 3 (3%) | 49 (49%) | 30 (30%) | 13 (13%) | 1 (1%) | 3 (3% |
| Breast biopsies | 7 (7%) | 44 (44%) | 25 (25%) | 19 (19%) | 1 (1%) | 3 (3%) |
| Lung nodule biopsies | 8 (8%) | 47 (47%) | 24 (24%) | 16 (16% | 1 (1%) | 3 (3%) |
For each cancer, participants were asked to complete the following sentence: “AI- and ML-tools will ____ the overall costs of screening in the United States” and “AI- and ML-tools will ____ the number of unnecessary procedures in the United States.”
Table 4.
Preferred Recommendation Format and Workflow Implementations for Hypothetical AI Algorithms Based on the Cancer. a
| Survey Question of Interest | Colorectal Cancer | Breast Cancer | Lung Cancer |
|---|---|---|---|
| Format of recommendation | |||
| 5-y probability of cancer diagnosis (0%–100% with 95% CI) | 16 (16%) | 16 (16%) | 13 (13%) |
| Recommended time until next cancer screening (years, with 95% CI) | 51 (52%) | 39 (39%) | 37 (37%) |
| Binary recommendation of 1) screen now or 2) defer until the next USPTSF-recommended screening interval | 14 (14%) | 15 (15%) | 19 (19%) |
| Segment suspicious findings on imaging | N/A | 10 (10%) | 13 (13%) |
| Would not use any of these tools | 8 (8%) | 9 (9%) | 7 (7%) |
| No response | 10 (10%) | 10 10%) | 10 (10%) |
| Workflow implementation | |||
| Automatically generate EHR chart flag for eligible patients | 56 (57%) | 55 (56%) | 56 (57%) |
| Have an easily accessible risk calculator hosted online or through a smartphone application | 13 (13%) | 12 (12%) | 11 (11%) |
| Provide predictions to the practice’s licensed care manager once per year | 8 (8%) | 10 (10%) | 11 (11%) |
| None of the above | 6 (6%) | 6 (6%) | 5 (5%) |
| No response | 16 (16%) | 16 (16%) | 16 (16%) |
| Communication of suspicious imaging findings to the healthcare team | |||
| Alert attending radiologist | N/A | 70 (71%) | 68 (69%) |
| Alert PCP | N/A | 53 (54%) | 53 (54%) |
| Alert licensed care manager | N/A | 27 (27%) | 25 (25%) |
| None of the above | N/A | 1 (1%) | 1 (1%) |
| No response | N/A | 15 (15%) | 15 (15%) |
| Radiologist workflow implementation | |||
| Attending radiologist may view AI-generated report for a given study before signing her final report. Both reports are available in the EMR. | N/A | 44 (44%) | 44 (44%) |
| Attending radiologist signs her final radiology report before the AI-generated report can be viewed. Both reports are available in the EMR. | N/A | 10 (10%) | 10 (10%) |
| Attending radiologist views the AI-generated report for a given study before viewing the study images herself. The attending then views images and signs the final report. Both reports are available in the EMR. | N/A | 21 (21%) | 21 (21%) |
| None of the above | N/A | 7 (7%) | 8 (8%) |
| No response | N/A | 17 (17%) | 16 (16%) |
AI, artificial intelligence; CI, confidence interval; EMR, electronic medical record; N/A, not applicable; PCP, primary care provider; USPSTF, United States Preventative Services Task Force.
Each vignette was single choice and single attribute. The preferred method of communication of imaging-related AI findings to the health care team and implementation in radiologist workflow. Low-dose chest computed tomography and mammogram for lung and breast cancers, respectively.
Concerning workflow implementation, most (56% or 57%) of the respondents preferred that a flag be generated in the electronic health record describing AI-generated cancer risk, greater than those who preferred involving a practice’s licensed care manager (8%–11%) or using a smartphone application (11%–13%). Respondents indicated a preference for hypothetical AI systems that detect suspicious nodules on mammograms or alert the attending radiologist (71%) and the patient’s PCP (54%) over a licensed care manager (27%). Similarly, for nodule detection on low-dose chest CT, 69% and 54% preferred that the AI alert the radiologist and PCP, respectively, compared with 25% for a licensed care manager.
For breast and lung cancer screening, a plurality (44%) preferred that AI interpretations of breast and lung imaging be made available after the radiologist makes her preliminary report but before providing a final attestation. This finding is in contrast to the radiologist making a final interpretation without using an AI at all (selected by 10%) or the radiologist accessing the AI interpretation before viewing the imaging herself (selected by 21%).
Finally, in terms of explainability for a breast cancer risk prediction algorithm (Table 5), while a plurality (33%) opted for a delineation of pertinent risk factors influencing the model’s prediction, 18% preferred more models that highlight suspicious findings on breast imaging, 8% preferred breast density estimations, and 23% had no preference regarding explainability.
Table 5.
Preferences for Explainability for a Hypothetical AI Algorithm That Informs Breast Cancer Screening Risk. a
| Explainability Method | |
|---|---|
| Model reports percentage probability of developing breast cancer within the next 5 y with delineation of the most relevant risk factors that influenced prediction | 33 (33%) |
| Model estimates volumetric breast density from a mammogram and assigns BI-RADS density category | 8 (8%) |
| Model segments suspicious masses on previous mammograms | 18 (18%) |
| I have no preference for any of these explainability methods | 23 (23%) |
| No response | 17 (17%) |
BI-RADS, Breast Imaging-Reporting and Data System.
This vignette was single choice and single attribute.
Discussion
This study fills an important knowledge gap by elucidating the user interface and user experience features that will best translate the power of AI models to cancer screening. Previous work has leveraged qualitative and quantitative methods to understand patient and provider preferences regarding AI in health care. Hendrix et al. 49 used discrete choice experiments and qualitative interviews to understand what features would increase clinician trust in an AI method for breast tumor detection on screening mammograms. The authors found that for clinicians to trust the predictions from such an algorithm, sensitivity should be prioritized over specificity, radiologist confirmation of the AI prediction should be present, and the model should be trained on a diverse patient population. 49 While this information is helpful, questions remain regarding the deployment of AI in the clinical context.
In our study, most clinicians reported optimism regarding the advent of AI in cancer screening, and few felt concerned about the impact of AI on their future job prospects, which aligns with previous work. 50 Yet, clinicians do not feel that their medical training has adequately prepared them to engage with AI, indicating a need for medical education to adapt. Given that no current guidelines or regulations exist regarding how medical education should teach trainees to interact with AI tools and a literature describing a desire for a more rigorous AI curriculum,50–54 leaders in medical education should consider incorporating formal AI instruction into their programs.
Our findings support a preliminary framework to guide model deployment for software developers and health systems. Many previously published and/or widely publicized AI models rely on either highlighting suspicious findings in imaging,55–57 a process that is interpretable by humans, or risk stratification (i.e., low risk versus high risk for future cancer status).58–60 However, in the context of primary care–based cancer screening, respondents indicated a preference for prescriptive recommendations, such as an interval until the next screening, in contrast to more objective but less actionable assessments of risk. Developers may need to alter AI models so that they provide more prescriptive outputs, since those are more clinically actionable.
We now turn our focus to workflow implementation. Our findings suggest that health systems should focus on integrating AI models into the EHR and ensuring that clinician assessments are not initially biased by an AI-generated recommendation. These results are in concert with previous work concerning a hypothetical melanoma detection tool 21 and AI-enabled medical devices in diabetes care, 61 underscoring a desire for mitigating provider bias, robust user support services, and a limiting of device alerts to those that are pertinent and actionable. In addition, delegating the responsibility for following up on an AI-generated prediction from a PCP to another healthcare team member was not popular among providers. Ensuring that AI is conducive to the clinician’s busy schedule is paramount to maximizing its potential.
Interestingly, there was no clear consensus on a preferred method for explainability in a hypothetical AI model for breast cancer screening, with one-quarter explicitly having no preference for any of the methods we offered. While this contrasts with a substantial literature describing explainability’s importance,17,19,22 respondents may not have had sufficient AI literacy to understand the implications of the options. Some scholars have argued that the black box nature of AI should not necessitate its removal from healthcare spaces because some widely accepted clinical decisions rely on “opaque” reasoning.62–64 Future work should further investigate how best to communicate explainability and the need, if any, of explainability in the cancer screening context.
Our results apply to other emerging areas. While three-quarters of respondents in our study believe the onus of regulation lies with the federal government, other bodies are seen as responsible, including individual hospital systems, physician organizations, and software developers. This finding aligns with a literature indicating a lack of clarity regarding who is morally and/or legally accountable for the outcomes associated with this technology.50,65–67 One way in which individual healthcare systems can exercise oversight over this technology is through governance committees, an increasingly common review board at many institutions that ensures that AI tools promote usability, equity, and explainability in patient care.68–70 A UI/UX framework for such tools can guide the vetting and in-house regulation of AI models from potential vendors.
As payers begin to reimburse providers for care involving AI decision support, a new framework for future AI tools may affect whether payers choose to reimburse care that leverages AI, as well as whether AI is needed to meet certain quality metrics in value-based care models.71,72 Further, standardizing outputs for AI models may facilitate future cost-effectiveness studies.50,73,74
From a practice standpoint, a clear framework for AI models and improved AI training in medical education should allow the PCP to better engage the patient in shared decision-making regarding their cancer screening. If the clinician can better communicate her reasoning to the patient, this will lead to decisions that better align with a patient’s values. 27 Previous work demonstrates reporting superior collaborative decision-making for patients with osteoarthritis who received an AI-enabled decision aid and patient education, as compared with those who received only patient education. 75
This study has some limitations. First, our aims are primarily descriptive, and the study was not designed to infer relationships between AI preferences and respondent attributes including age, healthcare role, or time in practice or across cancer types. However, the directionality of the responses was consistent between each of the cancer types. Second, most respondents were affiliated with clinics in an academic medical center in the Northeast, which may not capture the preferences of providers in more community and/or rural clinical settings. Clinicians at an academic medical center may have more exposure to AI tools than those in a community setting, 69 which may lead to them having a more favorable view of AI. Clinicians in academic settings who have a range of roles including clinical work, education, and research may have a lower concern of AI tools encroaching on their job prospects compared with those who are primarily focused on clinical work. Third, our examination of preferences regarding AI in radiology workflow relied on answers from PCPs rather than radiologists who play a key role in the breast and lung cancer screening process. Lastly, we restricted this study to the management of colorectal, breast, and lung cancers. Screening and detection of other conditions in the primary care setting could also benefit from a more standardized framework for AI models. Finally, we do not have access to the demographics of those who did not respond, so we cannot speculate about possible self-selection bias. Future work is needed to extend this study to a broader range of clinicians and clinical scenarios as the vignette options were not mutually exclusive nor exhaustive of all knowledge gaps present in the AI space.
In conclusion, through a survey of primary care clinicians, we identified key attitudes and preferences regarding AI in clinical decision-making for cancer screening. These findings constitute a preliminary framework for optimizing the deployment of AI tools in the primary care setting as well as in other contexts.
Supplemental Material
Supplemental material, sj-docx-1-mpp-10.1177_23814683251329007 for Primary Care Provider Preferences Regarding Artificial Intelligence in Point-of-Care Cancer Screening by Vinayak S. Ahluwalia, Marilyn M. Schapira, Gary E. Weissman and Ravi B. Parikh in MDM Policy & Practice
Footnotes
The authors declared the following potential conflicts of interest with respect to the research, authorship, and/or publication of this article: There are potential conflicts of interest. RBP has received research funding from the National Institutes of Health, Department of Defense, Prostate Cancer Foundation, National Palliative Care Research Center, NCCN Foundation, Conquer Cancer Foundation, Humana, Emerson Collective, Schmidt Futures, Arnold Ventures, Mendel.ai, and Veterans Health Administration; personal fees and equity from GNS Healthcare, Thyme Care, and Onc.AI; personal fees from the ConcertAI, Cancer Study Group, Biofourmis, Archetype Therapeutics, CreditSuisse, G1 Therapeutics, Humana, and Nanology; honoraria from Flatiron and Medscape; has board membership (unpaid) at the Coalition to Transform Advanced Care and American Cancer Society; and serves on a leadership consortium (unpaid) at the National Quality Forum. The authors disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: VSA and MMS report no funding. GEW received support from NIH R35GM155262 and R03HL171424 during the study period. RBP received funding from NIH 5K08CA263541 during the study period. RBP has received research funding from the National Institutes of Health, Department of Defense, Prostate Cancer Foundation, National Palliative Care Research Center, NCCN Foundation, Conquer Cancer Foundation, Humana, Emerson Collective, Schmidt Futures, Arnold Ventures, Mendel.ai, and Veterans Health Administration; personal fees and equity from GNS Healthcare, Thyme Care, and Onc.AI; personal fees from the ConcertAI, Cancer Study Group, Biofourmis, Archetype Therapeutics, CreditSuisse, G1 Therapeutics, Humana, and Nanology; honoraria from Flatiron and Medscape; has board membership (unpaid) at the Coalition to Transform Advanced Care and American Cancer Society; and serves on a leadership consortium (unpaid) at the National Quality Forum. The funding agreement ensured the authors’ independence in designing the study, interpreting the data, writing, and publishing the report.
Ethical Considerations: This study was exempt from institutional review board approval at the University of Pennsylvania (protocol No. 855353).
Consent to Participate: Written.
Patient Consent: Not applicable.
Consent for Publication: Not applicable.
Disclaimer: The views and opinions expressed in this article are those of the authors and do not reflect the official positions or policies of the institutions with which the authors are affiliated.
ORCID iDs: Vinayak S. Ahluwalia
https://orcid.org/0000-0002-2227-6644
Marilyn M. Schapira
https://orcid.org/0000-0002-3775-9462
Gary E. Weissman
https://orcid.org/0000-0001-9588-3819
Ravi B. Parikh
https://orcid.org/0000-0003-2692-6306
Data Availability: Anonymized data can be made available upon request.
Contributor Information
Vinayak S. Ahluwalia, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, USA Leonard Davis Institute of Health Economics, University of Pennsylvania, Philadelphia, PA, USA.
Marilyn M. Schapira, Leonard Davis Institute of Health Economics, University of Pennsylvania, Philadelphia, PA, USA Department of Medicine, University of Pennsylvania, Philadelphia, PA, USA.
Gary E. Weissman, Leonard Davis Institute of Health Economics, University of Pennsylvania, Philadelphia, PA, USA Department of Medicine, University of Pennsylvania, Philadelphia, PA, USA; Palliative and Advanced Illness Research (PAIR) Center, University of Pennsylvania Perelman School of Medicine, Philadelphia, PA, USA.
Ravi B. Parikh, Emory University School of Medicine, Atlanta, GA, USA; Emory Winship Cancer Institute, Atlanta, GA, USA.
References
- 1. Herreros-Villanueva M, Duran-Sanchon S, Martín AC, et al. Plasma MicroRNA signature validation for early detection of colorectal cancer. Clin Transl Gastroenterol. 2019;10(1):e00003. doi: 10.14309/ctg.0000000000000003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2. Afshar S, Afshar S, Warden E, Manochehri H, Saidijam M. Application of artificial neural network in miRNA biomarker selection and precise diagnosis of colorectal cancer. Iran Biomed J. 2019;23(3):175–83. doi: 10.29252/.23.3.175 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Soares F, Becker K, Anzanello MJ. A hierarchical classifier based on human blood plasma fluorescence for non-invasive colorectal cancer screening. Artif Intell Med. 2017;82:1–10. doi: 10.1016/j.artmed.2017.09.004 [DOI] [PubMed] [Google Scholar]
- 4. Wang P, Berzin TM, Brown JRG, et al. Real-time automatic detection system increases colonoscopic polyp and adenoma detection rates: a prospective randomised controlled study. Gut. 2019;68(10):1813–9. doi: 10.1136/gutjnl-2018-317500 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5. McKinney SM, Sieniek M, Godbole V, et al. International evaluation of an AI system for breast cancer screening. Nature. 2020;577(7788):89-94. doi: 10.1038/s41586-019-1799-6 [DOI] [PubMed] [Google Scholar]
- 6. Ahluwalia VS, Doiphode N, Mankowski WC, et al. Volumetric breast density estimation from three-dimensional reconstructed digital breast tomosynthesis images using deep learning. JCO Clin Cancer Inform. 2024;(8):e2400103. doi: 10.1200/CCI.24.00103 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Swiderski B, Kurek J, Osowski S, Kruk M, Barhoumi W. Deep learning and non-negative matrix factorization in recognition of mammograms. In: Eighth International Conference on Graphic and Image Processing (ICGIP 2016). Vol 10225. Tokyo, Japan: SPIE; 2017. pp 53–59. doi: 10.1117/12.2266335 [DOI] [Google Scholar]
- 8. Antropova N, Huynh BQ, Giger ML. A deep feature fusion methodology for breast cancer diagnosis demonstrated on three imaging modality datasets. Med Phys. 2017;44(10):5162–71. doi: 10.1002/mp.12453 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Gould MK, Huang BZ, Tammemagi MC, Kinar Y, Shiff R. Machine learning for early lung cancer identification using routine clinical and laboratory data. Am J Respir Crit Care Med. 2021;204(4):445–53. doi: 10.1164/rccm.202007-2791OC [DOI] [PubMed] [Google Scholar]
- 10. Kang G, Liu K, Hou B, Zhang N. 3D multi-view convolutional neural networks for lung nodule classification. PLoS One. 2017;12(11):e0188290. doi: 10.1371/journal.pone.0188290 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Ardila D, Kiraly AP, Bharadwaj S, et al. End-to-end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography. Nat Med. 2019;25(6):954–61. doi: 10.1038/s41591-019-0447-x [DOI] [PubMed] [Google Scholar]
- 12. Ciompi F, Chung K, van Riel SJ, et al. Towards automatic pulmonary nodule management in lung cancer screening with deep learning. Sci Rep. 2017;7(1):46479. doi: 10.1038/srep46479 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Gunasekeran DV, Zheng F, Lim GYS, et al. Acceptance and Perception of Artificial Intelligence Usability in Eye Care (APPRAISE) for ophthalmologists: a multinational perspective. Front Med. 2022;9:875242. doi: 10.3389/fmed.2022.875242 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Leenhardt R, Fernandez-Urien Sainz I, Rondonotti E, et al. PEACE: perception and expectations toward artificial intelligence in capsule endoscopy. J Clin Med. 2021;10(23):5708. doi: 10.3390/jcm10235708 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Scheetz J, Rothschild P, McGuinness M, et al. A survey of clinicians on the use of artificial intelligence in ophthalmology, dermatology, radiology and radiation oncology. Sci Rep. 2021;11(1):5193. doi: 10.1038/s41598-021-84698-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Doraiswamy PM, Blease C, Bodner K. Artificial intelligence and the future of psychiatry: insights from a global physician survey. Artif Intell Med. 2020;102:101753. doi: 10.1016/j.artmed.2019.101753 [DOI] [PubMed] [Google Scholar]
- 17. Yoon CH, Torrance R, Scheinerman N. Machine learning in medicine: should the pursuit of enhanced interpretability be abandoned? J Med Ethics. 2022;48(9):581–5. doi: 10.1136/medethics-2020-107102 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Al-Edresee T. Physician acceptance of machine learning for diagnostic purposes: caution, bumpy road ahead! Stud Health Technol Inform. 2022;295:83–86. doi: 10.3233/SHTI220666 [DOI] [PubMed] [Google Scholar]
- 19. Shortliffe EH, Sepúlveda MJ. Clinical decision support in the era of artificial intelligence. JAMA. 2018;320(21):2199–200. doi: 10.1001/jama.2018.17163 [DOI] [PubMed] [Google Scholar]
- 20. Ghosh A, Kandasamy D. Interpretable artificial intelligence: why and when. Am J Roentgenol. 2020;214(5):1137–8. doi: 10.2214/AJR.19.22145 [DOI] [PubMed] [Google Scholar]
- 21. Haggenmüller S, Maron RC, Hekler A, et al. Patients’ and dermatologists’ preferences in artificial intelligence-driven skin cancer diagnostics: a prospective multicentric survey study. J Am Acad Dermatol. 2024;91(2):366–70. doi: 10.1016/j.jaad.2024.04.033 [DOI] [PubMed] [Google Scholar]
- 22. Amann J, Blasimme A, Vayena E, Frey D, Madai VI; the Precise4Q consortium. Explainability for artificial intelligence in healthcare: a multidisciplinary perspective. BMC Med Inform Decis Mak. 2020;20(1):310. doi: 10.1186/s12911-020-01332-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Qiao EM, Guram K, Kotha NV, et al. Association between primary care use prior to cancer diagnosis and subsequent cancer mortality in the veterans affairs health system. JAMA Netw Open. 2022;5(11):e2242048. doi: 10.1001/jamanetworkopen.2022.42048 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Owens-Jasey C, Chen J, Xu R, et al. Implementation of health IT for cancer screening in US primary care: scoping review. JMIR Cancer. 2024;10(1):e49002. doi: 10.2196/49002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Mańczuk M, Przepiórka I, Cedzyńska M, et al. Actual and potential role of primary care physicians in cancer prevention. Cancers. 2023;15(2):427. doi: 10.3390/cancers15020427 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Trimble M, Hamilton P. The thinking doctor: clinical decision making in contemporary medicine. Clin Med. 2016;16(4):343. doi: 10.7861/clinmedicine.16-4-343 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Schrager S, Phillips G, Burnside E. Shared decision making in cancer screening. Fam Pract Manag. 2017;24(3):5. [PMC free article] [PubMed] [Google Scholar]
- 28. Kiviniemi MT, Hay JL, James AS, et al. Decision making about cancer screening: an assessment of the state of the science and a suggested research agenda. Cancer Epidemiol Biomarkers Prev. 2009;18(11):3133. doi: 10.1158/1055-9965.EPI-18-11-ASPO [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Tiffen J, Corbridge SJ, Slimmer L. Enhancing clinical decision making: development of a contiguous definition and conceptual framework. J Prof Nurs. 2014;30(5):399–405. doi: 10.1016/j.profnurs.2014.01.006 [DOI] [PubMed] [Google Scholar]
- 30. United States Preventive Services Taskforce. A and B recommendations. 2024. Available from: https://www.uspreventiveservicestaskforce.org/uspstf/recommendation-topics/uspstf-a-and-b-recommendations. [Accessed 18 November, 2024].
- 31. Peters Y, van Grinsven E, van de Haterd M, van Lankveld D, Verbakel J, Siersema PD. Individuals’ preferences for esophageal cancer screening: a discrete choice experiment. Value Health. 2020;23(8):1087–95. doi: 10.1016/j.jval.2020.03.013 [DOI] [PubMed] [Google Scholar]
- 32. Gyrd-Hansen D, Søgaard J. Analysing public preferences for cancer screening programmes. Health Econ. 2001;10(7):617–34. doi: 10.1002/hec.622 [DOI] [PubMed] [Google Scholar]
- 33. Pignone M, Bucholtz D, Harris R. Patient preferences for colon cancer screening. J Gen Intern Med. 1999;14(7):432–7. doi: 10.1046/j.1525-1497.1999.00018.x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Wong SF, Norman R, Dunning TL, Ashley DM, Lorgelly PK. A protocol for a discrete choice experiment: understanding preferences of patients with cancer towards their cancer care across metropolitan and rural regions in Australia. BMJ Open. 2014;4(10):e006661. doi: 10.1136/bmjopen-2014-006661 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Asan O, Bayrak AE, Choudhury A. Artificial intelligence and human trust in healthcare: focus on clinicians. J Med Internet Res. 2020;22(6):e15154. doi: 10.2196/15154 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36. Kapoor N, Lacson R, Khorasani R. Workflow applications of artificial intelligence in radiology and an overview of available tools. J Am Coll Radiol. 2020;17(11):1363–70. doi: 10.1016/j.jacr.2020.08.016 [DOI] [PubMed] [Google Scholar]
- 37. O’Connor SD, Dalal AK, Sahni VA, Lacson R, Khorasani R. Does integrating nonurgent, clinically significant radiology alerts within the electronic health record impact closed-loop communication and follow-up? J Am Med Inform Assoc JAMIA. 2016;23(2):333–8. doi: 10.1093/jamia/ocv105 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38. Tomoaia-Cotisel A, Farrell TW, Solberg LI, et al. Implementation of care management: an analysis of recent AHRQ research. Med Care Res Rev MCRR. 2018;75(1):46–65. doi: 10.1177/1077558716673459 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39. Leach B, Morgan P, Strand de Oliveira J, Hull S, Østbye T, Everett C. Primary care multidisciplinary teams in practice: a qualitative study. BMC Fam Pract. 2017;18:115. doi: 10.1186/s12875-017-0701-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Case MAB. Oncology nurse navigator. Clin J Oncol Nurs. 2011;15(1):33–40. doi: 10.1188/11.CJON.33-40 [DOI] [PubMed] [Google Scholar]
- 41. Emfield Rowett K, Christensen D. Oncology nurse navigation: expansion of the navigator role through telehealth. Clin J Oncol Nurs. 2020;24(3):24–31. doi: 10.1188/20.CJON.S1.24-31 [DOI] [PubMed] [Google Scholar]
- 42. Role of the oncology nurse navigator throughout the cancer trajectory. Oncol Nurs Forum. 2018;45(3):283. doi: 10.1188/18.ONF.283 [DOI] [PubMed] [Google Scholar]
- 43. Lacson R, O’Connor SD, Sahni VA, et al. Impact of an electronic alert notification system embedded in radiologists’ workflow on closed-loop communication of critical results: a time series analysis. BMJ Qual Saf. 2016;25(7):518–24. doi: 10.1136/bmjqs-2015-004276 [DOI] [PubMed] [Google Scholar]
- 44. Carrodeguas E, Lacson R, Swanson W, Khorasani R. Use of machine learning to identify follow-up recommendations in radiology reports. J Am Coll Radiol JACR. 2019;16(3):336–43. doi: 10.1016/j.jacr.2018.10.020 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45. Blezek DJ, Olson-Williams L, Missert A, Korfiatis P. AI integration in the clinical workflow. J Digit Imaging. 2021;34(6):1435–46. doi: 10.1007/s10278-021-00525-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46. Melnikow J, Fenton JJ, Whitlock EP, et al. Table 1, BI-RADS breast density descriptions. January 2016. Available from: https://www.ncbi.nlm.nih.gov/books/NBK343794/table/ch1.t1/. [Accessed 24 April, 2024].
- 47. He W, Juette A, Denton ERE, Oliver A, Martí R, Zwiggelaar R. A review on automatic mammographic density and parenchymal segmentation. Int J Breast Cancer. 2015;2015. doi: 10.1155/2015/276217 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48. Hassan SA, Sayed MS, Farag F. Segmentation of breast cancer lesion in digitized mammogram images. In: Cairo International Biomedical Engineering Conference (CIBEC), Giza, Egypt. New York (NY): IEEE; December 2014. pp 103–6. doi: 10.1109/CIBEC.2014.7020928 [DOI] [Google Scholar]
- 49. Hendrix N, Hauber AB, Lee CI, Bansal A, Veenstra DL. Provider preferences for attributes of artificial intelligence in breast cancer screening: a discrete choice experiment. J Clin Oncol. 2020;38(15 suppl):e14118. doi: 10.1200/JCO.2020.38.15_suppl.e14118 [DOI] [Google Scholar]
- 50. Vo V, Chen G, Aquino YSJ, Carter SM, Do QN, Woode ME. Multi-stakeholder preferences for the use of artificial intelligence in healthcare: a systematic review and thematic analysis. Soc Sci Med 1982. 2023;338:116357. doi: 10.1016/j.socscimed.2023.116357 [DOI] [PubMed] [Google Scholar]
- 51. Sun L, Yin C, Xu Q, Zhao W. Artificial intelligence for healthcare and medical education: a systematic review. Am J Transl Res. 2023;15(7):4820. [PMC free article] [PubMed] [Google Scholar]
- 52. Dumić-Čule I, Orešković T, Brkljačić B, Kujundžić Tiljak M, Orešković S. The importance of introducing artificial intelligence to the medical curriculum - assessing practitioners’ perspectives. Croat Med J. 2020;61(5):457–64. doi: 10.3325/cmj.2020.61.457 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53. Liu D-l, Li J. Cultivating practical literacy of machine learning for medical students. Basic Clin Med. 2021;41(7):1071–5. Available from: https://journal11.magtechjournal.com/Jwk_jcyxylc/EN/Y2021/V41/I7/1071. [Accessed 18 November, 2024]. [Google Scholar]
- 54. Wartman SA, Combs CD. Medical education must move from the information age to the age of artificial intelligence. Acad Med. 2018;93(8):1107. doi: 10.1097/ACM.0000000000002044 [DOI] [PubMed] [Google Scholar]
- 55. Vlasimsky R. VisiRad - IMIDEX. 31 March, 2020. Available from: https://imidex.com/visirad/. [Accessed 18 November, 2024].
- 56. Barnett AJ, Schwartz FR, Tao C, et al. A case-based interpretable deep learning model for classification of mass lesions in digital mammography. Nat Mach Intell. 2021;3(12):1061–70. doi: 10.1038/s42256-021-00423-x [DOI] [Google Scholar]
- 57. Lancaster HL, Zheng S, Aleshina OO, et al. Outstanding negative prediction performance of solid pulmonary nodule volume AI for ultra-LDCT baseline lung cancer screening risk stratification. Lung Cancer. 2022;165:133–40. doi: 10.1016/j.lungcan.2022.01.002 [DOI] [PubMed] [Google Scholar]
- 58. Stark GF, Hart GR, Nartowt BJ, Deng J. Predicting breast cancer risk using personal health data and machine learning models. PLoS One. 2019;14(12):e0226765. doi: 10.1371/journal.pone.0226765 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59. Nartowt BJ, Hart GR, Muhammad W, Liang Y, Stark GF, Deng J. Robust machine learning for colorectal cancer risk prediction and stratification. Front Big Data. 2020;3:6. doi: 10.3389/fdata.2020.00006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60. Hart GR, Roffman DA, Decker R, Deng J. A multi-parameterized artificial neural network for lung cancer risk prediction. PLoS One. 2018;13(10):e0205264. doi: 10.1371/journal.pone.0205264 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61. Robinson R, Liday C, Lee S, et al. Artificial intelligence in health care-understanding patient information needs and designing comprehensible transparency: qualitative study. JMIR AI. 2023;2:e46487. doi: 10.2196/46487 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62. London AJ. Artificial intelligence and black-box medical decisions: accuracy versus explainability. Hastings Cent Rep. 2019;49(1):15–21. doi: 10.1002/hast.973 [DOI] [PubMed] [Google Scholar]
- 63. Pierce RL, Biesen WV, Cauwenberge DV, Decruyenaere J, Sterckx S. Explainability in medicine in an era of AI-based clinical decision support systems. Front Genet. 2022;13:903600. doi: 10.3389/fgene.2022.903600 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64. Véliz C, Prunkl C, Phillips-Brown M, Lechterman TM. We might be afraid of black-box algorithms. J Med Ethics. 2021;47(5):339–40. doi: 10.1136/medethics-2021-107462 [DOI] [Google Scholar]
- 65. Bartoletti I. AI in healthcare: ethical and privacy challenges. In: Riaño D, Wilk S, ten Teije A, eds. Artificial Intelligence in Medicine. Cham (Switzerland): Springer International Publishing; 2019. pp 7–10. DOI: 10.1007/978-3-030-21642-9_2 [DOI] [Google Scholar]
- 66. Scherer MU. Regulating artificial intelligence systems: risks, challenges, competencies, and strategies. Harv J Law Technol. 2015;29(2). doi: 10.2139/ssrn.2609777 [DOI] [Google Scholar]
- 67. Carter SM, Rogers W, Win KT, Frazer H, Richards B, Houssami N. The ethical, legal and social implications of using artificial intelligence systems in breast cancer care. Breast. 2020;49:25–32. \10.1016/j.breast.2019.10.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68. Borkowski AA, Jakey CE, Thomas LB, Viswanadhan N, Mastorides SM. Establishing a hospital artificial intelligence committee to improve patient care. Fed Pract. 2022;39(8):334–6. doi: 10.12788/fp.0299 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69. Nong P, Hamasha R, Singh K, Adler-Milstein J, Platt J. How academic medical centers govern AI prediction tools in the context of uncertainty and evolving regulation. NEJM AI. 2024;1(3):AIp2300048. doi: 10.1056/AIp2300048 [DOI] [Google Scholar]
- 70. Roski J, Maier EJ, Vigilante K, Kane EA, Matheny ME. Enhancing trust in AI through industry self-governance. J Am Med Inform Assoc JAMIA. 2021;28(7):1582. doi: 10.1093/jamia/ocab065 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71. Parikh RB, Helmchen LA. Paying for artificial intelligence in medicine. NPJ Digit Med. 2022;5(1):1–5. doi: 10.1038/s41746-022-00609-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72. Abràmoff MD, Roehrenbeck C, Trujillo S, et al. A reimbursement framework for artificial intelligence in healthcare. NPJ Digit Med. 2022;5(1):1–6. doi: 10.1038/s41746-022-00621-w [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73. Khanna NN, Maindarkar MA, Viswanathan V, et al. Economics of artificial intelligence in healthcare: diagnosis vs. treatment. Healthcare. 2022;10(12):2493. doi: 10.3390/healthcare10122493 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74. EIT. EIT Health Think Tank—Healthcare Workforce and Organisational Transformation with AI. 2021. Available from: https://eit.europa.eu/library/eit-health-think-tank-healthcare-workforce-and-organisational-transformation-ai. [Accessed 19 November, 2024].
- 75. Jayakumar P, Moore MG, Furlough KA, et al. Comparison of an artificial intelligence-enabled patient decision aid vs educational material on decision quality, shared decision-making, patient experience, and functional outcomes in adults with knee osteoarthritis: a randomized clinical trial. JAMA Netw Open. 2021;4(2):e2037107. doi: 10.1001/jamanetworkopen.2020.37107 [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Supplemental material, sj-docx-1-mpp-10.1177_23814683251329007 for Primary Care Provider Preferences Regarding Artificial Intelligence in Point-of-Care Cancer Screening by Vinayak S. Ahluwalia, Marilyn M. Schapira, Gary E. Weissman and Ravi B. Parikh in MDM Policy & Practice

