Table 1.
Characteristics of included studies.
| Authors | Study design and methodology/data analysis | Method of data collection | Study population and sampling strategy | Number of participants | Participant characteristics | Topic of interest | AI studied | Main findings |
|---|---|---|---|---|---|---|---|---|
| Abuzaid et al. (2021)29 | Mixed methods, thematic analysis | Focus groups | Magnetic Resonance Imaging (MRI) technologists, snowball sampling | 98 | NR | Radiology | Hypothetical use of AI in radiology | MRI technologists have an understanding of AI and believe it would improve MRI protocols, reduce scan time, enhance post-processing and play a role in image interpretation. They highlighted a need for bespoke for AI training and practical sessions. |
| Adams et al. (2020)30 | Qualitative, thematic analysis | Focus groups | Patient and family advisors from urban and rural Canada, NR | 17 | 11 (55%) female | Radiology | Hypothetical AI used in radiology | Patient perceptions were described in four themes: 1. Fear of the unknown, 2. Trust, 3. Human connection, 4. Cultural acceptability. Patient priorities for AI were described in five themes: 1. Improving access to imaging and reducing wait times, 2. Reducing time to diagnosis, 3. Increasing diagnostic accuracy, 4. Improving communication 5. Empowering patients. |
| Bergquist et al. (2023)31 | Qualitative, NR | Semi-structured interviews | Mixed group of clinicians, researchers, and healthcare leaders in Sweden, purposive sampling | 25 | 5 (20%) female; median age 53 years; 8 radiologists, 6 other medical professionals, 7 management, 4 engineers/developers | Radiology | Hypothetical AI used in radiology | Trustworthiness of AI is related to: 1. Reliability, 2. Transparency, 3. Quality verification and 4. Inter-organizational compatibility |
| Buck et al. (2022)32 | Qualitative, grounded theory | Semi-structured interviews | General practitioners in Germany, convenience sampling | 18 | 9 (50%) female, 9 (50%) male | General | Hypothetical AI used in medical diagnosis | Three determinants of GPs' attitudes towards AI: concerns, expectations, and minimum requirements of AI-enabled systems. Individual characteristics and environmental influences as the 2 conditional determinants of GPs' attitudes toward AI-enabled systems. |
| Carter et al. (2023)33 | Mixed-methods, NR | Focus groups | Female members of the public aged 50–74 years old, Australia | 50 | Age: 20 (40%) aged 50–59, 21 (42%) aged 60–69, 9 (18%) aged 70–74 Personal history of breast cancer: 12 (24%) yes Education: 26 (52%) university educated |
Radiology | Hypothetical AI used in radiology (breast cancer screening) | There is broad acceptance of the use of AI in breast screening, conditioned on human involvement and AI performance |
| Cartolovni, Malesevic and Poslon (2023)34 | Qualitative, thematic analysis | Semi-structured interviews | Patients, clinicians, and healthcare leaders working in Croatia | 38 | Patients 11, clinicians 12, leaders 11; age range 18–65 years old | General | Hypothetical clinical AI | Four themes were developed: 1. The current state of healthcare and the patient-physician relationship, 2. Expectation of AI, 3. A synergistic effect between physicians and AI, 4. The future of healthcare and the patient-physician relationship |
| Chen et al. (2021)35 | Qualitative, thematic analysis | Semi-structured interviews and focus groups | Radiologists and radiographers working in five NHS organisations in England, snowball sampling/convenience sampling | Interviews: 18 FG: 8 |
Interviews: 12 (60%) radiologists, 8 (40%) radiographers FG: 8 (100%) radiographers |
Radiology | Hypothetical AI used in radiology | Considered responses to the use of AI in radiology in 1. Knowledge and 2. Attitudes, finding differences in attitudes towards AI between professional groups. |
| Drogt et al. (2022)36 | Qualitative, NR | Semi-structured interviews | Mixed group of professionals working in the pathology department of two hospitals in the Netherlands, convenience sampling | 24 | 15 (63%) pathologists, 17 (29%) laboratory technicians, 2 (8%) computer scientists | Pathology | Hypothetical AI used in pathology | Three recommendations for embedding AI in pathology: 1. Foster a pragmatic attitude toward AI development, 2. Provide task-sensitive information and training to health care professionals working at pathology departments, 3. Take time to reflect upon users' changing roles and responsibilities. |
| Faric et al. (2023)37 | Qualitative, thematic analysis | Semi-structured interviews | Mixed group of clinicians, healthcare leaders, and patients across 5 hospitals in the UK, Belgium, and the Netherlands | 39 | 12 patients: 10 (83%) female; 3 (25%) aged 40–50 years old, 3 (25%) aged 50–60 years old, 3 (25%) aged 60–70 years old, 3 (25%) aged 70–80 years old 25 clinicians/healthcare leaders: 6 (24%) female 2 healthcare leaders: 2 male |
Radiology | AI in use for diagnosing lung nodules | Four main themes were developed: 1. Perceived drivers and benefits, 2. Design of the tool and integration, 3. Appropriation of the tool by expert labour, 4. Clinical governance, quality assurance, maintenance, and post-market surveillance |
| Fazakarley et al. (2023)38 | Qualitative, thematic analysis | Semi-structured interviews | Clinicians and AI researchers in the UK | 13 | 5 (38%) female; mean age 38 years old, SD 9.1 years; 9 (69%) White British; 2 (15%) mixed or multiple ethnicities; 1 (8%) Asian; 1 (8%) other 3 (23%) doctors, 4 (31%) nurses, 2 (15%) IT technicians, 4 (31%) AI developer/researcher |
Cardiology | AI in use within a randomised-control trial, to diagnose coronary artery disease | Four themes were identified: 1. Positive perceptions of AI, 2. Potential barriers to using AI, 3. Concerns regarding AI use, 4. Steps needed to ensure the acceptability of future AI tools |
| Gillner (2024)39 | Qualitative, thematic analysis | Semi-structured interviews, 1 focus group | A mixed group of AI providers across Europe and clinicians | Interviews: 17 FG: 5 |
Interviews: 17 AI researchers/leaders FG: 5 clinicians |
General | Hypothetical clinical AI | Two major themes were developed: 1. Subsystems of complex healthcare systems, 2. Emergent practices of AI providers in healthcare |
| Haan et al. (2019)40 | Qualitative, grounded theory | Semi-structured interviews | Patients attending the radiology department of a tertiary care, academic hospital in the Netherlands for CT chest and abdomen, purposive sampling | 20 | 9 (45%) female; age 39–79 years old (mean age 64) | Radiology | Hypothetical AI used in radiology | Six key domains related to AI use in radiology: 1. Proof of technology, 2. Procedural knowledge, 3. Competence, 4., Efficiency, 5. Personal interaction, 6. Accountability |
| Hallowell et al. (2022)41 | Qualitative, NR | Semi-structured interviews | Membership of the “Minerva Consortium” and personal contacts of the study authors, convenience/snowballing sampling | 20 | Expertise: 9 (45%) clinical genetics, 2 (10%) paediatric genetics, 5 (25%) bioinformatics, 2 (10%) commercial, 3 (15%) other | Rare disease | Hypothetical AI used in diagnosing facial dysmorphology | Discussion of the value of trust in using AI for dysmorphology, concluding that trust in AI is grounded in its reliability, competence and “intentions.” |
| Held et al. (2022)42 | Qualitative, thematic content analysis | Semi-structured interviews | Mixed group of clinicians in Germany, convenience sampling | 24 | 10 (42%) female; average year of birth 1971; 16 (67%) general practitioner, 3 (13) medical assistant, 5 (21%) ophthalmologists | Ophthalmology | Hypothetical AI used to diagnose diabetic retinopathy | Main determinants of implementation have been identified: personal attitude, organisation, time, financial factors, education, support, technical requirement, influence on profession and patient welfare. |
| Helenason et al. (2024)43 | Mixed-methods, NR | Semi-structured Interviews | Primary care clinicians in Sweden | 15 | NR | Dermatology | AI used to diagnose skin lesions, proposed for use | Three major themes were identified: trust, usability and user experience and clinical context |
| Henry et al. (2022)44 | Qualitative, grounded theory | Semi-structured Interviews | Mixed group of clinicians in a 285 bed acute-care, US hospital, purposive sampling | 20 | 13 physicians (4 emergency department, 4 critical care, 5 general ward) and 7 nurses (3 emergency department, 4 critical care) | Sepsis | AI used to diagnose sepsis, in use by the institution | Themes identified included: lack of understanding of the difference between ML-based and conventional CDSS; ML-based systems play a supporting role; an overall willingness to trust AI despite lack of full understanding. Barriers highlighted included over-reliance on AI leading to deskilling. |
| Joshi et al. (2022)45 | Qualitative, thematic content analysis | Semi-structured interviews | Hospital leaders in the USA | 21 | 5 (24%) informatics leadership, 10 (48%) clinical leadership e.g., CMO, 6 (29%) other executive leadership, convenience sampling | Sepsis | AI for diagnosis of sepsis, in use by the institution | Identified several barriers and facilitators to implementation of sepsis-detection AI, identifies consideration of workflow integration, and clinician buy-in as two key approaches to overcome identified barriers. |
| Jussupow et al. (2021)46 | Qualitative, grounded theory | Semi-structured interviews and ethnography | Radiologists working in a hospital in Germany, with experience of using an AI system to diagnose stroke, snowball sampling | 14 | 2 chief radiologists, 4 senior radiologists, 8 assistant radiologists; mean self-reported diagnostic confidence (1–10) ranging from 4.3–10.0 | Radiology | AI for stroke diagnosis, in use at the institution | Described three patterns of AI use. “Sensedemanding”radiologists will evaluate AI results in both confirming and disconfirming AI, “Sensegiving” radiologists will reinforce use if AI confirms their findings. “Sensebreaking” radiologists find no benefit from AI. |
| Kim et al. (2024)47 | Qualitative, ethnography with abductive reasoning | Semi-structured interviews and ethnography | Mixed group of clinicians and healthcare leaders working at a hospital in the Netherlands | Ethnographic observation over 3 years; 18 interviews | NR | Radiology | 15 individual AI pipelines in use, for cross-specialty diagnostic tasks | Three key themes were developed to inform AI implementation: 1. Technology level, 2. Workflow level, 3. People and organisational level |
| King et al. (2023)48 | Qualitative, framework approach | Semi-structured interviews | Pathologists employed in UK hospitals, purposive sampling | 25 | 20 pathology consultants/attendings, 5 pathology trainees. 14 (70%) male, 11 (30%) female. | Pathology | Hypothetical AI used in pathology | Required features of AI identified by pathologists were trustworthiness and explainability, usability and workflow integration. Key contextual information and concerns about AI included the context of AI deployment, pathologists involvement with AI development, liability, evaluation and validation of AI and resources for AI. |
| Lebovitz et al. (2022)17 | Qualitative, grounded theory | Semi-structured interviews and ethnography | Radiologists working in three departments utilising diagnostic AI, NR | 33 + 500 h of ethnographic observation | NR | Radiology | AI in use for diagnosing breast cancer, classifying lung nodules and determining bone age | Only radiologists diagnosing lung cancer engaged with AI tools, despite high accuracy of all AI tools in the study. Explainability of AI is a necessary feature for clinician engagement, but on its own is permissive rather than sufficient. |
| Lombi and Rossero (2023)49 | Qualitative, template analysis | Semi-structured interviews | Radiologists working in a mixture of private and public hospitals in Italy, purposive sampling | 12 | 1 (8%) female, age range 36–64 years, 5 (42%) employed by private hospitals | Radiology | Hypothetical AI used in radiology | Three themes were developd: 1. ‘It will take time’ 2. ‘This is what being a radiologist means’ 3. ‘Don't be a DIY diagnostician!’ |
| Massey et al. (2023)50 | Mixed-methods, content analysis | Semi-structured interviews | Otolaryngologists, working in the USA, purposive sampling | 19 | 11 (58%) general otolaryngologists, 8 (42%) subspecialty rhinologists; 11 (58%) practicing in an academic setting. | Radiology | Hypothetical AI used in radiology for sinus CT interpretation | Six themes were identified: 1. Conventional reporting was indispensable for extra-sinus analysis, 2. Relationship with radiologist dictates trust in reporting, 3. Clinicians were open to utilizing AI, 4. Standardization of reporting was valued, 5. Anatomical analysis was preferred over descriptive assessments, 6. Trust in AI could be improved with additional validation in the literature |
| Mosch et al. (2022)51 | Qualitative, thematic analysis | Semi-structured interviews | Mixed group of participants with expertise in the field of AI in medicine, medical education, and training, purposive sampling | 24 | Professional background: 15 (63%) medical, nine (38%) computer science, 3 (13%) medical education, 8 (23%) other | General | Hypothetical clinical AI | Three themes were developed: 1. Specific tasks of physicians will be taken over by AI systems, 2. AI will not replace physicians, 3. Ways of work: AI will transform how healthcare is delivered. |
| Nelson et al. (2020)16 | Qualitative, grounded theory | Semi-structured interviews | Patients attending general dermatology clinics and melanoma clinics at a hospital in the USA, purposive sampling | 48 | 26 (54%) female; mean (SD) age 53.3 (21.7) years old; 16 (33%) history of melanoma, 16 (33%) history of non-melanomatous skin cancer, 16 (33%) no history of skin cancer; 45 (94%) White, 2 (4%) American Indian or Alaskan Native, 1 (2%) African American | Dermatology | Hypothetical AI used in dermatology for skin lesion classification | Patients describe a preference for AI as an assistive tool, rather than a replacement for a clinician. Increased diagnostic speed, accuracy and healthcare access were commonly perceived benefits of AI, but perceived risks included increased patient anxiety, AI errors and loss of human interaction. |
| Ng et al. (2022)52 | Qualitative, Phenomenology/thematic analysis |
Focus groups | Radiographers working in public institutions in Singapore, purposive sampling | 22 | 11 (50%) female; age 23–42 years old (median 30.5 years); working experience 1–18 years (median six years) | Radiology | Hypothetical AI used in radiography | Four themes were developed from the data: 1. Knowledge of AI and its applications, 2. Perceptions on the use of AI in radiographic practice, 3. Patients' perceptions as viewed by radiographers, 4. Prospective applications and expectations of AI. |
| Pelayo et al. (2023)53 | Qualitative, framework analysis | Semi-structured interviews | Latinx patients with T2DM at a single health center in the USA | 20 | 12 (60%) female; mean age 59.8, range 14 | Ophthalmology | Hypothetical AI used to diagnose diabetic retinal disease | Patients strongly prefer human review rather than AI; if AI is integrated it should be as a tool rather than a replacement |
| Prakash et al. (2021)54 | Qualitative, thematic analysis | Netnography, semi-structured interviews | Radiologists working in India, purposive sampling | 15 | 5 (33%) female; mean age 40.7 years old, range 28–62 | Radiology | Hypothetical AI used in radiology | Themes were developed from qualitative data: 1. Perceived threat, 2. Medico-legal risk, 3. Performance risk, 4. Performance expectancy, 5. Trust, 6. User resistance |
| Pumplun et al. (2021)24 | Qualitative, directed content analysis | Semi-structured interviews | Mixed group of AI experts, with detailed knowledge of clinical processes and AI, theoretical sampling approach | 22 | 5 (23%) clinicians, 8 (36%) clinicians with leadership roles, 9 (25%) managers or IT staff; between 3 and 40 years of work experience | General | Hypothetical AI used in diagnosis | Developed a maturity model to score the readiness of a clinic for AI adoption, spanning three dimensions: organisation, adopter system and patient data. |
| Rabinovich et al. (2022)55 | Mixed-methods, NR | Structured interviews | Mixed group of clinicians in a hospital in Argentina, with experience of using diagnostic AI, NR | 6 | 3 (50%) emergency physicians and 3 (50%) radiology residents | Radiology | AI in use in the institution, for diagnosing pneumothoraces, rib fractures, pleural effusions, and lung opacities on chest radiographs | Participants in general had positive experiences with using the diagnostic AI. They describe using it as a second opinion, to reduce human error, and valued its use in diagnostic confirmation. |
| Redrup Hill et al. (2023)56 | Qualitative, NR | Focus groups | Mixed group of patients/clinicians, researchers and healthcare leaders | 31 | 4 software developers, 7 pathologists, 11 leaders, 9 patients/clinicians | Pathology | Existing AI to diagnose Barett's oesophagus or adenocarcinoma from pathology specimens | Six themes were developed: 1. Risks and potential harms, 2. Impacts on human experts, 3. Equity and bias, 4. Transparency and oversight, 5. Patient information and choice, 6. Accountability, moral responsibility and liability for error |
| Richardson et al. (2021)26 | Qualitative, grounded theory | Focus groups | Patients who had a recent primary care visit at a large academic health centre in the USA, convenience sampling | 87 | 49% female; average age 53.5 years old; 93% white and 94% non-Hispanic/Latino; 87% education level higher than a high school degree; 20% employment history in technology or computer science; 45% employment history in healthcare/health science | General | Hypothetical clinical AI, using case studies to ground discussion | Description of six themes: excitement about healthcare AI but needing safety assurances, and expectation for clinicians to ensure AI safety, preservation of patient choice and autonomy, concerns about healthcare costs and insurance coverage, ensuring data integrity, and the risks of technology-dependent systems. |
| Richardson et al. (2022)27 | As above | As above | As above | As above | As above | As above | As above | Developed a conceptual framework for understanding how patients evaluate healthcare AI, based on patient experiences (with illness, health technology, relationship with clinicians, social context and familiarity with technology), beliefs (about healthcare and technology) and attitudes towards AI in healthcare (attitude formation, perceived acceptability and support for development). |
| Robertson et al. (2023)57 | Mixed-methods, NR | Semi-structured interviews | Patients recruited from cardiac clinics in Tucson, Arizona; convenience sampling | 24 | 16 (67%) female; age range 19–92 years old; 10 (42%) White, 8 (33%) Hispanic, 3 (13%) Black, 2 (8%) Native American, 1 (4%) Asian, 7 (29%) University education | General | Hypothetical clinical AI | Narrative overview of qualitative data; patients discussed fallibility of AI systems, trust related to healthcare systems, knowledge of AI in use, confidence in human physicians and religious belief |
| Sangers et al. (2021)58 | Qualitative, thematic content analysis | Focus groups | Members of the public who took part in a customer panel of a Dutch health insurer, and social media platforms; purposive sampling | 27 | 18 (67%) female; mean age 37.3 years, range 19–73; all use a smartphone at least every half day; 20 (74%) no history of skin cancer, 4 (15%) personal history of skin cancer, 3 (11%) family history of skin cancer | Dermatology | Hypothetical AI used in diagnosing skin cancer | Barriers to using AI apps for skin cancer diagnosis were: perceived lack of value, perception of untrustworthiness, preference for humans, concerns about privacy, complex user interface and increased costs. The facilitators were high perceived value, transparent and trustworthy identity of AI developers, endorsement by clinicians and regulatory bodies, easy to use interface and low costs. |
| Sangers et al. (2023)18 | Qualitative, grounded theory | Focus groups | Dutch dermatologists and GPs identified through social media and via specialty newsletters, purposive sampling | 33 | Mean age 35.6 years, range 31–62; 17 (52%) female; 17 (52%) general practitioner, 16 (49%) dermatologist | Dermatology | Hypothetical AI used in diagnosing skin cancer | Dermatologists and GPs described preconditions for implementation: adequacy of algorithms, sufficient usability and accessibility, validation and regulation/clear liability, national guidance; they described benefits including improved health outcomes, care pathways and education. They described perceived barriers as doubts about AI accuracy, exacerbation of health inequalities, fear of replacement by AI, extra time taken to use AI and commercialization and privacy concerns. |
| Satterfield et al. (2019)59 | Qualitative, thematic analysis | Semi-structured interviews | 3 groups of researchers: diagnosis, AI, “Learning Health Systems”, NR | 32 | 18 (56%) from the “improving diagnosis” research group, 6 (19%) from AI research, 8 (25%) from the “Learning Health Systems” group | General | Hypothetical AI used in diagnosis | There is limited collaboration between the research communities, and the authors emphasise the importance of forming a multi-disciplinary “learning community” to ensure uptake of AI in diagnosis. |
| Scheetz et al. (2021)60 | Mixed-methods, thematic analysis | Semi-structured interviews | Mixed group of clinicians, including doctors and AHP, with experience of using an AI tool to screen for diabetic retinopathy, in outpatient clinicians in Australia, convenience sampling | 8 | 3 (37.5%) male doctors, 5 (62.5%) female AHP | Ophthalmology | AI to screen for diabetic retinopathy | Participants agreed that the AI tool was easy to use and interpret, but reported challenges in explaining findings to patients, and allocating enough time to use the tool. They reported the requirement for validation of any AI tool to increase trust, and the value of AI was felt to be reducing the burden on individual clinicians. |
| Sibbald et al. (2022)61 | Qualitative, content analysis | Semi-structured interviews | Emergency department physicians with personal experience of using an AI tool to support differential diagnosis (EDS) at triage, purposive sampling | 13 | 2 (15%) female; 5 (38%) <5 years of practice, 4 (31%) 5–10 years, 1 (7%) 11–20 years, 3 (23%) >20 years; 6 (46%) family medicine specialists with subspecialisation in emergency medicine, 7 (54%) emergency medicine specialists | Emergency medicine | AI in use to generate differential diagnosis for emergency medicine triage | Four themes were identified: 1. The quality of EDS was inferred from the scope and prioritization of the diagnoses, 2. Trusting EDS differential diagnoses was linked to varied beliefs around the diagnostic process and potential for bias, 3. Who benefits? Not me, 4. Information flow between EDS and the Electronic Medical Record. |
| Strohm et al. (2020)25 | Qualitative, NR | Semi-structured interviews | Mixed group of radiologists, managers, implementation consultants and data scientists with experience using an AI for automating bone maturity assessments (BoneXpert), sampling for maximal variation | 24 | 20 (83%) radiologists, 5 of which have a dual role as data scientists/managerial, 4 (17%) managers | Radiology | Hypothetical AI used in radiology, with reference to BoneXpert, an AI developed by a commercial company (Visiana) that automated bone maturity assessments using paediatric hand X-rays | Using the NASSS framework, identified facilitating and hindering factors for AI implementation, with one of the most important barriers identified as the non-acceptance of AI by clinicians. |
| Townsend et al. (2023)62 | Mixed-methods, thematic analysis | Semi-structured interviews | Clinicians with current or previous emergency department roles, located in the UK | 9 | 4 (44%) female; age range 20–59 years; experience in emergency medicine range 1 month–22 years | Emergency department | AI in use to generate differential diagnosis for emergency medicine triage | The overarching theme is ‘trust’, with five subthemes: 1. Social, 2. Legal, 3. Ethical, 4. Empathetic, 5. Cultural |
| van Cauwenberge et al. (2022)63 | Mixed-methods, thematic analysis | Think-aloud interviews | Physicians working in a large tertiary care academic hospital in the Netherlands, purposive sampling | 30 | 16 (53%) female; 7 (23%) in training, 8 junior (27%), 15 (50%) senior | General | Hypothetical AI for general clinical and diagnostic support | Four themes were developed: 1. Transparency, 2. Obstructivity, 3. Type of problem, 4., Certainty of advice |
| Wenderott et al. (2024)64 | Qualitative, | Semi-structured interviews | Radiologists and radiology residents in a hospital in Germany, convenience sampling | 12 | 8 (67%) 2–4 years of work experience; 5 (42%) worked in department <1 year, 4 (33%) worked in department 1–3 years | Radiology | AI in use to diagnose prostate lesions on MRI | Findings were categorised into AI benefits/risks, barriers/facilitators, external factors influencing AI adoption and contradictory statements |
| Winter and Carusi (2022)65 | Qualitative, thematic analysis | Focus groups | Mixed group of professionals involved developing AI for clinical use, and patients/carers with lived experience of pulmonary hypertension, NR | 21, split into two FG (10, 11) | FG1: 4 (19%) computer scientists, 4 (19%) clinicians,2 (10%) researchers, 1 (4%) patient representative FG2: 6 (29%) patients, 4 (19%) carers, 1 (4%) patient representative |
Respiratory | Hypothetical AI used to diagnose pulmonary hypertension | Four themes were developed: 1. AI can result in early diagnosis, 2. Early diagnosis outweighs data risks of privacy and reuse, 3. Responsibility lies with specialist clinicians, 4. AI will result in deskilling of professionals. |
| Winter and Carusi (2022)66 | Qualitative, thematic analysis | Semi-structured interviews and ethnography | Mixed group of professionals involved in the development of a screening algorithm for pulmonary hypertension, NR | 3 | 2 (67%) researchers, 1 (33%) clinician | Respiratory | AI under development to screen for pulmonary hypertension. | Collaboration between clinicians and researchers is encouraged, particularly in 1. Querying datasets, 2. Building the software and 3. Training the model. |
NR, Not reported.