Skip to main content
PLOS One logoLink to PLOS One
. 2023 Feb 27;18(2):e0281733. doi: 10.1371/journal.pone.0281733

Implementing artificial intelligence in Canadian primary care: Barriers and strategies identified through a national deliberative dialogue

Katrina Darcel 1,2, Tara Upshaw 1,3, Amy Craig-Neil 1, Jillian Macklin 1,2,4,5, Carolyn Steele Gray 6,7, Timothy C Y Chan 8, Jennifer Gibson 4,5, Andrew D Pinto 1,5,9,10,*
Editor: Ardashir Mohammadzadeh11
PMCID: PMC9970060  PMID: 36848339

Abstract

Background

With large volumes of longitudinal data in electronic medical records from diverse patients, primary care is primed for disruption by artificial intelligence (AI) technology. With AI applications in primary care still at an early stage in Canada and most countries, there is a unique opportunity to engage key stakeholders in exploring how AI would be used and what implementation would look like.

Objective

To identify the barriers that patients, providers, and health leaders perceive in relation to implementing AI in primary care and strategies to overcome them.

Design

12 virtual deliberative dialogues. Dialogue data were thematically analyzed using a combination of rapid ethnographic assessment and interpretive description techniques.

Setting

Virtual sessions.

Participants

Participants from eight provinces in Canada, including 22 primary care service users, 21 interprofessional providers, and 5 health system leaders

Results

The barriers that emerged from the deliberative dialogue sessions were grouped into four themes: (1) system and data readiness, (2) the potential for bias and inequity, (3) the regulation of AI and big data, and (4) the importance of people as technology enablers. Strategies to overcome the barriers in each of these themes were highlighted, where participatory co-design and iterative implementation were voiced most strongly by participants.

Limitations

Only five health system leaders were included in the study and no self-identifying Indigenous people. This is a limitation as both groups may have provided unique perspectives to the study objective.

Conclusions

These findings provide insight into the barriers and facilitators associated with implementing AI in primary care settings from different perspectives. This will be vital as decisions regarding the future of AI in this space is shaped.

Introduction

With advancements in artificial intelligence (AI) and vast volumes of health data produced through electronic health records (EHRs), we are in an opportune time to explore ways in which AI can be applied in healthcare settings. Although many of the advancements have taken place in the field of radiology [1], AI applications such as online scheduling tools, drug dosing algorithms, and digitization of medical records have been implemented in many other areas of healthcare, including primary care [2].

Applying AI in primary care settings poses unique challenges as this field of medicine encompasses a wide variation of health concerns in diverse patient populations [3]. Primary care EHR data is unique in that it is longitudinal, it covers undifferentiated and broad health concerns, it serves populations of differing socioeconomic statuses, it contributes to public health records, and it contains data to inform preventative care practices. Despite these distinct qualities, AI is beginning to be used in primary care in the areas of clinical decision making, risk prediction, care management, and proactive detection [4]. Examples of risk prediction AI tools applicable to the primary care space include machine learning algorithms that identify patients at risk for cancer [5], predict which patients may experience post-partum depression [6, 7], and estimate those who may be at risk of multiple chronic diseases including diabetes, osteoarthritis, and hypertension [8, 9]. Although a flurry of research is focused on the development of AI tools, implementation into practice remains one of the most challenging barriers to overcome.

To identify and ease these barriers, researchers have developed models related to technology implementation including the ‘Technology Acceptance Model’ (TAM) and the ‘Unified Theory of Acceptance and the Use of Technology’ (UTAUT) model [10, 11]. These are well studied models that explore the reasons that users accept or reject a given technology and how acceptance can be improved, but unfortunately are not specific to primary healthcare or even to the healthcare at all. Another model, Sittig and Singh’s sociotechnological model for studying health information technology includes eight dimensions (hardware and software computing infrastructure, clinical content, human-computer interface, people, workflow and communication, internal organizational features, external rules and regulations, and measurement and monitoring [12]. Sittig and Singh’s model focuses specifically on implementation of technology in complex adaptive healthcare systems and will be used as a basis of comparison for the findings of this study.

Other researchers including Lovejoy et al. and Singh et al. have conducted reviews to explore the challenges of implementing AI in healthcare settings. The recent review by Lovejoy et al. explores the challenges of implementing AI to healthcare by breaking down the AI innovation process into three stages–invention, development, and implementation. Specifically with respect to the implementation stage, they highlight three main barriers of focus (generalizability, regulation, and deployment). Generalizability refers to the need for diversity within the training set, regulation refers to the need for ongoing monitoring of an algorithm or tool throughout its lifetime to ensure continued safety, and deployment refers to the need to minimize disruption to workflow and account for interoperability between healthcare technologies [13]. Although this review is not specific to the primary care space, it provides insights into barriers that are commonly encountered in the healthcare AI space, all of which will help inform our study. Another review by Singh et al. outlines the barriers in relation to the perspective of healthcare organizations, healthcare providers, and patients. Interestingly, from the healthcare organization perspective, they discuss the role of “company culture” as important when considering the use of embracing AI in their day-to-day practice. The provider perspective outlined concerns for the consideration of racial, ethnic, gender, and other sociodemographic characteristics within the algorithm. The perspective of patients was centered on trust, including the concern for physician-patient confidentiality if AI were to be used in their care [14]. This review provides vital insight from various stakeholder perspectives and emphasizes the importance of doing so.

To effectively implement AI in primary care settings, it is critical that these barriers are explored from multiple perspectives, as emphasized by Singh et al. above. Thus, we engaged patients, providers, and healthcare leaders in a deliberative dialogue series to gain insight into perceived challenges associated with AI. These findings will help assist those that are being called upon to make pivotal decisions regarding the future of AI in primary healthcare.

Methods

Setting

This study was conducted virtually within Canada, a high-income country with provincially and territorial-based public health insurance that covers all necessary hospital and physician services [15]. Both within and across the provinces and territories of Canada, primary healthcare is delivered through different models. Largely, it is funded following a fee-for-service (FFS) model, but has adopted new funding models such as the enhanced fee-for-service model and capitation model. For example, in Ontario, the enhanced FFS model is similar to the traditional FFS model, except that it includes additional targeted fee increases, extended hours, performance-based initiatives, and patient enrollment. In contrast, the capitation model reimburses primary care physicians for each person attributed to them, based on age and sex adjusted factors [16]. More recently, Ontario has been forming Ontario Health Teams including family physicians, nurse practitioners, social workers, dietitians, and other healthcare professionals to increase interdisciplinary care for a given community [17].

Design

This emergent qualitative study used virtual deliberative dialogues. Deliberative dialogue is a participatory method in which people affected by an issue are gathered to provide advice to decisionmakers [1820]. With facilitator support, participants work together to interpret existing evidence and identify approaches to the issue that reflect their—sometimes conflicting—values and tacit knowledge [21, 22]. We selected this method because we believe that the values and wisdom of primary care stakeholders can guide health AI in directions that support equity and advance the Quadruple Aim (optimizing health system performance through improved patient experience, improved clinical experience, better outcomes, and lower costs [2123]).

To adhere to COVID-19 public health restrictions, we held twelve 90-minute dialogues by video conference between September 8th and October 15th, 2020. Dialogues were divided into three rounds and all sessions were facilitated by one trained Master’s-level graduate student (TLU). Before taking part, participants reviewed an information package that provided an overview of health AI, key definitions, examples of AI in medicine, possible uses in primary care, and ethical challenges. The study was approved by the research ethics boards of Unity Health Toronto and the University of Toronto. Verbal informed consent was obtained from all participants and documented prior to scheduling their first dialogue session.

Participants

Patients and primary care providers were recruited under distinct purposive variation sampling frames from various online advertisement platforms and email distribution lists. This recruitment method is outlined further by Patton et al. [24], and was chosen in order to optimize variation in age, gender, race and ethnicity, education level, income, province of residence for patients and gender, race and ethnicity, provider type, country of health professions education, years in practice, practice size, and province of practice for primary care providers. Specifically, patients were recruited using Kijiji and social media channels across 10 provinces as well as patient advisor distribution lists (patient advisory network in British Columbia, patient-family advisory council of downtown Toronto hospitals). Primary care physicians were recruited through social media channels, primary care research network news channels, and through email invitations from AP to practice colleagues and those in his network interested in digital health and health informatics. To be eligible, patients must have seen their primary care provider once within the last year and providers had to work at least one day in the clinic per week. To promote variation, we collected sociodemographic information during consent interviews and then iteratively adjusted recruitment strategies. We used a critical case frame for system leaders, seeking those involved in digital health, health informatics, or primary care governance [24]. Patients and providers were invited to take part in up to three rounds, while system leaders were invited to join one final-round session or to provide feedback on the dialogue from the first two sessions by survey. The sessions were designed this way to allow for interactive dialogue between the different perspectives of patients, providers, and health system leaders in real-time. An honorarium of $32.50 USD was offered to participants for every session attended. TLU was acquainted with two patient participants. AP was known to most providers in the study and observed three sessions.

Data collection & analysis

Audiovisual recordings and fieldnotes were generated for each session by one of the observers (ACN, JM). Following a session, TLU and the designated observer met after each session to analyze fieldnotes using interpretive description methods, coding for concerns, perceived barriers, and facilitators of AI implementation in primary care [2528]. This analysis guided the emergent nature of the study (e.g., participant-facing summaries, use cases, etc.) and generated an initial code list. Early codes were later confirmed and expanded upon through dual-coder thematic analysis of verbatim transcripts (TLU, KD). Transcript analysis occurred in two passes, the first being entirely inductive and the second using agreed upon codes. TLU, KD, and AP then met to review coded transcripts and agree on final interpretations. The final interpretations were compiled into four major themes and then compared to Sittig and Singh’s sociotechnical model for studying health information technology (HIT) in complex adaptive health systems [12]. Data was managed using Microsoft Excel and Word. Saturation was reached for most concepts after nine sessions.

Rigor

Dialogue guides for rounds 2 and 3 were not pilot tested because of the study’s emergent nature. These guides were iterated upon and further developed as the study progressed. Investigator triangulation was achieved through the involvement of multiple researchers in data collection and analysis [29]. Data triangulation was achieved through careful comparison of field notes, primary transcripts, researcher interpretive reflections, and participant activities throughout the study. We created multiple opportunities for participants to member-check researcher interpretations. Eighteen participants provided reflections on preliminary findings at the study’s conclusion.

Results

Participants

Three rounds of deliberative dialogue sessions were completed with 48 participants from 8 Canadian provinces. This included 22 patients, 21 primary care providers, and 5 health system leaders; participants’ characteristics can be found in S1 Table and have been published elsewhere [30].

Findings

From these dialogue sessions, a series of barriers and strategies to implement AI in primary care emerged, which were grouped into four major themes: (1) system and data readiness, (2) the potential for bias and inequity, (3) the regulation of AI and big data, and (4) the importance of people as technology enablers. Within each of these 4 themes, participants spoke about trust, either in reference to trust in the technology or in the provider using it, so trust will be discussed within each theme. When asked what they thought about algorithm explainability, all but one of the patient participants agreed that they would trust AI algorithms used in their care by proxy of the trust they placed in their provider. Patients in one session suggested that health AI-enabled harms could damage public trust in health institutions. Patients and providers across sessions identified regulatory oversight, scientific evidence, and participatory co-development processes as builders of trust.

System and data readiness

Significant concern around adoption and the readiness of Canada’s health system came up in many sessions and was particularly focused on both the readiness of our current technology/data, as well as the willingness for clinics to engage with and adopt new AI technologies. Some primary care clinics still operate using paper and fax machines, and therefore implementing AI may be too advanced. Additionally, primary care clinics operate independently and are fragmented from one another. Each clinic "enters things in all different ways. We have to make sure that the tool is good enough to manage the many different ways that people get information and enter information" (PR-7905). Creating an adoption strategy and designing tools to accommodate each clinic may be challenging.

Participants acknowledged the potential for AI to improve health and healthcare. However, a few patients and most providers recognized that the optimal performance of health AI tools depends upon a level of HIT interoperability that does not exist in many jurisdictions. For AI tools to reliably inform clinical decision-making and support integrated, equitable care, providers believed that algorithms must synthesize data from multiple sources, including administrative databases and EHRs across levels of care and care settings. They also believed that certain advanced AI tools (e.g., rare disease diagnosis) would have to synthesize an exponentially increasing body of evidence and ever-changing guidelines to be clinically useful.

Furthermore, when discussing the willingness to adopt AI technology, it was highlighted that gaining buy-in from clinic owners and operators may be challenging since many of them were promised a series of quality improvements during the EHR rollout that were not met. As one participant suggested, “the quality improvements that were expected didn’t happen” (PR-4404). They may be hesitant to invest and place trust in an AI system unless they can clearly and quickly realize the benefit of doing so.

Potential for bias and inequity

With varying levels of technological readiness and challenges with clinic uptake, the potential for biased algorithms and inequitable access to AI tools becomes increasingly concerning. Several providers worried that lagging digitization in resource-constrained settings (e.g., rural communities), poor interoperability of established HIT systems, or data monopolies held by leading EHR vendors could negatively impact health equity by driving the uneven application of AI tools across settings, or limiting datasets available for developing AI tools to smaller, non-representative populations, thereby increasing the risk of bias and inequity.

The potential for bias was a concern that came up in every dialogue session. Some participants brought up past studies where AI tools were proven to further perpetuate the biases of their developers and of society. As one patient participant mentioned, “there could be bias in your software. Because these factors are not all equal, they have different weights of importance. And so how that’s applied will affect how good the results are going to be” (PT-5981). There was also concern for the bias that could be present in the guidelines and criteria that feed the AI algorithms. Participants were worried that if pharmaceutical companies or other private enterprises sponsored the tool, they may bias the algorithms to produce outcomes that favor their product. Participants linked this theme to notions of trust. As evidenced by the following quotation, “we want to make sure that the tools we use don’t create new problems… they are an opportunity to address some of the biases that already exist in our system. . .and that these tools are probably only as good as the data we provide them” (PR-7905).

Equity and access to AI technology were discussed from many different perspectives throughout the dialogue sessions. From a patient perspective, there was expressed concern that some populations may not want to or be able to interact with AI tools, which as a result, may further deter them from the health system as a whole and widen the care gap that already exists. From a primary care clinic and provider level, the clinics that are frequently last to procure EHRs and new technology will likely also be the last to adopt a new AI tool; thus, we may further widen the divide between the advantaged and disadvantaged primary care practices and the individuals they treat. Overall, while participants are hopeful for the potential that AI holds, it will likely not serve populations that most frequently “fall through the cracks” (PT-4653), and therefore it is important to consider these populations and their ability to access and trust the tool before it is implemented widely in primary care.

Regulation of AI and big data

Participants from all three groups believed that inadequate regulatory oversight of health AI tools places the public at risk of harm to physical health or wellbeing. This was a perceived barrier to health AI adoption because it decreased patient and provider trust in available tools. To feel comfortable using an AI tool in clinical decision-making, providers wanted assurance that it had been approved by a regulatory authority after meeting rigorous standards of performance, in terms of accuracy, risk of bias, and patient safety. Participants from all three groups, but especially patients, were concerned that existing privacy laws do not adequately prevent AI developers from monetizing PHI or allowing PHI to be used in ways that impact patient wellbeing (e.g., insurance or loan decisions). One provider stated that they would like to be aware of “how this data is being used and whether there were any concerns about this data being used by a [private] company” (PR-7905) and a patient mentioned, “I think it’s really important to clearly set out what the data be used for and ensure that the patient understands how the confidential confidentiality and privacy still applies in those situations" (PT-4653). Evidently, all participants stressed the importance of regulatory oversight to ensure the tools are frequently monitored for safety and that there is insight into the ownership, control, and access of health data to ensure it is not being monetized or used inappropriately.

Importance of people as technology enablers

People play a vital role in the success of any given technology; therefore, the barriers that impact the people using and affected by the technology must be properly mitigated. From the dialogue sessions, these barriers included the need for provider training and skill preservation, the fear of changing the provider-patient relationship, and the challenge of designing tools to properly suit the user’s needs. Firstly, many providers were unsure about the training that would be needed to successfully adopt a given AI tool and were worried that it may impede their clinic workflow. As stated by one provider, “oftentimes, you have to adapt your workflow and sometimes your clinical processes to accommodate the tool where I really strongly feel it needs to be the reverse” (PR-4404). After being trained on the tool, some were concerned that the clinical skills and reasoning of primary care providers could be at risk due to dependence on the tool. As one provider said, “we don’t want to lose that ‘art of medicine’ that we have learned over the years, because it’s a very challenging profession where you see people with undifferentiated complaints, so you have to know something about everything. And with AI, if we rely too much on it, we might lose that clinical judgment." (PR-1783).

Secondly, some participants feared that these technologies could harm patient-provider relationships, while others thought it may enhance them. For example, one patient mentioned, “I want my doctor to be present with me in a conversation and not staring at the computer and looking at all the tools or boxes that they need to check off” (PT-4159). Many agreed with this notion, but some did refute and stated that if the AI tools are designed correctly, they may provide more space and time for relationship building and trust between the patient and the provider. Lastly, all participants agreed that the tools must be designed deliberately, to beneficially serve the needs of the users and not contribute to patient, provider, or administrative burnout.

Strategies to promote adoption

To promote the adoption of health AI in primary care and beyond, participants offered a host of strategies to build patient and provider trust in the technology based on each theme (summarized in Table 1). Of these, the most robustly supported was a participatory AI co-development model that meaningfully involves patients, interprofessional primary care providers, clinical administrative staff, HIT specialists, and health policymakers. Participants believed that stakeholder engagement must occur upstream of piloting, ideally at the conception phase, and remain ongoing. They also specified that stakeholders must be diverse in terms of sociodemographic characteristics (e.g., race or ethnicity, gender, income), geographic location, and healthcare needs, as well as professional settings for providers (e.g., telehealth, walk-in, community health centers). Through this model, described by a patient as “designing with, rather than for,” stakeholders could co-construct AI tools that address shared application priorities and promote patient-centered practices. Within the model, participants viewed processes that facilitate iterative stakeholder input (e.g., a participatory piloting program) as particularly useful for ensuring usability and workflow integration, correcting for bias, and improving algorithm performance through feedback. To specifically help overcome the barriers associated with system and data readiness, investments are required to allow for different technological systems to “talk to one another”. This would thereby enable integration between primary care clinics using different EHR technologies.

Table 1. Strategies to promote adoption based on each theme.

All four themes
Participatory Co-design Ensure all stakeholders are at the table during both the design and implementation of these tools. Include representative stakeholders from different communities and different types of users (administrative clerks, providers, patients etc.)
Design and Implement Iteratively Use iterative design and implementation strategies to allow for stakeholder feedback throughout. This will ensure the tool is designed for those that it will affect (and not the agenda of a private business)
Learn from Pilot Projects Use pilot projects as examples to learn
Theme 1: System and data readiness
Interoperability Invest in interoperability to enable different technological systems to “talk to one another” thereby enabling integration between clinics and hospitals
Technology Infrastructure and Architecture Organize technology infrastructure and architecture to support AI tools
Create a Robust Adoption Strategy Arrange for adoption strategy where providers can slowly ease into using the technology. It may begin with tighter constraints at the start, and overtime as trust is built, it can be given more autonomy (staged approach). This strategy should also be versatile enough to be applicable to many different areas of primary care clinics.
Theme 2: Potential for bias and inequity
Unbiased Training Data Use unbiased, representative, good quality data to feed algorithms
Prevent Bias through Frequent Re-assessment Ensure AI tool remains free of bias overtime by conducting routine re-assessments and analyses of the tools
Ethics Training Consider designing ethics training courses for all of the stakeholders who will be involved in the design, testing, implementation, and use of AI tools
Theme 3: Regulation of AI and big data
Leadership Commitment Encourage health leaders and government players to demonstrate commitment to AI (lead by example)
National Rules & Regulations Create and implement national AI rules and regulations for healthcare (there is a need for content standards and protocols to further enable interoperability, and a need for regulatory standards to ensure representation and bias mitigation within the algorithms and the training data sets)
Design Regulations based on Level of Risk Create regulations based on the level of risk that the tool may pose to its users. For example, there may be different regulations for tools that conduct administrative-related functions than those that are designed for clinical decision making and diagnostic functions
Theme 4: Importance of people as technology enablers
Value Proposition Create a clear value proposition for each user/person affected by technology (what’s in it for them? What benefits will they see from this technology?)
Ease into the Technology Start with AI technology that can complete basic operational tasks before transitioning into the implementation of complex AI systems
Transparency Outline all of the benefits and limitations of the technology to all affected and communicate these to the users to manage expectations
AI Literacy Provide education and transparency into what AI tools are and how they are used in order to reduce concern of the unknown and enable all stakeholders to participate in co-development conversations
Deliberate Design Deliberately select AI technologies to align with what we want humans to do versus what we want technology to do. There is a need to be deliberate and weigh the benefits and risks before implementing AI technologies
Create a Robust Adoption Strategy Arrange for adoption strategy where providers can slowly ease into using the technology. It may begin with tighter constraints at the start, and overtime as trust is built, it can be given more autonomy (staged approach). This strategy should also be versatile enough to be applicable to many different areas of primary care clinics.

The second theme, the potential for bias and inequity, may begin to be mitigated by promoting the use of unbiased, representative, good quality data to feed algorithms. It was recommended by one participant that to ensure the AI tool remains free of bias overtime, it should be re-assessed and analyzed routinely to check for bias. There was an expressed desire for ethics training courses for those involved at many stages of the design, implementation, and evaluation life cycle of the technology.

Strategies to assist with the regulation of AI and big data will require strong leadership commitment by health leaders and government players. Providers felt that this commitment to health AI would act as an example for primary care clinics to follow. Additionally, to increase trust in the technology, national AI rules and regulations for healthcare must be designed and implemented. There was an expressed need for content standards and protocols to further enable interoperability, and a need for regulatory standards to ensure representation and bias mitigation within the algorithms and training data sets.

Building trust, preventing burnout, and preserving patient-provider relationships begins with outlining a clear value proposition for each user and individual affected by the technology. It was highlighted that the benefits, limitations, and uses of the technology must be transparently communicated. Following this, simple AI technology that can complete basic operational tasks should be implemented first, before transitioning into the use of complex AI systems to encourage uptake and trust in the technology.

Discussion

The barriers and strategies were derived from dialogue discussions conducted with patients, providers, and healthcare leaders across the country. The four major themes that arose from these discussions were in relation to system and data readiness, the potential for bias and inequity, the regulation of AI and big data, and the importance of people as technology enablers. Participatory co-development and iterative design were the two most important strategies that were heard during the sessions, and span across all four themes. This study also yielded rich insights on priority applications, which have been published elsewhere.

Sittig and Singh’s sociotechnological model for studying health information technology in complex adaptive healthcare systems demonstrates the intricate challenges involved in the design, development, implementation, use, and evaluation of technology in healthcare settings [12]. It was designed to address the both the sociological and technological challenges that are present within healthcare settings that are high-pressured, fast-paced, and fragmented. Their model introduces eight interacting dimensions: hardware and software computing infrastructure, clinical content, human-computer interface, people, workflow and communication, internal organizational features, external rules and regulations, and measurement and monitoring [12]. This model was used as a comparison of our four themes and was chosen over other models such as the Technology Acceptance Model (TAM) and the Unified Theory of Acceptance and Use of Technology (UTAUT) for its applicability to complex health systems [10, 11]. The TAM and UTAUT aim to understand why users accept or reject a given technology and how acceptance can be improved, but recent reviews show that they have failed to predict acceptance of technologies in healthcare [31]. This may be because of healthcare’s unique cultural, social, and organizational factors that influence technology adoption. Therefore, Sittig and Singh’s model remained the most appropriate choice for our comparison due to its specificity to health information technology.

Interestingly, the four themes captured throughout this dialogue do align with most of the eight dimensions quite remarkably. Most notably, Sittig and Singh’s “people” dimension aligns with our fourth theme, the importance of people as technology enablers, as both recognize the importance of provider training and the user’s perception of the technology.

Additionally, our regulation of AI and big data theme correlates with Sittig and Singh’s “external rules and regulations” dimension since both are focused on the directives that must be designed at a government level to ensure all AI technology is held to a safe and reliable standard.

The system and data readiness theme does not align strongly with any one given dimension but correlates loosely with both the “hardware and software computing infrastructure” and “clinical content” dimensions. The former covers the technological perspective, including equipment and software, which could be extended to cover the contents of our theme including interoperability and digitization of medical records, while the latter addresses personal health information, data quality, and data structure.

The potential for bias and inequity theme is also not explicitly present in Sittig and Singh’s eight dimensions, possibly because AI has the potential to present biases in ways that technology was less capable of before. Furthermore, trust, the overarching foundation to all of our themes is also not present in the eight dimensions but was very important to the participants during the dialogue sessions.

Overall, Sittig and Singh’s sociotechnological framework is advantageous to consider both the social and technical perspectives of technology implementation projects, but based on our deliberative dialogue series, we believe that new dimensions (e.g., trust, bias, and equity) would be required to translate this model to be applicable to AI implementation projects in the primary care space now and in the future.As mentioned in the introduction, the recent review conducted by Lovejoy et al. divides the AI innovation process into three stages (invention, development, and implementation) and outlines key considerations for innovators during each phase [13]. The considerations of the third stage, implementation, include generalizability, regulation, and deployment. Generalizability refers to the need for diversity within the training set, regulation refers to the need for ongoing monitoring of an algorithm or tool throughout its lifetime to ensure continued safety, and deployment refers to the need to minimize disruption to workflow and account for interoperability between healthcare technologies [13]. When comparing the four themes uncovered in our study to the implementation stage of this review, it is evident that our deliberative dialogue results greatly support those of Lovejoy et al. For example, within our ‘potential for bias and inequity’ and ‘regulation of AI and big data’ themes, we heard multiple times in our sessions that diversity within training sets and continuous monitoring of a given tool are both vital to ensure bias and safety are maintained over time. This is very similar to Lovejoy et al.’s generalizability and regulation sections as they state, “The training data must be at least as diverse as the population that the algorithm intends to serve” and “Good performance at the time of deployment does not guarantee that the model will continue to perform well. This introduces the need to regulate throughout the lifetime of an algorithm, and the need to continually demonstrate safe and effective practices” [13]. Therefore, our ‘potential for bias and inequity’ theme is similar to Lovejoy et al.’s concept of generalizability, and our ‘regulation of AI and big data’ compliments the regulation section explained in their review.

The review by Singh et al. outlines AI implementation barriers in relation to the perspective of healthcare organizations, healthcare providers, and patients. These three perspectives are similar to those included in our study, except that instead of healthcare organizations we heard from health system leaders. From the review’s healthcare organization, they discussed the role of “company culture” as important when considering the use of embracing AI in their day-to-day practice–a notion that did not come up explicitly in our deliberative dialogue sessions. The provider perspective outlined concerns for the consideration of racial, ethnic, gender, and other sociodemographic characteristics within the algorithm. This was similar to what was heard in our ‘Potential for bias and inequity’ theme, except that we heard the concern from all three types of stakeholders and not just from the providers alone. The perspective of patients was centered on trust, including the concern for physician-patient confidentiality if AI were to be used in their care [14]. This compliments our finding of the importance of trust, however, in our sessions it was woven into all four themes and from all perspectives (patients, providers and health system leaders).

The strengths of this study include the use of a two-coder analysis approach, its transferability to non-Canadian settings with similar primary care models, and the use of informed interprofessional stakeholder engagement. First, for the analysis of the data, a traditional two-coder approach was employed to ensure trustworthiness and intercoder reliability. Second, based on the services and values that are offered by most primary care clinics [32], our findings are applicable to jurisdictions outside of Canada, confirming interjurisdictional transferability. Last, participants in these sessions were from many different backgrounds (patients, primary care providers, and system leaders), but all learned about AI technology before taking part, nurturing an informed discussion. The interprofessional design of these sessions enabled participants to refute and agree on discussion topics and fostered the ability to explore different perspectives of using AI in primary care.

To further interpret this study, it is essential to consider the limitations. It was completed by engaging participants from eight Canadian provinces but unfortunately did not include the perspectives from Canada’s territories or any self-identifying Indigenous people, both of whom may have unique perspectives regarding barriers to using AI in primary healthcare [33]. To further improve this study and its transferability, future work should focus on the perspectives of those that self-identify as indigenous as well as individuals form Canada’s territories. This would enable more diverse perspectives and would be vital before any AI implementation projects are considered. Additionally, due to the ongoing COVID-19 pandemic, we were only able to include five health system leaders in the study. The health leaders that were able to attend provided rich system-level perspectives that were unique to those heard from patients and providers, thus including more system leaders would have been preferred.

This level of study (48 participants from 8 Canadian provinces including 22 patients, 21 primary care providers, and 5 health system leaders) provides an excellent starting point for the barriers and strategies to use when implementing AI in primary care settings, but a study with larger reach would be recommended before practical AI applications become commonplace in primary care. To increase reach and gain further insight, a deliberative dialogue would not be practical but instead a survey could gain the perspectives of more of the population. This in addition to a focused dialogue with individuals from Canada’s territories and those who self-identify as indigenous would be vital before practical AI solutions are implemented widely in primary care.

Conclusions

Participants were hopeful for AI technology, but perceived barriers to its impact, most of which converged on trust in the technology, and thereby patient and provider willingness to adopt. These barriers were consolidated into four key themes and strategies to overcome these barriers were proposed, including most notably, participatory co-design and iterative implementation. These findings will act as a critical catalyst to address these barriers, as they highlight patient, provider and health system leaders concerns regarding implementation of AI in primary care settings.

Supporting information

S1 Checklist. COREQ.

(PDF)

S1 Table. Study participants.

(DOCX)

S2 Table. Iterative coding process.

(PDF)

S1 Appendix. Participant AI learning module.

(DOCX)

Acknowledgments

We are grateful to the study participants who made this research possible by sharing their time and rich perspectives. We thank Millie Upshaw, Sally Headrick, Monica Brands, and Jane Cooney for reviewing the online informational module used to support participants.

Data Availability

All relevant data are in the paper and its Supporting Information files.

Funding Statement

This project was supported by the Canadian Institutes for Health Research (FRN 156885), with ADP as Principal Investigator. TLU's time was supported by a CIHR Frederick Banting and Charles Best Canada Graduate Scholarship. ADP is supported by the Department of Family and Community Medicine, Temerty Faculty of Medicine at the University of Toronto and at St. Michael’s Hospital, and by the Li Ka Shing Knowledge Institute, Unity Health Toronto. He is also supported by a CIHR Applied Public Health Chair and a fellowship from the Physicians’ Services Incorporated Foundation. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Mayo R. C. and Leung J., “Artificial intelligence and deep learning–Radiology’s next frontier?,” Clinical Imaging, vol. 49. Elsevier Inc., pp. 87–88, May 01, 2018. doi: 10.1016/j.clinimag.2017.11.007 [DOI] [PubMed] [Google Scholar]
  • 2.Amisha P. Malik M Pathania, and Rathaur V, “Overview of artificial intelligence in medicine,” J Family Med Prim Care, vol. 8, no. 7, 2019, doi: 10.4103/jfmpc.jfmpc_440_19 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.American Academy of Family Physicians, “Primary Care—All Policies,” American Academy of Family Physicians, 2021. https://www.aafp.org/about/policies/all/primary-care.html (accessed Jun. 18, 2021). [Google Scholar]
  • 4.Mistry P., “Editorials artificial intelligence in primary care,” British Journal of General Practice, vol. 69, no. 686. Royal College of General Practitioners, pp. 422–423, 2019. doi: 10.3399/bjgp19X705137 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Soerensen P. D., Christensen H., Gray Worsoe Laursen S., Hardahl C., Brandslund I., and Madsen J. S., “Using artificial intelligence in a primary care setting to identify patients at risk for cancer: A risk prediction model based on routine laboratory tests,” Clin Chem Lab Med, 2021, doi: 10.1515/cclm-2021-1015 [DOI] [PubMed] [Google Scholar]
  • 6.Wang S., Pathak J., and Zhang Y., “Using electronic health records and machine learning to predict postpartum depression,” in Studies in Health Technology and Informatics, Aug. 2019, vol. 264, pp. 888–892. doi: 10.3233/SHTI190351 [DOI] [PubMed] [Google Scholar]
  • 7.Amit G. et al. , “Estimation of postpartum depression risk from electronic health records using machine learning,” BMC Pregnancy Childbirth, vol. 21, no. 1, Dec. 2021, doi: 10.1186/s12884-021-04087-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Black J. E., Kueper J. K., Terry A. L., and Lizotte D. J., “Development of a prognostic prediction model to estimate the risk of multiple chronic diseases: Constructing a copula-based model using Canadian primary care electronic medical record data,” Int J Popul Data Sci, vol. 6, no. 1, Jan. 2021, doi: 10.23889/ijpds.v5i1.1395 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Black J. E., Terry A. L., and Lizotte D. J., “Development and evaluation of an osteoarthritis risk model for integration into primary care health information technology,” Int J Med Inform, vol. 141, Sep. 2020, doi: 10.1016/j.ijmedinf.2020.104160 [DOI] [PubMed] [Google Scholar]
  • 10.Wallace L. G. and Sheetz S. D., “The adoption of software measures: A technology acceptance model (TAM) perspective,” Information and Management, vol. 51, no. 2, pp. 249–259, Mar. 2014, doi: 10.1016/j.im.2013.12.003 [DOI] [Google Scholar]
  • 11.Venkatesh V., Morris M. G., and Davis G. B., “User acceptance of information technology: toward a unified view,” MIS Quarterly, vol. 27, no. 3, pp. 425–478, 2003, 10.2307/30036540. [DOI] [Google Scholar]
  • 12.Sittig D. F. and Singh H., “A new sociotechnical model for studying health information technology in complex adaptive healthcare systems,” Qual Saf Health Care, vol. 19, no. 1, pp. 68–74, 2010, doi: 10.1136/qshc.2010.042085 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Lovejoy C. A., Arora A, Buch V, and Dayan I, “Key considerations for the use of artificial intelligence in healthcare and clinical research,” Future Healthc J, vol. 9, no. 1, pp. 75–78, Mar. 2022, doi: 10.7861/fhj.2021-0128 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Singh R. P., Hom G. L., Abramoff M. D., Campbell J. P., and Chiang M. F., “Current challenges and barriers to real-world artificial intelligence adoption for the healthcare system, provider, and the patient,” Transl Vis Sci Technol, vol. 9, no. 2, pp. 1–6, Jan. 2020, doi: 10.1167/tvst.9.2.45 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Martin D., Miller A. P., Quesnel-Vallée A., Caron N. R., Vissandjée B., and Marchildon G. P., “Canada’s universal health-care system: achieving its potential,” The Lancet, vol. 391, no. 10131. Lancet Publishing Group, pp. 1718–1735, Apr. 28, 2018. doi: 10.1016/S0140-6736(18)30181-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Glazier R. H., Klein-Geltink J., Kopp A., and Sibley L. M., “Capitation and enhanced fee-for-service models for primary care reform: A population-based evaluation,” CMAJ. Canadian Medical Association Journal, vol. 180, no. 11, May 2009, doi: 10.1503/cmaj.081316 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Ministry of Health and Long Term Care Ontario and L.-T.Care of Ontario, “Ontario Health Teams: Guidance for Health Care Providers and Organizations.”
  • 18.McCoy M. L. and Scully P. L., “Deliberative Dialogue to Expand Civic Engagement: What Kind of Talk Does Democracy Need?,” National Civic Review2, vol. 91, no. 2, pp. 117–135, 2002. [Google Scholar]
  • 19.Mulvale G., Chodos H., Bartram M., MacKinnon M. P., and Abud M., “Engaging civil society through deliberative dialogue to create the first Mental Health Strategy for Canada: Changing Directions, Changing Lives,” Soc Sci Med, vol. 123, pp. 262–268, Dec. 2014, doi: 10.1016/j.socscimed.2014.07.029 [DOI] [PubMed] [Google Scholar]
  • 20.Ridde V. and Dagenais C., “What we have learnt (so far) about deliberative dialogue for evidence-based policymaking in West Africa,” BMJ Glob Health, vol. 2, no. 4, p. e000432, Dec. 2017, doi: 10.1136/bmjgh-2017-000432 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Lavis J. N., Boyko J., Oxman A. D., Lewin S., and Fretheim A., “SUPPORT Tools for evidence-informed health Policymaking (STP) 14: Organising and using policy dialogues to support evidence-informed policymaking,” Health Res Policy Syst, vol. 7, no. S14, 2009, doi: 10.1186/1478-4505-7-S1-S14 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Boyko J. A., Lavis J. N., Abelson J., Dobbins M., and Carter N., “Deliberative dialogues as a mechanism for knowledge translation and exchange in health systems decision-making,” Soc Sci Med, vol. 75, no. 11, pp. 1938–1945, 2012, doi: 10.1016/j.socscimed.2012.06.016 [DOI] [PubMed] [Google Scholar]
  • 23.Bodenheimer T. and Sinsky C., “From triple to Quadruple Aim: Care of the patient requires care of the provider,” Ann Fam Med, vol. 12, no. 6, pp. 573–576, Nov. 2014, doi: 10.1370/afm.1713 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Patton M., “Purposeful Sampling,” in Qualitative Research& Evaluation Methods, 3rded., Laughton C, D Axelsen, and Peterson K, Eds. Thousand Oaks, California: Sage Publications Ltd., 2001, pp. 230–242. [Google Scholar]
  • 25.Thorne S., SR K., and M.-E J, “Interpretive description: a noncategorical qualitative alternative for developing nursing knowledge,” Res Nurs Health, vol. 20, no. 2, pp. 169–177, 1997. doi: [DOI] [PubMed] [Google Scholar]
  • 26.Thorne S., Kirkham S., and O’Flynn-Magee K., “The Analytic Challenge in Interpretive Description,” Int J Qual Methods, vol. 3, no. 1, 2004, [Online]. Available: http://www.ualberta.ca/~iiqm/backissues/3_1/ pdf/thorneetal.pdf [Google Scholar]
  • 27.Plamondon K. M., Bottorff J. L., and Cole D. C., “Analyzing Data Generated Through Deliberative Dialogue,” Qual Health Res, vol. 25, no. 11, pp. 1529–1539, Nov. 2015, doi: 10.1177/1049732315581603 [DOI] [PubMed] [Google Scholar]
  • 28.Sangaramoorthy T. and Kroeger K. A. , Rapid Ethnographic Assessments: A Practical Approach and Toolkit for Collaborative Community Research. Routledge, 2020. [Google Scholar]
  • 29.Tobin G. A. and Begley C. M., “Methodological rigour within a qualitative framework,” J Adv Nurs, vol. 48, no. 4, pp. 388–396, Nov. 2004, doi: 10.1111/j.1365-2648.2004.03207.x [DOI] [PubMed] [Google Scholar]
  • 30.Upshaw TL, Craig-Neil A, Macklin J, Steele Gray C, Chan TCY, and Gibson JL, “Priorities for artificial intelligence in primary care: results of a national deliberative dialogue,” Journal of the American Board of Family Medicine, 2022. [DOI] [PubMed] [Google Scholar]
  • 31.Ammenwerth E., “Technology Acceptance Models in ealth nformatics: TAM and UTAUT,” Stud Health Technol Inform, vol. 263, pp. 64–71, 2019, doi: 10.3233/SHTI190111 [DOI] [PubMed] [Google Scholar]
  • 32.“Task Force 1. Report of the Task Force on Patient Expectations, Core Values, Reintegration, and the New Model of Family Medicine,” The Annals of Family Medicine, vol. 2, no. suppl_1, pp. S33–S50, Mar. 2004, doi: 10.1370/afm.134 [DOI] [Google Scholar]
  • 33.Walter M. et al. , “Indigenous Data Sovereignty in the Era of Big Data and Open Data,” Australian Journal of Social Issues, vol. 56, no. 2, Jun. 2021, doi: 10.1002/ajs4.141 [DOI] [Google Scholar]

Decision Letter 0

Ardashir Mohammadzadeh

6 Sep 2022

PONE-D-22-20397Implementing artificial intelligence in Canadian primary care: barriers and strategies identified through a national deliberative dialoguePLOS ONE

Dear Dr. Pinto,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Oct 21 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Ardashir Mohammadzadeh, Phd

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2.We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. 

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

Additional Editor Comments:

Revise the paper according to the comments of reviewers; add some comparisions; add some statements and references about the potential improvement by type-3 fuzzy logic systems;

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: N/A

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Overall the article is impressive and it highlights insights into practicalities of AI integration in primary care. It shows good practice by adhering to COREQ reporting guidelines for qualitative work, and it makes good use of quotes and descriptors to illustrate the themes. The work is novel and the conclusions will have an impact in the field. I would recommend accepting for publication but have a few suggestions for ways to add value to the current work:

1. There needs to be more examples of existing uses of AI in the introduction to put the current work in better context. In relation to this, any examples of AI application in primary care that were discussed in the dialogues should be expanded on if it would help explain the existing comments and quotes by participants.

2. I note with interest the sections mentioning the limitation of processing longitudinal clinical data for AI use, and think this is worth expanding on as a significant practical difficulty.

3. It would benefit your international audience to explain more about the health system in Canada as well as the terminology you have used to describe it e.g. what is a team-based care model?

4. What is the quadruple aim and support equity?

5. Please provide the participant information sheets handed out before the dialogues as supplemental data

6. Please list which advertisement platforms were used and acknowledge any patient groups that were used.

7. More details on sampling method would be beneficial

8. Line 227 "publics" should be "populations"

9. It would be useful to write provide a further description of sittig and sing's sociotechnological model which was used as a framework for analysis. In particular, why was it chosen in lieu of other prevailing models such as the unified theory of acceptance and use of technology (UTAUT), and technology acceptance model (TAM)?

10. If there are any coding tables available to visualise how the final themes were reached from the primary codes, these would be useful to put into supplementary data so that the iterative process can be demonstrated.

Reviewer #2: This is a well written article on an important topic. The main barrier to publication, in my view, is the weakness of the literature review.

There are other papers which have explored barriers to implementation, for example: https://doi.org/10.7861/fhj.2021-0128 and https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7443115/

This paper is good opportunity to highlight challenges raised by existing literature with empirical evidence.

The Introduction section is very concise and superficial at the moment. Please consider expanding it with more detailed overview of existing literature. The authors write: "There is a paucity of research that focuses specifically on AI implementation challenges and the strategies to overcome them within the primary care space." but they do not really mention any of the existing research in the field. There are only four references in the introduction and all of these are related to generic AI papers.

I think that with an expanded literature review in the discussion and an enhanced introduction, this paper would be a good addition to the body of literature on AI implementation science.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Ahmed Al-Naher

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2023 Feb 27;18(2):e0281733. doi: 10.1371/journal.pone.0281733.r002

Author response to Decision Letter 0


7 Nov 2022

The attached file, Response to Reviewers, contains a point-by-point response to the specific reviewer comments table format. We also note in this document that the general comments from the Editor have been addressed: this feedback was greatly appreciated, and is incorporated throughout the manuscript. Below we have included the text of our response, however we would suggest using the attachment to more clearly understand the changes.

--

Dear PLoS ONE Editorial Team,

We appreciate the time taken by our reviewers and the editorial team to review our manuscript entitled, Implementing artificial intelligence in Canadian primary care: barriers and strategies identified through a national deliberative dialogue. We appreciate the opportunity to resubmit this paper, and we have addressed each concern raised.

We look forward to hearing from you about this manuscript.

--

Reviewer #1:

1. There needs to be more examples of existing uses of AI in the introduction to put the current work in better context. In relation to this, any examples of AI application in primary care that were discussed in the dialogues should be expanded on if it would help explain the existing comments and quotes by participants. • Examples added: “Examples of risk prediction AI tools applicable to the primary care space include machine learning algorithms that identify primary care patients at risk for cancer (3), predict which patients may experience post-partum depression (4,5), and estimate those who may be at risk of multiple chronic diseases including diabetes, osteoarthritis, and hypertension (6,7). Although a flurry of research is focused on the development of AI tools, implementation into practice remains one of the most challenging barriers to overcome.”

• Applications discussed in the dialogue were theoretical cases to provide a basis for discussion.

• Also added more literature review on AI implementation challenges in health care:

“A recent review by Lovejoy et al. explores these challenges to health care as a whole by breaking down the AI innovation process into three stages – invention, development, and implementation. Specifically with respect to the implementation stage, they highlight three main areas of focus (generalizability, regulation, and deployment). Generalizability refers to the need for diversity within the training set, regulation refers to the need for ongoing monitoring of an algorithm or tool throughout its lifetime to ensure continued safety, and deployment refers to the need to minimize disruption to workflow and account for interoperability between health care technologies. Although this review is not specific to the primary care space, it provides insight into some of the barriers to AI implementation that exist within the health care space. Another review by Singh et al. outlines the barriers in relation to the perspective of health care organizations, health care providers, and patients. Interestingly, from the health care organization perspective, they discuss the role of “company culture” as important when considering the use of embracing AI in their day-to-day practice. The provider perspective outlined concerns for the consideration of racial, ethnic, gender, and other sociodemographic characteristics within the algorithm. Finally, the patients perspective was based on trust, and what physician-patient confidentiality would look like when AI is used in their care. Although these reviews are not specific to the primary care space, they provide insight into some of the barriers to AI implementation that exist within the health care space.”

2. I note with interest the sections mentioning the limitation of processing longitudinal clinical data for AI use, and think this is worth expanding on as a significant practical difficulty. • Just to clarify, the introduction notes that longitudinal clinical data is one of the reasons that is unique to primary care data, but not necessarily a limitation.

3. It would benefit your international audience to explain more about the health system in Canada as well as the terminology you have used to describe it e.g. what is a team-based care model? • Re-wrote the “Setting” section to be more comprehensive of the models used in Canada, and to explain that healthcare is provincially/territorially managed:

“This study was conducted virtually within Canada, a high-income country with provincially and territorial-based public health insurance that covers all necessary hospital and physician services (5). Both within and across provinces and territories of Canada, primary health care is delivered through different models. Largely, it is funded following a fee-for-service (FFS) model, but has adopted new funding models such as the enhanced fee-for-service model and capitation model. In Ontario, the enhanced FFS model is similar to the traditional FFS model, except that it includes additional targeted fee increases, extended hours, performance-based initiatives, and patient enrollment. In contrast, the capitation model reimburses primary care physicians for each person attributed to them, based on age and sex adjusted factors (6). More recently, Ontario has been forming Ontario Health Teams including family physicians, nurse practitioners, social workers, dietitians, and other health care professionals to increase interdisciplinary care for a given community (7).”

4. What is the quadruple aim and support equity? • Supporting equity put before quadruple aim so it did not sound like it was all one thing.

• Added definition within text to clarify: “optimizing health system performance through improved patient experience, improved clinical experience, better outcomes, and lower costs”

5. Please provide the participant information sheets handed out before the dialogues as supplemental data • AI learning module will be attached when revisions are re-uploaded. It is named “Supplemental Material – Participant AI Learning Module”

6. Please list which advertisement platforms were used and acknowledge any patient groups that were used. • Added this to the manuscript under Participants section:

“Specifically, patients were recruited using Kijiji and social media channels across 10 provinces as well as patient advisor distribution lists (patient advisory network in British Columbia, patient-family advisory council of downtown Toronto hospitals). Primary care physicians were recruited through social media channels, primary care research network news channels, and through email invitations from AP to practice colleagues and those in his network interested in digital health and health informatics.”

7. More details on sampling method would be beneficial • Added more detail for sampling method: “This recruitment method is outlined further by Patton et al (21), and was chosen in order to optimize variation in age, gender, race and ethnicity, education level, income, province of residence for patients and gender, race and ethnicity, provider type, country of health professions education, years in practice, practice size, and province of practice for primary care providers.”

8. Line 227 "publics" should be "populations" • Switched from “publics” to “populations”

9. It would be useful to write provide a further description of sittig and sing's sociotechnological model which was used as a framework for analysis. In particular, why was it chosen in lieu of other prevailing models such as the unified theory of acceptance and use of technology (UTAUT), and technology acceptance model (TAM)? • Revised intro of model: “Sittig and Singh’s sociotechnological model for studying health information technology in complex adaptive healthcare systems demonstrates the complex challenges involved in the design, development, implementation, use, and evaluation of technology in health care settings. It was designed to address the both the sociological and technological challenges that are present within health care settings that are high-pressured, fast-paced and distributed.”

• Added explanation: “This model was used as a comparison of our four themes and was chosen over other models such as the Technology Acceptance Model (TAM) and the Unified Theory of Acceptance and Use of Technology (UTAUT) for its applicability to health care. The TAM and UTAUT aim to understand why users accept or reject a given technology and how acceptance can be improved, but recent reviews show that they have failed to predict acceptance of technologies in health care (23). This may be because of health care’s unique cultural, social, and organizational factors that influence technology adoption. Therefore, Sittig and Singh’s model remained the most appropriate choice for our comparison due to its specificity to health information technology.”

10. If there are any coding tables available to visualise how the final themes were reached from the primary codes, these would be useful to put into supplementary data so that the iterative process can be demonstrated. • Coding table attached to re-submission

Reviewer #2:

This is a well written article on an important topic. The main barrier to publication, in my view, is the weakness of the literature review. There are other papers which have explored barriers to implementation, for example: https://doi.org/10.7861/fhj.2021-0128 and https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7443115/ This paper is good opportunity to highlight challenges raised by existing literature with empirical evidence. The Introduction section is very concise and superficial at the moment. Please consider expanding it with more detailed overview of existing literature. The authors write: "There is a paucity of research that focuses specifically on AI implementation challenges and the strategies to overcome them within the primary care space." but they do not really mention any of the existing research in the field. There are only four references in the introduction and all of these are related to generic AI papers. I think that with an expanded literature review in the discussion and an enhanced introduction, this paper would be a good addition to the body of literature on AI implementation science. • Added findings by one of the reccomended papers by Reviewer #2: “A recent review by Lovejoy et al. explores these challenges to health care as a whole by breaking down the AI innovation process into three stages – invention, development, and implementation. Specifically with respect to the implementation stage, they highlight three main areas of focus (generalizability, regulation, and deployment). Generalizability refers to the need for diversity within the training set, regulation refers to the need for ongoing monitoring of an algorithm or tool throughout its lifetime to ensure continued safety, and deployment refers to the need to minimize disruption to workflow and account for interoperability between health care technologies. Although this review is not specific to the primary care space, it provides insight into some of the barriers to AI implementation that exist within the health care space.”

• Also from the second recommended manuscript from reviewer #2: “Another review by Singh et al. outlines the barriers in relation to the perspective of health care organizations, health care providers, and patients. Interestingly, from the health care organization perspective, they discuss the role of “company culture” as important when considering the use of embracing AI in their day-to-day practice. The provider perspective outlined concerns for the consideration of racial, ethnic, gender, and other sociodemographic characteristics within the algorithm. Finally, the patients perspective was based on trust, and what physician-patient confidentiality would look like when AI is used in their care. Although these reviews are not specific to the primary care space, they provide insight into some of the barriers to AI implementation that exist within the health care space.”

Attachment

Submitted filename: AI for primary care - PLOS ONE - response to reviewers.docx

Decision Letter 1

Ardashir Mohammadzadeh

2 Dec 2022

PONE-D-22-20397R1Implementing artificial intelligence in Canadian primary care: barriers and strategies identified through a national deliberative dialoguePLOS ONE

Dear Dr. Pinto,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jan 16 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Ardashir Mohammadzadeh, Phd

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #2: (No Response)

Reviewer #3: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #2: N/A

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #2: Yes

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #2: My original review focussed on two deficiencies: the brevity of the Introduction and of the Discussion. The authors have adequately expanded the Introduction and I hope they agree that this has strengthened the manuscript. However, the lack of comparison to existing literature in the Discussion remains a weakness of the manuscript. The Discussion section of the manuscript remains sparse and there is only one other relevant paper discussed: "Sittig and Singh’s sociotechnological model for studying health information technology". It would be advisable to research previous studies using surveys and or other similar methodologies to identify barriers to limitations.

Reviewer #3: The paper is sufficiently innovative but needs some improvement as follows to be ready for publication in the journal:

1- The comparison with other suggested or conventional methods and their discussion is not well-highlighted in the introduction. It can helps to demonstrate the strength of the work and its weakness practically.

2-Neural network and fuzzy networks are very efficient intelligent methods. How can the offered method be improved by using fuzzy systems and neural networks? It is strongly recommended to include the following paper in the introduction refer and explain .

An interval type-3 fuzzy system and a new online fractional-order learning algorithm: theory and practice.

3-"Only five health system leaders were included in the study and no self-identifying Indigenous people". According to the mentioned limitation, what is the suitable practical solution applicable for the future works improvement?

4-Is this level of study of the mentioned community effective for practical implementations? Please explain?

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: No

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2023 Feb 27;18(2):e0281733. doi: 10.1371/journal.pone.0281733.r004

Author response to Decision Letter 1


20 Jan 2023

We appreciate the time taken by the reviewers and the editorial team to review our manuscript entitled, "Implementing artificial intelligence in Canadian primary care: barriers and strategies identified through a national deliberative dialogue." We appreciate the opportunity to resubmit this paper, and we have addressed each concern raised. We have uploaded a response letter and point-by-point response. The content of the latter is copied below, however we would suggest referring to the table for ease of interpretation.

[Note: there was no Reviewer #1 in the Decision letter.]

Reviewer #2:

My original review focussed on two deficiencies: the brevity of the Introduction and of the Discussion. The authors have adequately expanded the Introduction and I hope they agree that this has strengthened the manuscript. However, the lack of comparison to existing literature in the Discussion remains a weakness of the manuscript. The Discussion section of the manuscript remains sparse and there is only one other relevant paper discussed: "Sittig and Singh’s sociotechnological model for studying health information technology". It would be advisable to research previous studies using surveys and or other similar methodologies to identify barriers to limitations.

RESPONSE:

Thank you for the guidance on the introduction. We agree that the changes to the introduction have improved the paper greatly.

• Some changes to the discussion were made in the last round of edits (introducing the TAM and UTAUT models and why Sittig & Singh was selected over them)

• Changes to the discussion during this round of edits includes a more in-depth comparison to the Sittig & Singh model:

“The system and data readiness theme does not align strongly with any one given dimension but correlates loosely with both the “hardware and software computing infrastructure” and “clinical content” dimensions. The former covers the technological perspective, including equipment and software, which could be extended to cover the contents of our theme including interoperability and digitization of medical records, while the latter addresses personal health information, data quality, and data structure.

The potential for bias and inequity theme is also not explicitly present in Sittig and Singh’s eight dimensions, possibly because AI has the potential to present biases in ways that technology was less capable of before. Furthermore, trust, the overarching foundation to all of our themes is also not present in the eight dimensions but was very important to the participants during the dialogue sessions.

Overall, Sittig and Singh’s sociotechnological framework is advantageous to consider both the social and technical perspectives of technology implementation projects, but based on our deliberative dialogue series, we believe that new dimensions (e.g., trust, bias, and equity) would be required to translate this model to be applicable to AI implementation projects in the primary care space now and in the future.”

• Further research was completed to compare our finding to two other reviews on barriers to implementation:

“As mentioned in the introduction, the recent review conducted by Lovejoy et al. divides the AI innovation process into three stages (invention, development, and implementation) and outlines key considerations for innovators during each phase (13). The considerations of the third stage, implementation, include generalizability, regulation, and deployment. Generalizability refers to the need for diversity within the training set, regulation refers to the need for ongoing monitoring of an algorithm or tool throughout its lifetime to ensure continued safety, and deployment refers to the need to minimize disruption to workflow and account for interoperability between healthcare technologies (13). When comparing the four themes uncovered in our study to the implementation stage of this review, it is evident that our deliberative dialogue results greatly support those of Lovejoy et al. For example, within our ‘potential for bias and inequity’ and ‘regulation of AI and big data’ themes, we heard multiple times in our sessions that diversity within training sets and continuous monitoring of a given tool are both vital to ensure bias and safety are maintained over time. This is very similar to Lovejoy et al.’s generalizability and regulation sections as they state, “The training data must be at least as diverse as the population that the algorithm intends to serve” and “Good performance at the time of deployment does not guarantee that the model will continue to perform well. This introduces the need to regulate throughout the lifetime of an algorithm, and the need to continually demonstrate safe and effective practices” (13). Therefore, our ‘potential for bias and inequity’ theme is similar to Lovejoy et al.’s concept of generalizability, and our ‘regulation of AI and big data’ compliments the regulation section explained in their review.

The review by Singh et al. outlines AI implementation barriers in relation to the perspective of healthcare organizations, healthcare providers, and patients. These three perspectives are similar to those included in our study, except that instead of healthcare organizations we heard from health system leaders. From the review’s healthcare organization, they discussed the role of “company culture” as important when considering the use of embracing AI in their day-to-day practice – a notion that did not come up explicitly in our deliberative dialogue sessions. The provider perspective outlined concerns for the consideration of racial, ethnic, gender, and other sociodemographic characteristics within the algorithm. This was similar to what was heard in our ‘Potential for bias and inequity’ theme, except that we heard the concern from all three types of stakeholders and not just from the providers alone. The perspective of patients was centered on trust, including the concern for physician-patient confidentiality if AI were to be used in their care (14). This compliments our finding of the importance of trust, however, in our sessions it was woven into all four themes and from all perspectives (patients, providers and health system leaders).”

• Please see response to reviewer #3 below for additional changes made to the discussion section.

REVIEWER #3 - 1/4

The comparison with other suggested or conventional methods and their discussion is not well-highlighted in the introduction. It can helps to demonstrate the strength of the work and its weakness practically.

RESPONSE:

Introduced the common tech implementation models (TAM and UTUAT). Also introduced the Sittig and Singh model in the introduction. These three models are referred to in the discussion so we agree, it is important to introduce them first in the introduction.

“To identify and ease these barriers, researchers have developed models related to technology implementation including the ‘Technology Acceptance Model’ (TAM) and the ‘Unified Theory of Acceptance and the Use of Technology’ (UTAUT) model (10,11). These are well studied models that explore the reasons that users accept or reject a given technology and how acceptance can be improved, but unfortunately are not specific to primary healthcare or even to the healthcare at all. Another model, Sittig and Singh’s sociotechnological model for studying health information technology includes eight dimensions (hardware and software computing infrastructure, clinical content, human-computer interface, people, workflow and communication, internal organizational features, external rules and regulations, and measurement and monitoring (12). Sittig and Singh’s model focuses specifically on implementation of technology in complex adaptive healthcare systems and will be used as a basis of comparison for the findings of this study.”

Additionally, the mention of the two reviews in the introduction (Lovejoy et al. and Singh et al.) were better highlighted and explained in relation to this paper.

“Other researchers including Lovejoy et al. and Singh et al. have conducted reviews to explore the challenges of implementing AI in healthcare settings. The recent review by Lovejoy et al. explores the challenges of implementing AI to healthcare by breaking down the AI innovation process into three stages – invention, development, and implementation. Specifically with respect to the implementation stage, they highlight three main barriers of focus (generalizability, regulation, and deployment). Generalizability refers to the need for diversity within the training set, regulation refers to the need for ongoing monitoring of an algorithm or tool throughout its lifetime to ensure continued safety, and deployment refers to the need to minimize disruption to workflow and account for interoperability between healthcare technologies (13). Although this review is not specific to the primary care space, it provides insights into barriers that are commonly encountered in the healthcare AI space, all of which will help inform our study. Another review by Singh et al. outlines the barriers in relation to the perspective of healthcare organizations, healthcare providers, and patients. Interestingly, from the healthcare organization perspective, they discuss the role of “company culture” as important when considering the use of embracing AI in their day-to-day practice. The provider perspective outlined concerns for the consideration of racial, ethnic, gender, and other sociodemographic characteristics within the algorithm. The perspective of patients was centered on trust, including the concern for physician-patient confidentiality if AI were to be used in their care (14). This review provides vital insight from various stakeholder perspectives and emphasizes the importance of doing so.”

Reviewer #3 - 2/4

Neural network and fuzzy networks are very efficient intelligent methods. How can the offered method be improved by using fuzzy systems and neural networks? It is strongly recommended to include the following paper in the introduction refer and explain.

An interval type-3 fuzzy system and a new online fractional-order learning algorithm: theory and practice.

RESPONSE:

We appreciate this recommendation. We have given it careful consideration, and believe the suggested citation and related recommendation for revision are beyond the scope of the work presented in this manuscript.

Reviewer #3 - 3/4

"Only five health system leaders were included in the study and no self-identifying Indigenous people". According to the mentioned limitation, what is the suitable practical solution applicable for the future works improvement?

RESPONSE:

Added a practical solution for future work:

“To further improve this study and its transferability, future work should focus on the perspectives of those that self-identify as indigenous as well as individuals form Canada’s territories. This would enable more diverse perspectives and would be vital before any AI implementation projects are considered.”

Reviewer #3 - 4/4

Is this level of study of the mentioned community effective for practical implementations? Please explain?

RESPONSE:

Expanded on this as follows:

“This level of study (48 participants from 8 Canadian provinces including 22 patients, 21 primary care providers, and 5 health system leaders) provides an excellent starting point for the barriers and strategies to use when implementing AI in primary care settings, but a study with larger reach would be recommended before practical AI applications become commonplace in primary care. To increase reach and gain further insight, a deliberative dialogue would not be practical but instead a survey could gain the perspectives of more of the population. This in addition to a focused dialogue with individuals from Canada’s territories and those who self-identify as indigenous would be vital before practical AI solutions are implemented widely in primary care.”

Attachment

Submitted filename: Response-to-Reviewers.docx

Decision Letter 2

Ardashir Mohammadzadeh

31 Jan 2023

Implementing artificial intelligence in Canadian primary care: barriers and strategies identified through a national deliberative dialogue

PONE-D-22-20397R2

Dear Dr. Pinto,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Ardashir Mohammadzadeh, Phd

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #2: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #2: Yes

Reviewer #3: Partly

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #2: N/A

Reviewer #3: I Don't Know

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #2: Yes

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #2: No further concerns. Comments from previous reviews have been answered. Manuscript appears much improved from original submission.

Reviewer #3: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: No

Reviewer #3: No

**********

Acceptance letter

Ardashir Mohammadzadeh

17 Feb 2023

PONE-D-22-20397R2

Implementing artificial intelligence in Canadian primary care: barriers and strategies identified through a national deliberative dialogue

Dear Dr. Pinto:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Ardashir Mohammadzadeh

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Checklist. COREQ.

    (PDF)

    S1 Table. Study participants.

    (DOCX)

    S2 Table. Iterative coding process.

    (PDF)

    S1 Appendix. Participant AI learning module.

    (DOCX)

    Attachment

    Submitted filename: AI for primary care - PLOS ONE - response to reviewers.docx

    Attachment

    Submitted filename: Response-to-Reviewers.docx

    Data Availability Statement

    All relevant data are in the paper and its Supporting Information files.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES