Skip to main content
The British Journal of General Practice logoLink to The British Journal of General Practice
editorial
. 2019 Aug 30;69(686):422–423. doi: 10.3399/bjgp19X705137

Artificial intelligence in primary care

Pritesh Mistry 1
PMCID: PMC6715470  PMID: 31467001

INTRODUCTION

Artificial intelligence (AI) in medicine captures the imagination with concepts such as the digital doctor, but also raises concerns such as replacing healthcare professionals, undermining trust in clinicians, and exacerbating inequalities. There is also continued uncertainty as to exactly what AI is. Given this confusion about definition, purpose, and potential, the RCGP has published an introductory report aimed at clarifying the position,1 in which a broad definition of AI as software with decision making capacity is used, ranging from software sequentially going through a series of yes/no questions (decision trees and algorithms) at the simplest, to complex software that learns from data (machine learning) where millions of datasets are used, for example, to find a pattern linking a set of symptoms and a disease.

Often, development of AI tools for place-based care and self-care is predominately driven by a focus on the technology and a commercial need to find a market, rather than the challenges and needs of healthcare professionals, practice and community providers, and patients. Unlocking the potential for AI means we must have healthcare professionals, patients, and technology experts working together and engaging with policy makers and commissioners.

USE OF AI FOR CLINICAL DECISION-MAKING AND PROACTIVE DETECTION

The development and provision of AI continues to progress rapidly. The Academic Health Science Network (AHSN), established by NHS England to spread innovation, released the results of a recent AI mapping survey,2 providing an overview of AI healthcare activity. While not exhaustive (organisations including Babylon and Livi are missing), the survey provides some useful insights. Primary and community care AI clusters around two areas. First, clinical decision making and care management, for example, symptom assessment, automating clinical coding, image recognition for dermatological conditions, triaging, and personalised self-management. Second, proactive detection, such as, analysing patient records to predict patients with undiagnosed conditions such as dementia or belonging to vulnerable groups such as frail older people. Surprisingly, less than 10% of the survey responders are applying AI to reduce or automate administrative tasks.

FRAMEWORKS SET THE FUTURE

While there is both enthusiasm and scepticism about AI (sometimes expressed by the same individuals) we need to separate the hype from the reality — the use of AI will depend upon infrastructure, operating models, and data. ‘The future of healthcare: our vision for digital, data and technology in health and care’ 3 is a policy paper published in October 2018 outlining a much needed approach to modernising healthcare computing systems, addressing current limitations in both infrastructure and operating models.

Data, on the other hand, can be a nebulous concept that continues to be a challenge. General practice records hold patient and practice information (or data) in various formats (codes, free text, letters) which can all be used to develop AI. However, biased, skewed, unrepresentative, and limited quantities of data affect AI functionality, potentially creating software with a higher likelihood of being erroneous and unsafe. Recent publications4,5 shine a light on the use and limitations of general practice data for narrow use cases: data need validating for each new application, which can be resource intensive. The use of patient data needs to be ethical from the perspective of patients, healthcare professionals, and the system,6 respecting choices on privacy, utilisation, access, and monetisation. Furthermore, these need to be mutually agreed, clearly communicated, and applied. Two recent publications aim to do just this; ‘Data Ethics Framework’,7 complemented and built upon by ‘Initial code of conduct for data-driven health and care technology‘.8 Together these set out the expectation the public sector holds on data use and transparency, providing reassurance when met. However, there are additional criteria related to the values held by primary care healthcare professionals which, if not met, could limit AI use. Continuity, evidence-based, overmedicalisation, safeguarding, anxiety induced demand, and management of uncertainty, are of high importance and could impede AI use if they are not incorporated and explained adequately.

‘The future of healthcare’,3 ‘Data Ethics Framework’,7 and ‘Code of Conduct’,8 set clear expectations, but it will take time to develop the guidance and mechanisms so that they are enforced and upheld, balancing self-policing and regulation of a technology that is mistrusted. In the meantime, the AI landscape risks being shaped by early established companies and decisions made with insufficient evaluations in place due to pressures to embrace technology. AI improves as more data becomes available; AI training is cumulative. It is easy to imagine the positive feedback loop: more data = more accurate AI = better product = more users = more data. Existing AI applications potentially have an advantage over newcomers; the gap widens with time producing monopolies and stifling competition.

Currently, AI needs to be developed for each use case (narrow AI) and retrained for different application areas, limiting lock-in and monopolies. However, research is progressing to make AI transferable across applications.9

THE NEW INTERFACE: CO-OPERATIVE AI AND THE HEALTHCARE PROFESSIONAL

The concept of the professional working alongside a supportive system is not new. In manufacturing industries, attempts to replace people with robots and automation failed, and it was realised that robots have strength and are tireless, but they do not have the dexterity, agility, or flexibility of humans. This gave rise to collaborative robots, or cobots, a revised approach pairing the advantages of humans with the benefits of robotics, for example, automotive manufacturing cobots do highly repetitive tasks like screwdriving while human dexterity guides and facilitates assembly. Similarly, AI has advantages over humans in some areas while being much less capable in others — the true benefit is unlocked through combination working. Combined AI/human working augments the leadership, teamwork, creativity, social skills, observational skills, and empathy of healthcare professionals (or person accessing care), with the speed, scalability, information access, and quantitative capabilities of software. Research on 1500 organisations across sectors showed significant improvements from AI/human working.10 AI tools that address unmet needs of patients and practice and community providers (for example, early detection of deterioration, automating communications, simplifying multi-login processes) will need a co-development approach with developers, healthcare professionals, provider organisations, patients, commissioners, and system level stakeholders as equal partners. Co-development is essential for AI to meet expectations and have acceptable limitations otherwise the bounds of function are the opinions of the developers, relying on quantitative data without qualitative data (preferences, context) with the risk of not necessarily being right.

AI has the potential to amplify human strengths, such as leadership and creativity, by providing relevant evidence-based information at the right time.10 Co-development leads to ‘relevant’ and ‘right time’ being defined in a way that meets the needs of healthcare professionals and of patients. AI can improve flexibility by creating chatbots which enable asynchronous one-to-many correspondences. With co-development, AI functionality and workflows embed the correct safeguards, have appropriate red flag functions, provide continuity, have acceptable clinical safety, avoids increasing anxiety, and have complementary processes supporting healthcare professionals. In an environment with insufficient numbers of healthcare professionals, AI can have a multiplying effect on available resources, such as automating identification of undiagnosed conditions or use digital systems to enable one-to-many remote monitoring. Through co-development we can achieve appropriate clinical oversight, safeguards, and management of risk.

To enable utilisation protocols, processes, workflows, and pathways will need altering to incorporate collaborative AI tools:

  • healthcare professionals must have access to education to learn new skills as AI users and work differently;

  • care providers using AI tools will need to be able to train AI systems to set them up to perform specified tasks. This means knowing what data to select and its quality;

  • clinicians will need to be able to interpret AI outputs, which requires an understanding of its limitations and bounds of function; and

  • users will need to know how the system learns and what constitutes appropriate use, so that ethical norms are upheld and any introduction of biases is avoided.

CONCLUSION

In future, some of this functionality will be built in, for example, self-selecting data to uphold original design principles, and commercial providers could offer to train an AI tool to a user’s specification. AI needs to be developed collaboratively with healthcare professionals and patients at all stages from idea to utilisation. This will create trusted solutions to real problems for patient benefit. It will support acceptance by ensuring expectations on evidence, ethics, and the values of healthcare professionals are incorporated in evaluation methodology and regulation. Lastly it will support adoption by identifying the training required to change provider protocols, business processes, administrative workflows, and pathways. If developed collaboratively, AI can bring together the advantages of people and computers to minimise potential negative impact and to maximise health gain.

Provenance

Commissioned; externally peer reviewed.

Competing interests

The author has declared no competing interests

REFERENCES


Articles from The British Journal of General Practice are provided here courtesy of Royal College of General Practitioners

RESOURCES