Abstract
Machine learning methods, combined with large electronic health databases, could enable a personalised approach to medicine through improved diagnosis and prediction of individual responses to therapies. If successful, this strategy would represent a revolution in clinical research and practice. However, although the vision of individually tailored medicine is alluring, there is a need to distinguish genuine potential from hype. We argue that the goal of personalised medical care faces serious challenges, many of which cannot be addressed through algorithmic complexity, and call for collaboration between traditional methodologists and experts in medical machine learning to avoid extensive research waste.
Introduction
Proponents of precision medicine make a compelling pitch: traditional approaches to health science have focused too much on comparing effectiveness in the average person and too little on the needs of actual individuals.1,2 The blame could lie with outdated statistical and epidemiological tools, which might offer decreasing relevance to the needs of contemporary clinical decision making.3 The proposed solution speaks to the zeitgeist: our newfound abundance of detailed and accessible longitudinal data on individuals combined with the practical realisation of various flexible machine learning approaches offer an exciting chance for revolution.2 At the apex sits the dream of precision medicine crafted by machine learning, a new framework that promises to revolutionise how we identify the best therapy for each person as an individual, while automating everyday tasks like diagnosis and prognostication with unprecedented accuracy.4
But how realistic are these claims? And when, if ever, can we expect them to be routinely realised? We consider the evidence underlying two of the most common claims about the potential of machine learning-powered precision medicine and call for a reality check of expectations.
Claim 1: machine learning will enable automated diagnoses with unprecedented accuracy
Machine learning is often heralded by health and medical commentators as a powerful prediction tool that will revolutionise disease screening and diagnosis. The inherent flexibility and scope for automation makes machine learning well suited to examining complex high-dimensional data (ie, with many variables or features) that would be challenging to model using conventional approaches. Such strategies have enabled the development of several innovative diagnostic algorithms—for example, to identify patients most in need of intervention from knee MRI,5 to detect cardiac arrhythmias from electrocardiograms,6 and to diagnose pneumonia from chest x-rays.7
Given such innovation, it is hard to dispute the revolutionary potential of machine learning for improving clinical diagnostics. However, acknowledging potential is a poor substitute for robust scientific evidence of actual benefit, and here research is lacking.8 Although news media is filled with enthusiastic stories about novel machine learning applications,9–12 a systematic review comparing the performance of deep learning versus health professional assessment in diagnosis of various diseases from medical images makes for sobering reading.13 Only 20 (24%) of the 82 studies identified evaluated the performance of their algorithm in an external cohort, and only 14 (17%) studies compared this out-of-sample performance with that of health professionals. This number is alarmingly small, especially given that many of the studies were flawed. The authors found that reporting standards were typically poor, internal validation was weak and, perhaps most worryingly, model performance was often evaluated under unrealistic conditions that had little relevance to routine clinical practice.13 For example, there is little use comparing the performance of a machine learning algorithm with health professional judgment for making diagnoses from medical images without providing further contextual information about the patient, as this would never happen in practice.14 A more recent systematic review of studies comparing deep learning with clinical judgement corroborated the widespread issues with poor study design and reporting, but identified some well designed randomised clinical trials that evaluated the technology.15
Emphasis on predictive performance over clinical utility is not unusual. The ability of machine learning to process high-dimensional data, for example, appears to be distracting from the often greater benefits of simple clinical variables. Volkmann and colleagues16 showed how predictions based on large dimensional omics data can easily be improved by including more common clinical information. Indeed, the added value of omics data for clinical prediction can be marginal once all relevant clinical variables are included. There is also an increasing focus on classifying patients into simple categories (eg, with and without the disease) rather than predicting a continuum of risk. This trend is an unfortunate departure from the supposed aim of increasing individual relevance and is particularly puzzling because it is not a requirement of most machine learning methods.
In terms of using machine learning to automate diagnoses, this remains more of a potential promise than a proven product, although notable exceptions exist, such as an automated system for detection of diabetic retinopathy.17 At present, the benefits of most proposals to automate clinical diagnoses with machine learning are unknown because they have not been meaningfully assessed. Novel machine learning studies are not unusual in this regard. Clinical journals are flooded with traditional prognostic models that were neither developed nor evaluated using appropriate methods.18 However, since few of these models ever end up being used, they are arguably benign. By contrast, there is tangible concern that the scale of enthusiasm around machine learning means substandard models might get unduly adopted by a clinical audience that is not equipped to assess them.
Regardless, the performance and utility of a machine learning algorithm is highly dependent on the quality and relevance of the data on which it is trained. Training in image recognition requires large numbers of images to be scrutinised and annotated by human experts, a burdensome task that itself carries a risk of error, highlighting the need for methods to enable medical experts to create high-quality annotations at scale. Furthermore, algorithms often perform badly and require retraining when introduced into environments that were not represented in their training data, as highlighted by the poor performance of a Google algorithm for detecting diabetic retinopathy from retinal images when deployed in poorly-lit eye clinics in Thailand.19
Claim 2: machine learning-powered precision medicine will enable identification of the best therapies for individuals rather than groups
The potential to revolutionise the individual tailoring of medical treatments is one of the most widely discussed and appealing promises of machine learning-powered precision medicine.20 Unfortunately, this promise is probably also the least likely to ever be fully achieved, for two fundamental epistemological reasons.
First, machine learning approaches are not (currently) able to identify cause and effect, because causal inference is fundamentally impossible to achieve without making assumptions.21,22 Causal inference requires the following three central assumptions: a clearly specified causal question, that the causal effect of interest can be identified, and that all treatment options of interest can be observed among all groups of interest. Several of these assumptions can be satisfied through study design or external contextual knowledge (ie, human input or true artificial intelligence), but none can be discovered solely from observational data. Since causal inference is necessary to find out what works and by how much, this issue poses a considerable barrier to the promise of ever being able to identify the best treatment for an individual person by machine learning.
This problem does not simply reflect the adage that correlation does not imply causation, nor the widely held belief that causal inference can only be achieved in experimental data. A suite of methodological approaches are available to aid estimation of causal effects in non-experimental data,23,24 such as electronic health records, which potentially offer more diverse and relevant samples than clinical trials (although representative samples might not be necessary to achieve representative results25). The limitation arises because almost all machine learning algorithms have been designed to make predictions (eg, to most accurately predict or classify those with a particular trait or prognosis) and this is fundamentally distinct from causal explanation and causal effect estimation.26,27
To estimate a causal effect, we instead need to estimate not just what is most likely to happen (ie, prediction) but what would most likely have happened if things had been different (ie, counterfactual prediction27). Accordingly, the prospect of automated causal inference in observational data remains beyond reach. Instead, causal inference requires external contextual information about the meaning of and relationships between the relevant variables in a given context.28 Machine learning cannot learn these things from a dataset because they are not necessarily there to learn—the data only include what has happened, not what might have happened if things had been different. Hence, data-driven prediction models cannot help us to determine the effect of different exposures and treatments on different outcomes, and in turn cannot deliver the promise of identifying the best treatment approach for specific individuals.28 Simply increasing the sample size (ie, collecting big data) or increasing algorithmic complexity does not help to resolve this fundamental epistemological mismatch of aims. Further tensions arise between adopting a data-driven approach and the preference towards pre-specification of analyses before data collection in clinical trials.
However, that machine learning cannot yet—if ever—conduct causal inference is arguably less of a barrier to achieving individual-level predictions than the second more fundamental problem that the majority of health states and events are so complex that we can only understand them probabilistically, and chance can never be predicted at the individual level. Indeed, although statistics—and hence machine learning—is excellent at helping us to understand and compare probabilities between groups, it is fundamentally unable to tell us what will happen to an individual. The power of statistics is precisely that it can describe and predict partly random events over large numbers of people. But no matter how accurately statistics can do this, it can never tell us with certainty what the next event will be.
For example, consider rolling two (fair) dice and counting the total score. Although we know what is most likely (seven), and we know with 100% certainty the probability of all other scores (eg, 1 in 32 for a double six), we are still no closer to knowing what the next roll will actually be. Likewise, even the most sophisticated causal inference methods cannot determine with certainty the effect of an individual variable in an individual person (eg, the effect of systolic blood pressure on the risk of stroke). This issue is not resolvable by simply collecting more data or building increasingly complex models, because it stems from limitations of our understanding of physical and biological processes. Since most health processes are effectively probabilistic, we must accept that individual outcomes will always be subject to chance, no matter how precisely we can describe these for groups of similar patients.
Pragmatic routes to personalisation
The scale and substance of these barriers leads to a larger question: is precision medicine itself an epistemological dead end? To answer, we must step back from the headline promises and revisit the core principles. The first requirement for precision medicine is that treatments have different effects in different people and the second is that this variation can, at least partly, be characterised and predicted.29 Where the cause of such variation can be easily identified and targeted, such as with a disease that has a dominant genetic component, this is potentially realistic; indeed this is exactly where precision medicine has found its success.30 However, in most cases, characterising and predicting treatment responses is undermined for the very reasons we seek to use precision medicine—health is complex—and understanding the determinants of variation in response is a formidable challenge. Indeed, for most health states and events, we remain unable to achieve the ostensibly much easier task of improving the average response in a group.
One of the biggest obstacles to precision medicine comes from our poor collective understanding of the nature of variation and the types of study needed to reveal it. Disease symptoms and the apparent treatment response can vary substantially within the same individual over time.29 Traditional clinical studies are woefully ill-equipped to identify who will respond consistently well, or indeed whether anyone will respond consistently at all. Even with more intensive study designs, such as repeated crossover studies or N-of-1 trials, it might not be possible to differentiate within-individual and between-individual variation without strong assumptions that are only plausible in specific circumstances (eg, where the symptoms and participant circumstances are stable over time and the treatments have no long-term effects).29,31
A more pragmatic route towards greater personalisation might be to shift the aim towards stratified medicine (ie, identifying and predicting subgroups with a better and worse response). Although somewhat more modest in ambition, this strategy would be compatible with our existing statistical epistemology and could still provide a genuine revolution. However, this route remains much more complex than simply applying machine learning approaches to experimental or observational datasets, because of the fundamental challenge of differentiating true signal from noise. For example, response variation is typically explored by simple subgroup analyses of specific predefined variables (eg, sex, age, race, or ethnicity). These analyses are usually exploratory, because even if there are sound expectations of treatment variation, experiments tend to be restricted to focus on the responsive group. However, doing multiple statistical tests for modest effects in relatively small samples is a recipe for disaster, including mistaking noise for signal, exaggerating trivial effects, and (ironically) overlooking true effects.32–34 Anything that is identified of course needs validation in subsequent experiments,35 which are rare and seldom confirm the initial suspicions.33,36
Conclusion
Both machine learning and precision medicine are genuine innovations and will undoubtedly lead to some great scientific successes. However, these benefits currently fall short of the hype and expectation that has grown around them. Such a disconnect is not benign and risks overlooking rigour for rhetoric and inflating a bubble of hope that could irretrievably damage public trust when it bursts. Such mistakes and harm are inevitable if machine learning is mistakenly thought to bypass the need for genuine scientific expertise and scrutiny. There is no question that the appearance of big data and machine learning offer an exciting chance for revolution, but revolutions demand greater scrutiny, not less. This scrutiny should involve a reality check on the promises of machine learning-powered precision medicine and an enhanced focus on the core principles of good data science—trained experts in study design, data system design, and causal inference asking clear and important questions using high-quality data.
Acknowledgments
JW is supported by a Wellcome Institutional Strategic Support Fund award (204796/Z/16/Z). MSG and PWGT are supported by The Alan Turing Institute (EP/N510129/1). CL is supported by the German Federal Ministry of Education and Research in the project KI-LAB-ITSE (project number 01|S19066). These funders had no role in any aspect of the conception or realisation of the manuscript. JW had final responsibility for the decision to submit for publication.
Footnotes
Declaration of interests
We declare no competing interests.
Contributor Information
Jack Wilkinson, Centre for Biostatistics, Manchester Academic Health Science Centre, Division of Population Health, Health Services Research and Primary Care, University of Manchester, Manchester, UK.
Kellyn F Arnold, Leeds Institute for Data Analytics, University of Leeds, Leeds, UK; Faculty of Medicine and Health, University of Leeds, Leeds, UK.
Eleanor J Murray, Department of Epidemiology, Boston University School of Public Health, Boston, MA, USA.
Maarten van Smeden, Department of Clinical Epidemiology, Leiden University Medical Center, Leiden, Netherlands.
Kareem Carr, Department of Biostatistics, Harvard TH Chan School of Public Health, Boston, MA, USA.
Rachel Sippy, Institute for Global Health and Translational Science, SUNY Upstate Medical University, Syracuse, NY, USA; Department of Geography, University of Florida, Gainesville, FL, USA; Emerging Pathogens Institute, University of Florida, Gainesville, FL, USA.
Marc de Kamps, Leeds Institute for Data Analytics, University of Leeds, Leeds, UK; School of Computing, University of Leeds, Leeds, UK.
Andrew Beam, Department of Epidemiology, Boston University School of Public Health, Boston, MA, USA.
Stefan Konigorski, Digital Health & Machine Learning Research Group, Hasso Plattner Institut for Digital Engineering, Potsdam, Germany; Hasso Plattner Institute for Digital Health at Mount Sinai, Icahn School of Medicine at Mount Sinai, New York, NY, USA.
Christoph Lippert, Digital Health & Machine Learning Research Group, Hasso Plattner Institut for Digital Engineering, Potsdam, Germany; Hasso Plattner Institute for Digital Health at Mount Sinai, Icahn School of Medicine at Mount Sinai, New York, NY, USA.
Mark S Gilthorpe, Leeds Institute for Data Analytics, University of Leeds, Leeds, UK; Faculty of Medicine and Health, University of Leeds, Leeds, UK; Alan Turing Institute, London, UK.
Peter W G Tennant, Leeds Institute for Data Analytics, University of Leeds, Leeds, UK; Faculty of Medicine and Health, University of Leeds, Leeds, UK; Alan Turing Institute, London, UK.
References
- 1.Macklon NS, Ahuja KK, Fauser B. Building an evidence base for IVF ‘add-ons’. Reprod Biomed Online 2019; 38: 853–56. [DOI] [PubMed] [Google Scholar]
- 2.Matheny ME, Whicher D, Thadaney Israni S. Artificial intelligence in health care: a report from the National Academy of Medicine. JAMA 2019; 323: 509–10. [DOI] [PubMed] [Google Scholar]
- 3.Gombar S, Callahan A, Califf R, Harrington R, Shah NH. It is time to learn from patients like mine. NPJ Digit Med 2019; 2: 16. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.The Lancet Public Health. Next generation public health: towards precision and fairness. Lancet Public Health 2019; 4: e209. [DOI] [PubMed] [Google Scholar]
- 5.Bien N, Rajpurkar P, Ball RL, et al. Deep-learning-assisted diagnosis for knee magnetic resonance imaging: development and retrospective validation of MRNet. PLoS Med 2018; 15: e1002699. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Hannun AY, Rajpurkar P, Haghpanahi M, et al. Cardiologist-level arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network. Nat Med 2019; 25: 65–69. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Rajpurkar P, Irvin J, Ball RL, et al. Deep learning for chest radiograph diagnosis: a retrospective comparison of the CheXNeXt algorithm to practicing radiologists. PLoS Med 2018; 15: e1002686. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Vollmer S, Mateen BA, Bohner G, et al. Machine learning and artificial intelligence research for patient benefit: 20 critical questions on transparency, replicability, ethics, and effectiveness. BMJ 2020; 368: l6927. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.ScienceDaily. AI system accurately detects key findings in chest X-rays of pneumonia patients within 10 seconds. Sept 30, 2019. https://www.sciencedaily.com/releases/2019/09/190930104505.htm (accessed July 27, 2020).
- 10.MedicalXpress. AI model based on deep learning detects ACL tears on knee MRI. June 4, 2019. https://medicalxpress.com/news/2019-06-ai-based-deep-acl-knee.html (accessed July 27, 2020). [Google Scholar]
- 11.Davis N AI equal with human experts in medical diagnosis, study finds. The Guardian. Sept 24, 2019. https://www.theguardian.com/technology/2019/sep/24/ai-equal-with-human-experts-in-medical-diagnosis-study-finds (accessed July 27, 2020).
- 12.Gallagher J Artificial intelligence diagnoses lung cancer. BBC News. May 20, 2019. https://www.bbc.co.uk/news/health-48334649 (accessed July 27, 2020).
- 13.Liu X, Faes L, Kale AU, et al. A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: a systematic review and meta-analysis. Lancet Digital Health 2019; 1: e271–97. [DOI] [PubMed] [Google Scholar]
- 14.van Smeden M, Van Calster B, Groenwold RHH. Machine learning compared with pathologist assessment. JAMA 2018; 319: 1725–26. [DOI] [PubMed] [Google Scholar]
- 15.Nagendran M, Chen Y, Lovejoy CA, et al. Artificial intelligence versus clinicians: systematic review of design, reporting standards, and claims of deep learning studies. BMJ 2020; 368: m689. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Volkmann A, De Bin R, Sauerbrei W, Boulesteix A-L. A plea for taking all available clinical information into account when assessing the predictive value of omics data. BMC Med Res Methodol 2019; 19: 162. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Abràmoff MD, Lavin PT, Birch M, Shah N, Folk JC. Pivotal trial of an autonomous AI-based diagnostic system for detection of diabetic retinopathy in primary care offices. NPJ Digit Med 2018; 1: 39. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Damen JA, Hooft L, Schuit E, et al. Prediction models for cardiovascular disease risk in the general population: systematic review. BMJ 2016; 353: i2416. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Beede E, Baylor E, Hersch F, et al. A human-centered evaluation of a deep learning system deployed in clinics for the detection of diabetic retinopathy. April, 2020. 10.1145/3313831.3376718 (accessed Aug 11, 2020). [DOI]
- 20.Zhang S, Bamakan SMH, Qu Q, Li S. Learning for personalized medicine: a comprehensive review from a deep learning perspective. IEEE Rev Biomed Eng 2019; 12: 194–208. [DOI] [PubMed] [Google Scholar]
- 21.Hernán MA, Robins JM. Causal inference: what if? Boca Raton: Chapman & Hall/ CRC, 2020. [Google Scholar]
- 22.Peters J, Janzing D, Schölkopf B. Elements of causal inference. Cambridge, MA, USA: The MIT Press, 2017. [Google Scholar]
- 23.Greenland S, Pearl J, Robins JM. Causal diagrams for epidemiologic research. Epidemiology 1999; 10: 37–48. [PubMed] [Google Scholar]
- 24.Angrist JD, Pischke S Jr. Mostly harmless econometrics: an empiricist’s companion. Princeton: Princeton University Press, 2009. [Google Scholar]
- 25.Rothman KJ, Gallacher JE, Hatch EE. Why representativeness should be avoided. Int J Epidemiol 2013; 42: 1012–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Shmueli G To explain or to predict? Statist Sci 2010; 25: 289–310. [Google Scholar]
- 27.Hernán MA, Hsu J, Healy B. A second chance to get causal inference right: a classification of data science tasks. Chance 2019; 32: 42–49. [Google Scholar]
- 28.Arnold KF, Davies V, de Kamps M, Tennant PWG, Mbotwa J, Gilthorpe MS. Reflections on modern methods: generalized linear models for prognosis and intervention—theory, practice and implications for machine learning. Int J Epidemiol 2020; published online May 7. 10.1093/ije/dyaa049. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Senn S Mastering variation: variance components and personalised medicine. Stat Med 2016; 35: 966–77. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Pearson ER, Flechtner I, Njølstad PR, et al. Switching from insulin to oral sulfonylureas in patients with diabetes due to Kir6.2 mutations. N Engl J Med 2006; 355: 467–77. [DOI] [PubMed] [Google Scholar]
- 31.Sundström J, Lind L, Nowrouzi S, et al. The Precision HYpertenSIon Care (PHYSIC) study: a double-blind, randomized, repeated cross-over study. Ups J Med Sci 2019; 124: 51–58. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Peto R Current misconception 3: that subgroup-specific trial mortality results often provide a good basis for individualising patient care. Br J Cancer 2011; 104: 1057–58. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Kent DM, Steyerberg E, van Klaveren D. Personalized evidence based medicine: predictive approaches to heterogeneous treatment effects. BMJ 2018; 363: k4245. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.van Klaveren D, Balan TA, Steyerberg EW, Kent DM. Models with interactions overestimated heterogeneity of treatment effects and were prone to treatment mistargeting. J Clin Epidemiol 2019; 114: 72–83. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Antoniou M, Kolamunnage-Dona R, Jorgensen AL. Biomarker-guided non-adaptive trial designs in phase 2 and phase 3: a methodological review. J Pers Med 2017; 7: 1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Burke JF, Sussman JB, Kent DM, Hayward RA. Three simple rules to ensure reasonably credible subgroup analyses. BMJ 2015; 351: h5651. [DOI] [PMC free article] [PubMed] [Google Scholar]