Abstract
There is increasing interest and use of artificial Intelligence algorithms and methods in biomedical research and practice, particularly as the technology has made significant advances in the past decade and become more accessible to more disciplines. This editorial briefly reviews this technology and its potential for injury prevention research and practice, proposing ways that it can be used to advance the discipline, as well as the potential pitfalls, concerns and biases that accompany it.
Over the past decade, and especially in recent years, artificial intelligence (AI) has permeated news, politics, and many aspects of everyday life (e.g., chatbots, virtual assistants, social media, smart devices). Biomedical and public health researchers and practitioners are also finding uses for AI. AI algorithms have been used to radiography and biomedical imagery, medical records,[1] and to identify built environment features associated with health outcomes. What potential do they have for injury prevention and control? A brief literature search suggests these methods are also being adopted by the field: examine road infrastructure safety and crashes,[2] predict the severity of motorcyclist injuries,[3] detect motorcycle helmet use,[4] predict and prevent sport injuries,[5] and to identify built environment typologies related to firearm violence.[6] What implications do these have for the field and how can we adopt them along with more traditional approaches?
AI refers to both a set of algorithmic approaches to analysing data, as well as the theoretical underpinnings of the discipline that has the goal of creating or simulating intelligence.[7] AI encompasses both machine learning and deep learning, the latter becoming most synonymous with AI in recent years. Deep learning algorithms underlie natural language processing models, speech recognition used in voice assistants, object detection and recognition in video and images like facial recognition and self-driving vehicles, and generative AI that create different media. AI as a field has existed since the 1940s and 1950s and many of the early successful efforts centred around text and language, though other areas of research included computer vision, speech and audio, and video.[7, 8] The field has passed through cyclical ups and downs, with current interest and progress in AI beginning in the early 2010s when newer models demonstrated high accuracy at or near human-level performance.[7] This has been possible by a combination of computing power available, distributed or cloud computing, algorithmic developments, and graphics processing units (GPUs).[9] The most common sets of algorithms that deep learning rely on are neural networks and diffusion models. While more technical details on these models can be found elsewhere, a brief overview will be provided here.[7, 10–12]
The essential structure of neural networks consists of various processing layers that examine and transform patterns in the data as they pass by each layer. The structure and connection between layers vary by the type of neural network being used with some of the most types being convolutional neural networks (CNNs), recurrent neural networks (RNNs), long short-term memory (LSTM), deep belief networks (DBNs), generative adversarial networks (GANs), and transformer neural networks. Neurons are the essential building blocks of neural networks and are meant to emulate the neurons of the human brain. The neurons are essentially a set of mathematical functions that receive input data that are weighted and summarized to then produce an output via a process known as activation. That output may then be used as input for the neurons in the next layer. Neurons are connected to one other throughout the layers and their combined weights and outputs are then used to produce the final outputs of the model (e.g., objects in images, words in text or speech, etc.). Neural networks also have other characteristics that determine a neural network’s structure (e.g., number of hidden layers and units, dropout, initialization, and activation function) and how the network is trained (e.g., learning rate, momentum, number of epochs, and batch size) known as hyperparameters. Most recent models also integrate backpropagation, which compares prediction outputs from the model with observed data to correct errors in the weights. Models like CNNs rely on feeding data sequentially forward through the network. RNNs also feed data forward and can also feed data backward. Transformer networks, in contrast, can process data in parallel due to their reliance on a self-attention mechanism that allows the model access to all data elements at once to learn the context or relationship of each element to every other element in the model.[13, 14]
Neural networks rely on training data to accomplish the tasks they are designed for and may be semi-supervised or unsupervised in this process. Semi-supervised models rely on training data that has been tagged, labelled or written by humans that can teach the neural network what the objects of interest are. Unsupervised models, on the other hand, are trained on data without human input to learn the objects of interest within the data. Unsupervised models typically need substantially more training data than other neural networks, for example some of the leading transformer networks available today have been trained on billions of data elements or more.[13, 15] In both cases, the models learn underlying characteristics and patterns in the training data through an iterative process of transforming the training data as it passes through each layer of the network. The performance of the models is typically measured through metrics of precision and recall, which are better known to epidemiologists and biostatisticians as the positive and negative predictive values, respectively. A summary metric, mean average precision (mAP), is created from precision and recall for evaluating models that is similarly interpreted as the area under the curve (AUC) metric in receiver operating characteristic (ROC) analyses.
Diffusion models are relative newcomers to AI, but are quickly becoming integral to generative AI enabling the generation of text, images, video, sound, etc. from pre-trained transformers.[16] They originate from the field of physics, particularly the study of non-equilibrium thermodynamics that includes the examination of the dispersion of substances like gases.[17, 18] Essentially, these models randomly add noise to the input data over several steps until all the data are uniform and then the process is reversed. The model learns from this process to create a set of probability densities of the input data, which can then be used to generate novel data.[16]
How can injury researchers and practitioners take advantage of these algorithms? First, these methods present immense opportunities to collect data on a much more massive scale than previously possible. Natural language processing (NLP) models can review records, extract data, and make clinical predictions from electronic health records (HER) more efficiently and quickly than human reviewers.[12, 19] Computer vision models can review millions of images to find relevant objects of interest, such as road traffic environment characteristics from street images,[20] neighbourhood characteristics from satellite imagery,[6] and identifying clinically significant objects in medical imagery.[21] These models can also identify people, vehicles and other moving objects in videos to produce counts that can be predictive of transportation mode.[22] Audio models can transcribe recordings of qualitative interviews, focus groups and other events, which can then be coded by text-focused models and analyzed.[23] These methods can also be used for mining social media to find patterns of interest, such as monitoring trends in mental health, violence, perceptions and sentiments.[24, 25] An emerging area of research is using generative models to produce simulated data to replace real-world data, particularly for protecting the privacy and confidentiality of medical data.[26, 27]
Second, just as machine learning has enabled analytical advances with large datasets, deep learning can further these efforts by replacing or enhancing some of the human effort or statistical methods. These methods can be more efficient for data reduction to create indices,[28] improve multiple imputation for missing data,[29] predicting health outcomes and risk factors.[30, 31] AI can also be used to assist with drafting communication materials,[32] assisting with analytical coding or programming,[33] visualizing changes to road infrastructure or other image and video outputs,[34] presenting and visualizing data,[35] and answering questions about injury prevention for educational purposes via chatbots.[12, 36, 37]
As with any new technology or methods, there are important pitfalls and biases that should be acknowledged and understood when using AI for research. A key challenge for most researchers will be coming to a fundamental understanding of how they work and how to apply them.[12] Injury researchers and professionals should be focused on finding expert collaborators and partners that can implement the desired data collection, analysis, or communication, just as we collaborate with other experts in other areas outside our specific discipline. Another challenge is ensuring the training data used for the models that researchers adapt to their needs are reliable and high quality.[38, 39] Many current models are trained on data from high resource settings in the U.S. and Western Europe and in English language, thus their generalizability to most of the world may not be appropriate. Many models are not entirely open, meaning the training data, the model weights, or other aspects are not available publicly; thus it is unknown how they were trained and validated. This, in part, contributes to “hallucinations” that some models may have, meaning generating text, imagery or video that is incorrect or inaccurate.[37] An additional issue with training data important to researchers is copyright and fair use, which differs from country-to-country.[40–42] Obtaining high quality data or permission to use high quality data can be expensive and challenging, thus researchers may turn to publicly available resources, yet those products may have important terms of use or service that limit use, even fair use in the US legal context.[41] The role of fair use in training artificial intelligence models is yet to be resolved, particularly for generative models.[42] The environmental impact of these models is also an important consideration as the computing power and resources needed to create new models is immense, as well as their use for customizing for a more specific use.[43] Another cause for concern is that they are a black box, even to those that have created them due to a fundamental inability to understand why these models work, how they work, and of why they may fail.[15] For causal epidemiologists and seeking to understand how exposures are related to disease, this could be problematic, particularly if relationships are observed that are unexpected or that cannot be explained by traditional means.[44, 45] Finally, there are many equity issues surrounding AI models and methods, such as the specialized knowledge needed to successfully use them, the access to the computational resources to train and implement them, the data that has been used to train them and access to those data, where they are being employed, who is involved in developing these models, and worsening existing disparities.[46–49] Some researchers have proposed all models should be accompanied with “model cards” or “datasheets” that provide standardized documentation that can make the models more transparent, ensuring they comply with FAIR (findable, accessible, interoperable, and reusable) data principles, and regulating the development and implementation of AI.[48, 50–52] There are also a growing number of recommendations for practitioners and researchers for promoting and ensuring health and healthcare equity with the integration of AI into the healthcare setting that could also be considered for developing similar recommendations for research use.[53–58]
In conclusion, artificial intelligence via deep learning has the potential to improve our efforts as injury prevention researchers, practitioners, and policymakers by improving efficiency in data collection and analysis, as well as increasing our reach and many other uses not described here. Injury Prevention welcomes submissions that examine and validate AI models and approaches in injury research and policy, including the evaluation of their ethical, equitable, and methodological soundness.
Acknowledgements.
Dr. Quistberg is supported by funding from the Fogarty International Center of the National Institutes of Health under awards K01TW011782 and 3K01TW011782-01S1. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.
References
- 1.Lewis AE, Weiskopf N, Abrams ZB, Foraker R, Lai AM, Payne PRO, Gupta A. Electronic health record data quality assessment and tools: a systematic review. Journal of the American Medical Informatics Association. 2023;30(10):1730–40. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Loo BPY, Fan Z, Lian T, Zhang F. Using computer vision and machine learning to identify bus safety risk factors. Accident Analysis & Prevention. 2023;185:107017. [DOI] [PubMed] [Google Scholar]
- 3.Das S, Dutta A, Dixon K, Minjares-Kyle L, Gillette G. Using Deep Learning in Severity Analysis of At-Fault Motorcycle Rider Crashes. Transportation Research Record. 2018;2672(34):122–34. [Google Scholar]
- 4.Siebert FW, Lin H. Detecting motorcycle helmet use with deep learning. Accident Analysis & Prevention. 2020;134:105319. [DOI] [PubMed] [Google Scholar]
- 5.Van Eetvelde H, Mendonça LD, Ley C, Seil R, Tischer T. Machine learning methods in sport injury prediction and prevention: a systematic review. Journal of Experimental Orthopaedics. 2021;8(1):27. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Jay J, Jong Jd, Jimenez MP, Nguyen Q, Goldstick J. Effects of demolishing abandoned buildings on firearm violence: a moderation analysis using aerial imagery and deep learning. Injury Prevention. 2022;28(3):249–55. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Russell SJ, Norvig P, Chang M-W. Artificial intelligence a modern approach. Harlow, England: Pearson; 2022. [Google Scholar]
- 8.Roser M. The brief history of artificial intelligence: The world has changed fast – what might be next? : OurWorldInData.org; 2022. [Available from: https://ourworldindata.org/brief-history-of-ai.
- 9.Sevilla J, Heim L, Ho A, Besiroglu T, Hobbhahn M, Villalobos P. Compute Trends Across Three Eras of Machine Learning. arXiv. 2022.
- 10.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521(7553):436–44. [DOI] [PubMed] [Google Scholar]
- 11.Esteva A, Robicquet A, Ramsundar B, Kuleshov V, DePristo M, Chou K, et al. A guide to deep learning in healthcare. Nature Medicine. 2019;25(1):24–9. [DOI] [PubMed] [Google Scholar]
- 12.Thirunavukarasu AJ, Ting DSJ, Elangovan K, Gutierrez L, Tan TF, Ting DSW. Large language models in medicine. Nature Medicine. 2023;29(8):1930–40. [DOI] [PubMed] [Google Scholar]
- 13.Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is All you Need. In: Guyon I, Luxburg UV, Bengio S, Wallach H, Fergus R, Vishwanathan S, Garnett R, editors.2017.
- 14.Raschka S. Ahead of AI [Internet]2023. [cited 2023]. Available from: https://sebastianraschka.com/blog/2023/self-attention-from-scratch.html.
- 15.Bommasani R, Hudson DA, Adeli E, Altman R, Arora S, von Arx S, et al. On the Opportunities and Risks of Foundation Models. arXiv. 2021.
- 16.Yang L, Zhang Z, Song Y, Hong S, Xu R, Zhao Y, et al. Diffusion Models: A Comprehensive Survey of Methods and Applications. ACM Comput Surv. 2023;56(4):Article 105. [Google Scholar]
- 17.Wiggers K. A brief history of diffusion, the tech at the heart of modern image-generating AI. TechCrunch; [Internet]. 2022 22 Dec 2022. Available from: https://techcrunch.com/2022/12/22/a-brief-history-of-diffusion-the-tech-at-the-heart-of-modern-image-generating-ai/. [Google Scholar]
- 18.Ananthaswamy A. The Physics Principle That Inspired Modern AI Art. Quanta Magazine [Internet]. 2023. Available from: https://www.quantamagazine.org/the-physics-principle-that-inspired-modern-ai-art-20230105/.
- 19.Wu S, Roberts K, Datta S, Du J, Ji Z, Si Y, et al. Deep learning in clinical natural language processing: a methodical review. Journal of the American Medical Informatics Association. 2019;27(3):457–70. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Nguyen QC, Keralis JM, Dwivedi P, Ng AE, Javanmardi M, Khanna S, et al. Leveraging 31 Million Google Street View Images to Characterize Built Environments and Examine County Health Outcomes. Public Health Reports. 2021;136(2):201–11. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Esteva A, Chou K, Yeung S, Naik N, Madani A, Mottaghi A, et al. Deep learning-enabled medical computer vision. npj Digital Medicine. 2021;4(1):5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Doiron D, Setton EM, Brook JR, Kestens Y, McCormack GR, Winters M, et al. Predicting walking-to-work using street-level imagery and deep learning in seven Canadian cities. Scientific Reports. 2022;12(1):18380. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Lennon RP, Fraleigh R, Scoy LJV, Keshaviah A, Hu XC, Snyder BL, et al. Developing and testing an automated qualitative assistant (AQUA) to support qualitative analysis. Family Medicine and Community Health. 2021;9(Suppl 1):e001287. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Singh T, Roberts K, Cohen T, Cobb N, Wang J, Fujimoto K, Myneni S. Social Media as a Research Tool (SMaaRT) for Risky Behavior Analytics: Methodological Review. JMIR Public Health Surveill. 2020;6(4):e21660. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Klein AZ, Banda JM, Guo Y, Schmidt AL, Xu D, Amaro JIF, et al. Overview of the 8th Social Media Mining for Health Applications (#SMM4H) Shared Tasks at the AMIA 2023 Annual Symposium. medRxiv. 2023:2023.11.06.23298168. [DOI] [PMC free article] [PubMed]
- 26.Chen RJ, Lu MY, Chen TY, Williamson DFK, Mahmood F. Synthetic data in machine learning for medicine and healthcare. Nature Biomedical Engineering. 2021;5(6):493–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Giuffrè M, Shung DL. Harnessing the power of synthetic data in healthcare: innovation, application, and privacy. npj Digital Medicine. 2023;6(1):186. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Howe B, Brown JM, Han B, Herman B, Weber N, Yan A, et al. Integrative urban AI to expand coverage, access, and equity of urban data. The European Physical Journal Special Topics. 2022;231(9):1741–52. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Getz K, Hubbard RA, Linn KA. Performance of Multiple Imputation Using Modern Machine Learning Methods in Electronic Health Records Data. Epidemiology. 2023: 10.1097/EDE.0000000000001578. [DOI] [PubMed]
- 30.Placido D, Yuan B, Hjaltelin JX, Zheng C, Haue AD, Chmura PJ, et al. A deep learning algorithm to predict risk of pancreatic cancer from disease trajectories. Nature Medicine. 2023;29(5):1113–22. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Savcisens G, Eliassi-Rad T, Hansen LK, Mortensen LH, Lilleholt L, Rogers A, et al. Using sequences of life-events to predict human lives. Nature Computational Science. 2023. [DOI] [PubMed]
- 32.Ray PP. ChatGPT: A comprehensive review on background, applications, key challenges, bias, ethics, limitations and future scope. Internet of Things and Cyber-Physical Systems. 2023;3:121–54. [Google Scholar]
- 33.Mollick E. One Useful Thing [Internet]. Philadelphia, PA: Substack. 2023. [cited 2023]. Available from: https://www.oneusefulthing.org/p/it-is-starting-to-get-strange. [Google Scholar]
- 34.Holland.com. Add a touch of Dutch to your Street: Holland.com; 2023. [Available from: https://dutchcyclinglifestyle.com/.
- 35.Wu A, Wang Y, Shu X, Moritz D, Cui W, Zhang H, et al. AI4VIS: Survey on Artificial Intelligence Approaches for Data Visualization. IEEE Transactions on Visualization & Computer Graphics. 2022;28(12):5049–70. [DOI] [PubMed] [Google Scholar]
- 36.Wilson L, Marasoiu M. The Development and Use of Chatbots in Public Health: Scoping Review . JMIR Hum Factors. 2022;9(4):e35882. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Oviedo-Trespalacios O, Peden AE, Cole-Hunter T, Costantini A, Haghani M, Rod JE, et al. The risks of using ChatGPT to obtain common safety-related information and advice. Safety Science. 2023;167:106244. [Google Scholar]
- 38.El-Mhamdi E-M, Farhadkhani S, Guerraoui R, Gupta N, Hoang L-N, Pinot R, et al. On the Impossible Safety of Large AI Models2022 September 01, 2022:[arXiv:2209.15259 p.]. Available from: https://ui.adsabs.harvard.edu/abs/2022arXiv220915259E https://arxiv.org/pdf/2209.15259.pdf.
- 39.Bender EM, Gebru T, McMillan-Major A, Shmitchell S . Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency; Virtual Event, Canada: Association for Computing Machinery; 2021. p. 610–23. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Dreben RN, Julyan MT. Lawflash [Internet]. Washington, DC: Morgan Lewis. 2023. [cited 2023]. Available from: https://www.morganlewis.com/pubs/2023/03/generative-artificial-intelligence-and-copyright-current-issues. [Google Scholar]
- 41.Henderson P, Li X, Jurafsky D, Hashimoto T, Lemley MA, Liang P. Foundation Models and Fair Use. arXiv. 2023;2303.15715.
- 42.Levine AM. Practicioner Insights Commentaries [Internet]: Westlaw Today. 2023. 21 Sep. [cited 2023]. Available from: https://today.westlaw.com/Document/I3f6be98a58ac11ee8921fbef1a541940/View/FullText.html.
- 43.Dodge J, Prewitt T, Combes RTd, Odmark E, Schwartz R, Strubell E, et al. Measuring the Carbon Intensity of AI in Cloud Instances. Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency; Seoul, Republic of Korea: Association for Computing Machinery; 2022. p. 1877–94. [Google Scholar]
- 44.Mitra N, Roy J, Small D. The Future of Causal Inference. American Journal of Epidemiology. 2022;191(10):1671–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Mooney SJ, Keil AP, Westreich DJ. Thirteen Questions About Using Machine Learning in Causal Research (You Won’t Believe the Answer to Number 10!). American Journal of Epidemiology. 2021;190(8):1476–82. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Smith MJ, Axler R, Bean S, Rudzicz F, Shaw J. Four equity considerations for the use of artificial intelligence in public health. Bulletin of the World Health Organization. 2020;98(4):29–02. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.de Hond AAH, van Buchem MM, Hernandez-Boussard T. Picture a data scientist: a call to action for increasing diversity, equity, and inclusion in the age of AI. Journal of the American Medical Informatics Association. 2022;29(12):2178–81. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Huerta EA, Blaiszik B, Brinson LC, Bouchard KE, Diaz D, Doglioni C, et al. FAIR for AI: An interdisciplinary and international community building perspective. Scientific Data. 2023;10(1):487. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Zack T, Lehman E, Suzgun M, Rodriguez JA, Celi LA, Gichoya J, et al. Assessing the potential of GPT-4 to perpetuate racial and gender biases in health care: a model evaluation study. The Lancet Digital Health. 2024;6(1):e12–e22. [DOI] [PubMed] [Google Scholar]
- 50.Mitchell M, Wu S, Zaldivar A, Barnes P, Vasserman L, Hutchinson B, et al. Model Cards for Model Reporting2018 October 01, 2018:[arXiv:1810.03993 p.]. Available from: https://ui.adsabs.harvard.edu/abs/2018arXiv181003993M https://arxiv.org/pdf/1810.03993.pdf.
- 51.Embi PJ. Algorithmovigilance—Advancing Methods to Analyze and Monitor Artificial Intelligence–Driven Health Care for Effectiveness and Equity. JAMA Network Open. 2021;4(4):e214622–e. [DOI] [PubMed] [Google Scholar]
- 52.Gebru T, Morgenstern J, Vecchione B, Vaughan JW, Wallach H, III HD, Crawford K. Datasheets for datasets. Commun ACM. 2021;64(12):86–92. [Google Scholar]
- 53.Clark CR, Wilkins CH, Rodriguez JA, Preininger AM, Harris J, DesAutels S, et al. Health Care Equity in the Use of Advanced Analytics and Artificial Intelligence Technologies in Primary Care. Journal of General Internal Medicine. 2021;36(10):3188–93. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Cerrato P, Halamka J, Pencina M. A proposal for developing a platform that evaluates algorithmic equity and accuracy. BMJ Health & Care Informatics. 2022;29(1):e100423. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Abràmoff MD, Tarver ME, Loyo-Berrios N, Trujillo S, Char D, Obermeyer Z, et al. Considerations for addressing bias in artificial intelligence for health equity. npj Digital Medicine. 2023;6(1):170. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Pham Q, Gamble A, Hearn J, Cafazzo JA. The Need for Ethnoracial Equity in Artificial Intelligence for Diabetes Management: Review and Recommendations. Journal of medical Internet research. 2021;23(2):e22320. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Brown S-A, Sparapani R, Osinski K, Zhang J, Blessing J, Cheng F, et al. Establishing an interdisciplinary research team for cardio-oncology artificial intelligence informatics precision and health equity. American Heart Journal Plus: Cardiology Research and Practice. 2022;13:100094. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Zhang J, Whebell S, Gallifant J, Budhdeo S, Mattie H, Lertvittayakumjorn P, et al. An interactive dashboard to track themes, development maturity, and global equity in clinical artificial intelligence research. The Lancet Digital Health. 2022;4(4):e212–e3. [DOI] [PMC free article] [PubMed] [Google Scholar]