Skip to main content
Springer logoLink to Springer
. 2022 Jan 12;37(Suppl 2):163–165. doi: 10.1007/s12028-021-01427-6

Big Data and Artificial Intelligence for Precision Medicine in the Neuro-ICU: Bla, Bla, Bla

Giuseppe Citerio 1,2,
PMCID: PMC9343268  PMID: 35023043

“Bla, bla, bla.” With these words, Greta Thunberg condemned the world leaders in Milan over the climate crisis. She said there are many fine words, but the science does not lie: CO2 emissions are still rising. The same could apply to artificial intelligence (AI) in neurocritical care. There are plenty of articles, conference talks, a lot of bla, bla, bla but no real AI application changing our approach in managing neurocritical care patients. The AI inventers in the previous century had envisioned a future with technologies that could sense and think like, or even better than, clinicians at the bedside, a hypothesis that is probable to linger in the territories of science fiction in my remaining professional career.

Please, do not misunderstand me. I envisage a future neurocritical care system with AI that can phenotype the patient, anticipate any pathophysiological derangements early, evaluate the better strategy to be applied in the individual neurocritical care patients, suggest it to the clinician in charge, and evaluate the response to the intervention, learning from the response of the patients. I would love to have such systems working properly, guiding and supporting me and my colleagues, but the reality is that we are barely there yet, and I do not see a strong, concrete, coordinate movement for reaching this aim.

Today, only used for quoting some potential scenarios in neurocritical care, AI is not guiding intracranial hypertension therapies at the bedside, seizure or stroke treatments, or disorder of consciousness understanding.

The intracranial pressure management strategies are summarized in a simple stepwise approach for everyone [1, 2]. The effect of the ICP dose [35] has been explored, but it has not yet been integrated in clinical reasoning. The selection of strategies remains not individualized and without any AI-guided approach [6].

Some advancements in seizure detection have been trumpeted with AI using data from conventional electroencephalography (EEG). Reasonable sensitivity and specificity were described [7, 8], but it did not become routine practice. These findings are limited, and generalizability is still an issue.

Some tools available today are US Food and Drug Administration approved [9], such as the Rapid software for the detection of large cerebral vessel occlusion (LVO). It identifies suspected LVO with a high sensitivity and specificity in a couple of minutes and notifies the stroke team members when suspected LVO is detected. This is great but it is not enough [10]. We are looking for a clinical decision support tool that, when combining anamnestic, clinical, and imaging information, generates accurate estimations of likely outcomes under different treatment scenarios and suggests the best one and then provides an evaluation of the results that will be integrated for ameliorating the future algorithm.

Evidence of covert consciousness in 15% of individuals with absence of behavioral responses to motor commands following an acute brain injury was confirmed by supervised learning algorithms from EEG [11]. Moreover, 12 h following cardiac arrest, a deep-learning artificial neural network trained on EEG data in comatose patients was able to predict 6-month functional outcome [12].

In my opinion these advances are interesting but minimal, and the potential of AI in neurocritical care has still to blossom. The so-called AI chasm [13], the hiatus between developing “a scientifically sound algorithm and its use in any meaningful real-world applications,” might explain the situation we are experiencing. The AI experts focus their efforts on the technical aspects of the algorithms, with scarce consideration for the interaction with human users, in our setting, the neurointensivists. Because neurointensivists and neurointensive care unit nurses play the pivotal role in patient care, the development and the evaluation of AI-based clinical algorithms have to be focused on augmenting rather than replacing human intelligence. However, AI-based decision support systems frequently lack the possibility to be understood in their logic (the so-called black box problem) and threaten the historical medical decision-making process. Hence, the challenge is to keep neurointensivists at the center of the design and assessment route of the new AI applications.

A potential solution to the AI chasm is an early limited clinical evaluation step between the in silico development of the algorithm and a large-scale clinical trial for evaluating AI applications.

Clinical decision-making processes are complicated, and we cannot expect that human users, i.e., the neurointensivists, who remain responsible for their clinical decisions, will exactly obey all the recommendations from an algorithm with perceived opacity as an inscrutable black box [14, 15]. Therefore, it is essential to assess the actual assisted human performance and algorithm usability at an early stage in the target patient population to confirm its relevance in the implementation settings and the need to be reported as a crucial outcome. Moreover, we are not sure that clinicians’ decisions will mirror the AI recommendations. We need to test the safety profile of new algorithms when used to guide human decisions to avoid exposing to risk of harm a large group of critically ill patients.

AI ergonomics should occur as early as possible, and it requires reiterative evaluation design cycles. Technical necessities often progress as a system starts being employed, and clinicians’ expectations of the performance of an AI system also grow after the initial acquaintance period, requiring, for example, additional variables to improve the algorithm’s recommendations. This would demand AI creators to integrate other parts of the electronic patient record.

These iterative design modifications and rapid prototyping are essential and have to be planned early and not during large-scale trials to avoid undermining the whole project. This step might be compared to a phase 1 or 2 trial for drug development. However, before using the new AI tool at the bedside, we need to have a large trial demonstrating an efficacy on valuable outcomes, as we are used to with phase 3 trials before adopting a new therapeutical approach. However, this stage cannot be escaped. Large-scale clinical trials are complicated and expensive activities that require meticulous preparation. A well-planned design is fundamental for obtaining acceptable and meaningful conclusions, and it requires background information about the therapeutical strategy under evaluation. The silico evaluation cannot give all this information, and some has to be gathered in limited prospective studies. Some fundamental information has to be defined while trial protocols are drafted, for example, the most suitable outcomes for the trial, the expected effect size, the best inclusion and exclusion criteria, the evolution of the users’ confidence in the algorithm, and how to use the algorithm’s output. This path toward an integration of AI in the care of the sickest neurointensive care patients is a long route that needs a collaborative effort and funding and, as highlighted in a recent World Health Organization publication [16], requires the satisfaction of some core principles to promote its ethical use: protect autonomy; promote human well-being, human safety, and the public interest; ensure transparency, explicability and intelligibility; foster responsibility and accountability; ensure inclusiveness and equity; and promote AI that is responsive and sustainable.

Topol [17] wrote the following:

The greatest opportunity offered by AI is not reducing errors or workloads, or even curing cancer: it is the opportunity to restore the precious and time-honored connection and trust—the human touch—between patients and doctors. Not only would we have more time to come together, enabling far deeper communication and compassion, but also we would be able to revamp how we select and train doctors.

I hope we will stop the bla, bla, bla and start to build a better plan for implementing AI in medicine and in neurocritical care using the opportunity that technology will offer us to be better doctors at the bedside.

Author contributions

GC ideational and writing in a “solo” mode.

Source of Support

None.

Declarations

Conflicts of Interest

The author declare that there is no conflict of interest.

Ethical approval/informed consent

The author confirm adherence to ethical guidelines and indicate ethical approvals (IRB) and use of informed consent, as appropriate.

Footnotes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Hawryluk GWJ, Aguilera S, Buki A, Bulger E, Citerio G, Cooper DJ, et al. A management algorithm for patients with intracranial pressure monitoring: the Seattle International Severe Traumatic Brain Injury Consensus Conference (SIBICC) Intensive Care Med. 2019;3:109–114. doi: 10.1007/s00134-019-05805-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Chesnut R, Aguilera S, Buki A, Bulger E, Citerio G, Cooper DJ, et al. A management algorithm for adult patients with both brain oxygen and intracranial pressure monitoring: the Seattle international severe traumatic brain injury consensus conference (SIBICC) Intensive Care Med. 2020;46:919–929. doi: 10.1007/s00134-019-05900-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Carra G, Elli F, Ianosi B, Flechet M, Huber L, Rass V, et al. Association of dose of intracranial hypertension with outcome in subarachnoid hemorrhage. Neurocrit Care. 2021;3:722–730. doi: 10.1007/s12028-021-01221-4. [DOI] [PubMed] [Google Scholar]
  • 4.Carra G, Güiza F, Depreitere B, Meyfroidt G. Prediction model for intracranial hypertension demonstrates robust performance during external validation on the CENTER-TBI dataset. Intensive Care Med. 2021;47:124–126. doi: 10.1007/s00134-020-06247-4. [DOI] [PubMed] [Google Scholar]
  • 5.Güiza F, Depreitere B, Piper I, Citerio G, Chambers I, Jones PA, et al. Visualizing the pressure and time burden of intracranial hypertension in adult and paediatric traumatic brain injury. Intensive Care Med. 2015;41:1067–1076. doi: 10.1007/s00134-015-3806-1. [DOI] [PubMed] [Google Scholar]
  • 6.Stocchetti N, Carbonara M, Citerio G, Ercole A, Skrifvars MB, Smielewski P, et al. Severe traumatic brain injury: targeted management in the intensive care unit. Lancet Neurol. 2017;16:452–464. doi: 10.1016/S1474-4422(17)30118-7. [DOI] [PubMed] [Google Scholar]
  • 7.Fergus P, Hignett D, Hussain A, Al-Jumeily D, Abdel-Aziz K. Automatic epileptic seizure detection using scalp EEG and advanced artificial intelligence techniques. Biomed Res Int. 2015;2015:986736. doi: 10.1155/2015/986736. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Daoud H, Bayoumi MA. Efficient epileptic seizure prediction based on deep learning. IEEE Trans Biomed Circuits Syst. 2019;13:804–813. doi: 10.1109/TBCAS.2019.2929053. [DOI] [PubMed] [Google Scholar]
  • 9.Benjamens S, Dhunnoo P, Meskó B. The state of artificial intelligence-based FDA-approved medical devices and algorithms: an online database. NPJ Digit Med. 2020;3:118. doi: 10.1038/s41746-020-00324-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Bivard A, Churilov L, Parsons M. Artificial intelligence for decision support in acute stroke - current roles and potential. Nat Rev Neurol. 2020;16:575–585. doi: 10.1038/s41582-020-0390-y. [DOI] [PubMed] [Google Scholar]
  • 11.Claassen J, Doyle K, Matory A, Couch C, Burger KM, Velazquez A, et al. Detection of brain activation in unresponsive patients with acute brain injury. N Engl J Med. 2019;380:2497–2505. doi: 10.1056/NEJMoa1812757. [DOI] [PubMed] [Google Scholar]
  • 12.Tjepkema-Cloostermans MC, da Silva LC, Ruijter BJ, Tromp SC, Drost G, Kornips FHM, et al. Outcome prediction in postanoxic coma with deep learning. Crit Care Med. 2019;47:1424–1432. doi: 10.1097/CCM.0000000000003854. [DOI] [PubMed] [Google Scholar]
  • 13.Keane PA, Topol EJ. With an eye to AI and autonomous diagnosis. NPJ Digi Med. 2018;1:40. doi: 10.1038/s41746-018-0048-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Price WN. Big data and black-box medical algorithms. Sci Transl Med. 2018;10:eaao5333. doi: 10.1126/scitranslmed.aao5333. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Burrell J. How the machine ‘thinks’: understanding opacity in machine learning algorithms. Big Data Soc. 2016;3:2053951715622512. doi: 10.1177/2053951715622512. [DOI] [Google Scholar]
  • 16.World Health Organization. Ethics and governance of artificial intelligence for health. 2021. https://www.who.int/publications/i/item/9789240029200.
  • 17.Topol EJ. Deep medicine: how artificial intelligence can make healthcare human again. New York: Basic Books; 2019. [Google Scholar]

Articles from Neurocritical Care are provided here courtesy of Springer

RESOURCES