Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2023 Jun 20;13:9975. doi: 10.1038/s41598-023-37114-z

A comparison of performance between a deep learning model with residents for localization and classification of intracranial hemorrhage

Salita Angkurawaranon 1,2, Nonn Sanorsieng 1, Kittisak Unsrisong 1, Papangkorn Inkeaw 3, Patumrat Sripan 4, Piyapong Khumrin 5, Chaisiri Angkurawaranon 2,5, Tanat Vaniyapong 6, Imjai Chitapanarux 1,
PMCID: PMC10282020  PMID: 37340038

Abstract

Intracranial hemorrhage (ICH) from traumatic brain injury (TBI) requires prompt radiological investigation and recognition by physicians. Computed tomography (CT) scanning is the investigation of choice for TBI and has become increasingly utilized under the shortage of trained radiology personnel. It is anticipated that deep learning models will be a promising solution for the generation of timely and accurate radiology reports. Our study examines the diagnostic performance of a deep learning model and compares the performance of that with detection, localization and classification of traumatic ICHs involving radiology, emergency medicine, and neurosurgery residents. Our results demonstrate that the high level of accuracy achieved by the deep learning model, (0.89), outperforms the residents with regard to sensitivity (0.82) but still lacks behind in specificity (0.90). Overall, our study suggests that the deep learning model may serve as a potential screening tool aiding the interpretation of head CT scans among traumatic brain injury patients.

Subject terms: Medical research, Neurology

Introduction

Among common neurological problems, traumatic brain injury (TBI) is one of the most prevalent and poses one of the most important burdens on public health1. A head computed tomography (CT) scan, an effective non-invasive modality, is almost always the first-line investigation of acute TBI, owing to the widespread availability for the procedure and also the short acquisition time. CT scans have the ability to detect intracranial hemorrhage (ICH), mass effect, and associated complications. As a result, patients requiring emergency neurosurgical intervention can be identified rapidly2.

Due to the emergency nature of trauma, doctors need to obtain and interpret CT scans as quickly as possible. This is especially important in the case of head injuries, where timely treatment can avoid cognitive and physical disability. Emergency physicians and neurosurgeons must decide whether to plan operative or conservative treatment for the patient. The potential subtypes of ICH that may necessitate surgical intervention include intraparenchymal hemorrhage (IPH), subdural hemorrhage (SDH), and epidural hemorrhage (EDH)3. Rapid trauma response systems, including the availability of CT scans and adequate personnel, are required to prevent the possibly long-lasting effects of secondary brain injury and enhance patient outcomes4.

However, the number of trained radiologists or even radiology trainees available to interpret the CT scans has often been limited, resulting in significant delays in analyzing and reporting results5,6. Thus, in resource-limited settings, treatment planning before the formal radiology report may result in misinterpretation and inappropriate clinical management.7,8.

A promising solution to tackle this problem is the usage of Artificial Intelligence (AI). Many studies have used deep learning methods to assist in the diagnosis of diseased, oncologic, and traumatic patients912. An automated ICH detection and classification tool may assist residents or clinicians when medical radiology experts are not immediately available13,14. Deep learning models have also been implemented to detect ICH and even assess mass effects from ICH in both retrospective and prospective studies1517. The majority of the studies1721 focused on the evaluation of diagnostic accuracy for the identification of ICH and classification into each ICH subtype using algorithms, but, to date, the accuracy of the detection of ICH into specific intracranial locations has not been well evaluated.

Recently, the authors of this study developed a deep learning model for segmenting SDH, EDH, and IPH22. The model proposed outperformed segmentation performance with a higher dice score when compared to reports in previously published literature22. In this study, we aim to compare the performance of the proposed deep learning segmentation model22 with that of radiology, emergency, and neurosurgery residents. We primarily focus on the detection of IPH, SDH, and EDH, as these subtypes are usually the ones being evaluated in identifying and selecting TBI patients for neurosurgical intervention3.

Material and methods

Development of the deep learning model

The deep learning model we investigated in this study was proposed in our prior published study22. The model is a variation of the DeepMedic23 model that has the ability to segment SDH, EDH, and IPH on a CT scan. Its architecture consists of four parallel pathways that process the input at different resolutions and two fully connected layers. It obtains a 2-channel voxel extracted from the subdural and bone windows of a brain CT scan as the input. All voxels in a CT scan were processed by the model to generate segmentation results with the value of each pixel as the class label of the hemorrhage type where the pixel was located. After the segmentation results were produced by the deep learning model, the regions of SDH and EDH with their major axis of less than 5 mm were considered as noise and removed from the segmentation result. The final results were drawn on the input CT scan by assigning different colors to hemorrhage types. The model is shared via https://github.com/RadiologyCMU/Hemorrhage-DeepMedic. In this study, the deep learning model was performed without any fine-tuning. We also adopted the same data pre-processing process used in our previous work. In addition, a new dataset was used as the test dataset, which differed from the dataset used to train the model in22. The samples in the test dataset were randomly selected from patients who were not in the training dataset. Patients who received multiple scans, we have chosen only the first scan.

Study cohort

Non-contrast head CT scans of adult patients aged 15 years or older suspected of head injury/trauma as the initial presentation during emergency department (ED) visits at Maharaj Nakorn Chiang Mai Hospital from January 1, 2014, to December 31, 2014, were included. Exclusion criteria included: (1) Follow-up CT studies of patients with known recent TBI; (2) Studies of patients with recent neurosurgical intervention; (3) Severe artifacts degrading study quality such as motion artifacts and metallic artifacts.

Brain CT scans were acquired with CT equipment from either of two manufacturers (Toshiba Aquilion 16 or Siemens SOMATOM Definition). Each slide was stored as a 512 × 512 pixels DICOM image. The typical image resolution of x and y is 0.4473 mm per pixel.

The number of image slices per patient may vary between 80 and 115, depending on factors such as the size of the patient's head, with a fixed separation distance between slices of 1.5 mm. With these criteria, a total of 300 head CT studies from different subjects were finally included, 166 studies were categorized under the intracranial hemorrhage (ICH) group, while 134 studies were classified under the non-ICH group. After thorough review and annotation from consensus of two experienced neuroradiologists out of the 300 head CT studies, 171 were identified as lesions of IPH (27.01%), 356 lesions of SDH (56.24%), and 106 lesions of EDH (16.75%). Detailed data describing locations of lesions are shown in Table 1.

Table 1.

Type and distribution of ICHs. Abbreviations: ICH, intracranial hemorrhage; IPH, intraparenchymal hemorrhage; EDH, epidural hemorrhage; SDH, subdural hemorrhage.

Type of ICH Number of lesions
IPH (lesions) 171
Locations
 Right hemisphere 84
 Left hemisphere 67
 Right cerebellar convexity 2
 Left cerebellar convexity 1
 Right Deep Nuclei 8
 Left Deep Nuclei 3
 Midbrain 5
 Pons 1
SDH (lesions) 356
Locations
 Right cerebral convexity 74
 Left cerebral convexity 75
 Falx cerebri 100
 Tentorium cerebelli 99
 Right cerebellar convexity 5
 Left cerebellar convexity 3
EDH (lesions) 106
Locations
 Right cerebral convexity 48
 Left cerebral convexity 49
 Right cerebellar convexity 3
 Left cerebellar convexity 5
Vertex 1

Classification of ICH by deep learning model

The deep learning model was then used to identify negative and positive studies. In the positive studies, the subtypes of ICH of concern were classified and segmented. The deep learning model segmented the area of ICH with different color maps at specific locations indicating different subtypes and locations of ICH. Correct segmentation means coloring true ICH in the expected locations, as in Fig. 1. The true location of ICH but incorrect color regarding ICH subtypes were considered false interpretations.

Figure 1.

Figure 1

ICH Segmentation with Different Colors Referring to Different Types of ICH. (a) SDHs are colored in green along the right cerebral convexity, falx cerebri, and tentorium cerebelli, equivalent to 3 locations of SDHs; (b) IPH is colored in blue at the left frontal lobe, equivalent to the left cerebral hemisphere, and is counted as 1 location; (c) EDH at the left cerebral convexity is colored in red.

Classification of ICH by residents

Four radiology residents (three junior radiology residents and one senior radiology resident) and four non-radiology residents (two senior emergency medicine residents and two senior neurosurgery residents) were recruited for the study. These three areas of specialist were chosen as the residents attached to these areas of expertise were most likely to be those making the initial CT interpretation at the emergency department. Blinded to the original CT results, all eight residents were independently required to interpret all head CT scans solely from the CT series consisting of a 1.5-mm thick slice in the axial plane. The residents could manually adjust the width and level of the window for each scan during interpretation.

A record form was created consisting of multiple-choice checkboxes regarding the location of each ICH subtype. The locations for IPH were: (1) Right cerebral hemisphere; (2) Left cerebral hemisphere; (3) Right cerebellar hemisphere; (4) Left cerebellar hemisphere; (5) Right deep nuclei; (6) Left deep nuclei; (7) Midbrain; (8) Pons. Deep nuclei encompassed either the caudate nucleus, lentiform nucleus, or thalamus. In the case of SDH, the locations could be: (1) Right cerebral convexity; (2) Left cerebral convexity; (3) Falx cerebri; (4) Tentorium cerebelli; (5) Right cerebellar convexity; (6) Left cerebellar convexity. For EDH, the locations included: (1) Right cerebral convexity; (2) Left cerebral convexity; (3) Right cerebellar convexity; (4) Left cerebellar convexity; (5) Vertex. If no ICH locations were identified in IPH, SDH, nor EDH these were classed as negative-ICH. A flowchart methodology of this study is available in the Fig. 2.

Figure 2.

Figure 2

Flowchart of this study.

Statistical analysis

In most cases, there were more than one ICH subtype and/or multiple lesions of the same subtype. The algorithm or trainee residents would have to correctly identify all ICH subtypes and a correct location was considered as “detected”. If any ICH remained undetected or was mis-identified either by subtype or location, it was counted as “missed”. The imaging of ICH subtypes in certain locations is presented in Fig. 3. We evaluated the performance of the algorithm and trainees using statistical metrics, including accuracy, sensitivity, and specificity, using the Python package scikit-learn. A significant difference was considered when p < 0.05.

Figure 3.

Figure 3

Identification and Localization of ICH: (a) EDH at left cerebral convexity, (b) SDH at right cerebral convexity, (c) IPH at right cerebral hemisphere.

Informed consent and ethical approval

This study was approved by the Research Ethics Committee of the Faculty of Medicine, Chiang Mai University (No.423/2021). Informed consent was obtained from all participants. All methods were performed in accordance with the relevant guidelines and regulations.

Results

We evaluated the performance of the deep learning model and compared it to that of the eight training residents in the classification and localization of ICHs based on the individual locations occupied by specific types of ICH. The accuracy, sensitivity, and specificity of the algorithm and residents are displayed in Table 2. In terms of ICH detection and localization, the model achieved an accuracy of 0.89 with a sensitivity, and specificity of 0.82 and 0.90, respectively. Overall, four radiology residents achieved accuracy, sensitivity, and specificity of 0.96 ± 0.00, 0.74 ± 0.04, and 0.99 ± 0.01, whereas four non-radiology residents scored values of 0.94 ± 0.01, 0.61 ± 0.08, and 0.99 ± 0.00, respectively.

Table 2.

Location-level performance of the algorithm, three junior radiology residents, a senior radiologist, two emergency residents, and two neurosurgery residents presented as accuracy, sensitivity, and specificity. Abbreviations: ICH, intracranial hemorrhage; IPH, intraparenchymal hemorrhage; EDH, epidural hemorrhage; SDH, subdural hemorrhage.

Hemorrhage Interpreters Accuracy Sensitivity Specificity
ICH Model 0.89 0.82 0.90
Junior radiology resident 1 0.96 0.78 0.98
Junior radiology resident 2 0.96 0.70 0.99
Junior radiology resident 3 0.96 0.71 0.99
Senior radiology resident 0.96 0.78 0.98
Radiology residents 0.96 ±  < 0.01 0.74 ± 0.04 0.99 ± 0.01
Senior EM resident 1 0.95 0.70 0.98
Senior EM resident 2 0.94 0.56 0.99
Senior NS resident 1 0.94 0.52 0.99
Senior NS resident 2 0.95 0.64 0.99
Non-radiology residents 0.94 ± 0.01 0.61 ± 0.08 0.99 ±  < 0.01
IPH Model 0.93 0.83 0.94
Junior radiology resident 1 0.97 0.78 0.99
Junior radiology resident 2 0.98 0.79 0.99
Junior radiology resident 3 0.98 0.77 0.99
Senior radiology resident 0.98 0.85 0.99
Radiology residents 0.98 ±  < 0.01 0.80 ± 0.04 0.99 ±  < 0.01
Senior EM resident 1 0.96 0.79 0.97
Senior EM resident 2 0.97 0.65 0.99
Senior NS resident 1 0.97 0.73 0.98
Senior NS resident 2 0.97 0.68 0.99
Non-Radiology residents 0.96 ±  < 0.01 0.71 ± 0.06 0.98 ± 0.01
EDH Model 0.91 0.72 0.93
Junior radiology resident 1 0.97 0.74 0.99
Junior radiology resident 2 0.97 0.61 1.00
Junior radiology resident 3 0.98 0.76 0.99
Senior radiology resident 0.97 0.71 0.99
Radiology residents 0.97 ±  < 0.01 0.71 ± 0.07 0.99 ±  < 0.01
Senior EM resident 1 0.97 0.71 0.99
Senior EM resident 2 0.96 0.55 0.99
Senior NS resident 1 0.96 0.52 0.99
Senior NS resident 2 0.97 0.74 0.99
Non-Radiology residents 0.97 ± 0.01 0.63 ± 0.11 0.99 ±  < 0.01
SDH Model 0.82 0.85 0.82
Junior radiology resident 1 0.93 0.79 0.96
Junior radiology resident 2 0.93 0.69 0.99
Junior radiology resident 3 0.92 0.67 0.99
Senior radiology resident 0.92 0.76 0.96
Radiology residents 0.93 ±  < 0.01 0.73 ± 0.06 0.97 ± 0.01
Senior EM resident 1 0.91 0.66 0.98
Senior EM resident 2 0.89 0.51 0.98
Senior NS resident 1 0.88 0.42 0.99
Senior NS resident 2 0.91 0.59 0.99
Non-Radiology residents 0.90 ± 0.02 0.54 ± 0.11 0.98 ± 0.01

Regarding the detection and localization of each ICH subtype, the deep learning model was the most sensitive in detecting SDH (sensitivity = 0.85). The sensitivities for detecting IPH and EDH were 0.83 and 0.72, respectively. The radiology residents performed similarly well to the model in the sensitivity of each ICH subtype. The sensitivity for IPH was 0.80, 0.71 for EDH, and 0.73 for SDH. The neurosurgery and emergency medicine residents had lower performance scores in detecting each ICH subtype, with a sensitivity of 0.71 for IPH detection, 0.63 for EDH, and 0.54 for SDH. The model had an overall higher sensitivity for ICH detection than the average performance of training residents across all ICH subtypes (p < 0.05).

In several cases, the deep learning model detected subtle SDH when this could not be detected by any of the residents. Most of these cases were studies containing thin SDHs either along the tentorium cerebelli or the cerebral convexities. Some of these cases had various hemorrhagic subtypes resulting in small hemorrhages being overlooked. Three subtle EDHs were missed by radiology residents, and five subtle EDHs were missed by non-radiology residents. However, all of these EDHs were picked up by the algorithm. These cases are illustrated in Figs. 4 and 5. Only two cases of small EDHs were missed by the deep learning but were able to be detected by all residents (Fig. 6).

Figure 4.

Figure 4

SDH Along Right Tentorium Cerebelli (a) are correctly colored green by the algorithm (b). Another case with subtle right cerebral convexity SDH (c) was also detected by the algorithm and colored in green (d).

Figure 5.

Figure 5

Missed EDH at Right Cerebral Convexity (a) is correctly colored by the algorithm in red (b) with associated skull fracture demonstrated on bone window image (c).

Figure 6.

Figure 6

The CT scan of the brain shows a small left frontal convexity EDH (white arrow) in narrowed window setting (a) and associated skull fracture in the bone window setting (b). The post-processing image reveals that the EDH has not been annotated by the algorithm c).

The specificities for ICH detection and subtypes of ICH by the deep learning and residents were relatively high. The overall specificity for ICH detection by the algorithm was 0.90, while the specificities for ICH detection by radiology and non-radiology residents were both 0.99. Specificity values for each subtype of ICH varied between 0.82 and 0.90 in the case of the deep learning, while the specificity values varied between 0.97 and 0.99 for the residents for each ICH subtype (Table 2). There were many cases in which the deep learning model overdiagnosed, usually involving basal ganglia calcification, beam hardening artifacts, dense cortical veins, and dural venous sinuses being interpreted as hemorrhage. Some of these studies are presented in Fig. 7.

Figure 7.

Figure 7

Demonstrating the misinterpretation of basal ganglia calcification by the algorithm (a), beam hardening artifact (b), cortical vein (c), and sigmoid sinus (d) as ICHs.

Discussion

Our results demonstrate non-inferior diagnostic accuracies in ICH detection of the deep learning model compared to residents (p > 0.05). In terms of sensitivity, the model yielded noticeably higher overall sensitivity values for ICH detection across nearly all subtypes compared to the residents (p < 0.05). Nonetheless, specificity of the model still falls behind that of the residents.

In our study, we determined the accuracy of the algorithm by evaluating the frequency of correct and incorrect ICH detection from each location. The deep learning model achieved high accuracy of overall ICH detection, and IPH, EDH, and SDH detection (0.89, 0.93, 0.91, and 0.82). The reported diagnostic performance of deep learning models in prior literature were variable, with accuracy values ranging from 0.70 to 0.9413. Similar studies based on convolutional neural networks yielded approximate accuracy values from 0.81 to 0.9020,24. The variations in accuracy are likely due to the classification methods among different models and ways of measurement for ICH segmentation which might not allow direct comparison between studies.

Few studies have compared the diagnostic performance between algorithms and trainee residents. Ye et al. demonstrated superior performance of the deep learning neural network. They concluded that their algorithm was fast and accurate, indicating its potential role in assisting junior radiology residents in reducing misinterpretation of head CT scans15. However, they primarily focused on ICH detection and classification, while our study also stresses the importance of identifying both the subtype and correct location.

Although our deep learning model demonstrated high specificity, it still lags behind that of the residents. From a review of the lesion segmentation by the algorithm, many non-ICH findings had been misinterpreted as ICH. This was possibly because identification of ICH required the region be of higher attenuation or higher Hounsfield unit (HU) than surrounding normal brain parenchyma. Other findings with high HU, such as basal ganglia calcification, beam hardening artifacts, dense cortical veins, and dural venous sinuses, were misclassified as hemorrhage. This is concordant with prior results in that common AI overcalls are calcification and beam-hardening artifacts25,26. While radiologists and clinicians gain experience in identifying actual ICHs, AI analysis software must also be trained to recognize these ICH mimickers.

In terms of sensitivity, the deep learning model yielded noticeably higher overall sensitivity values for ICH detection across nearly all subtypes compared to the residents (p < 0.05). Moreover, the deep learning model could detect subtle hemorrhages missed by residents, such as thin SDH and small EDH. Waite et al. suggest that in their study of interpretative errors in radiology perceptual errors account for 60%-80% of radiological errors27. Since perception and detection are the initial phases in image interpretation, an error in this phase can abruptly terminate the diagnostic process and result in a mistaken (false-negative) diagnosis. Perceptual errors in radiology have been linked to a variety of causes, including fatigue of the interpreter and the increased pace of interpretations28. The error rate could also change depending on the time of day the interpretation was given, and long and overnight shifts are associated with increased rates of inaccuracy, according to previous studies29,30. More crucially, in approximately 1% of cases, the diagnostic error results in incorrect or inadequate patient management31,32. This reinforces the vital role of a deep learning model as a potential screening tool in emergency cases requiring rapid ICH detection or as an assistant for trainee residents in generating emergency CT reports. Similar work has been done using an active automated tool to detect acute intracranial conditions, including hemorrhage, reprioritizing study, and notifying radiologists if ICH was identified as present. Thus, these resulted in significant reductions in diagnostic waiting time with high sensitivity with regard to detection33,34.

Different sensitive and specificity tradeoff is likely to be different depending on their role in the care pathway. If primary role of the operator, whether it be the radiologist, emergency physician or AI, is to screen/inform or refer for appropriate interventions, due to the serious and urgent nature of TBI and importance of timely surgical intervention for some subtypes of TBI, it may be better to set a high sensitivity and allow more false positives (lower positive predictive value). As a reference, the American College of Surgeons Committee on Trauma has set the national benchmark for field triage at ≥ 95% for sensitivity and ≥ 50% specificity35. However, if the primary role of the operator, whether it be the radiologist, neurosurgeon or AI, is to decide the definitive treatment on whether or not to perform neurosurgery, it would be better to set a high specificity.

Following this argument, outcomes of this study suggest that the deep learning model may be a useful screening tool for the detection and localization of ICH from CT scans in the cases of traumatic brain injury in the majority of emergency departments (ED), in particular where there may not be 24-hour coverage of trained radiologists. However, confirmation by trained personnel is required before definitive surgical treatment plans can be made. This is particularly the case in low- and middle-income settings when emergency physicians and neurosurgeons frequently evaluate emergency computed tomography (CT) scans without support from trained radiologists.

There are three main limitations to our study. First, the performance evaluation is done by scoring the detected ICH according to crude locations, which sometimes represent a wide and less specific area within the cranium. For example, hematoma in the cerebral hemisphere could be in either the frontal, parietal, temporal, or occipital lobes. Multiple separated hematomas may be present in different locations within the same hemisphere. If only one hematoma is segmented or identified in the setting of multiple discrete hematomas confined within the recording area (e.g., a hemisphere), this will be erroneously considered as all hematomas being detected when the remaining hematomas have been missed. However, unlike a previous study15, ours is one of the few that has attempted to match the type of ICH and their precise locations with performance comparisons to training residents rather than simply detecting ICH subtype alone. The second limitation is the sample size; with only 300 cases of head CT studies, the study may not have the same level of statistical power in defining the exact diagnostic performance of the deep learning model in comparison to some other studies in the literature16,17,36. In addition, all the subjects were limited to 15 or more years of age. In future studies, it would be useful if the data set were expanded to include all age groups. Lastly, we did not validate the deep learning model with the external dataset (e.g. from another hospital or geographical area). Samples in the test dataset were collected from the same hospital and CT machines used for the training dataset. This limitation needs to be addressed in future validation.

Conclusion

In conclusion, our study is one of the first to validate the efficacy of the role of the deep learning model for ICH detection and localization by comparing the level of diagnostic accuracy with radiology, emergency medicine, and neurosurgery residents. Based on the results, our study highlighted the potential use of AI as a useful intracranial hemorrhage screening tool in traumatic brain injury patients. However, its slightly lower specificity and tendency to misinterpret some benign lesions with high attenuation into hemorrhage remain an issue to be addressed. Further model training with a larger data set and a larger sample size is expected to improve the overall capability of our deep learning model in a real clinical setting.

Acknowledgements

This research was supported by the Program Management Unit for Human Resources and Institutional Development, Research, and Innovation, NXPO [grant number B04G640072]. This work was partially supported by Chiang Mai University [grant number 8/2565].

Author contributions

S.A. contributed to Conceptualization, Methodology, Investigation, Resources, Data Curation, Writing—Original Draft, Visualization. N.S. contributed to Resources, Writing—Review & Editing. K.U. contributed to Resources, Writing—Review & Editing. P.I. contributed to Conceptualization, Software, Formal Analysis, Writing—Original Draft, Visualization. P.K. contributed to Software, Resources, Writing—Review & Editing. P.S. contributed to formal analysis, Writing—Review & Editing. C.A. contributed to Methodology, Formal Analysis, Writing—Review & Editing. T.V. contributed to investigation, writing- review&editing. I.C. contributed to Conceptualization, Writing—Review & Editing, Project Administration, Funding Acquisition.

Data availability

The datasets generated during and/or analysed during the current study are available from the corresponding author on reasonable request.

Competing interests

The authors declare no competing interests.

Footnotes

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Maas AIR, Menon DK, Manley GT, Abrams M, Åkerlund C, Andelic N, et al. Traumatic brain injury: Progress and challenges in prevention, clinical care, and research. Lancet Neurol. 2022;21(11):1004–1060. doi: 10.1016/S1474-4422(22)00309-X. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Fink KR. Imaging of head trauma. Semin. Roentgenol. 2016;51(3):143–151. doi: 10.1053/j.ro.2016.05.001. [DOI] [PubMed] [Google Scholar]
  • 3.Kvint S, Gutierrez A, Blue R, Petrov D. Surgical management of trauma-related intracranial hemorrhage-a review. Curr. Neurol. Neurosci Rep. 2020;20(12):63. doi: 10.1007/s11910-020-01080-0. [DOI] [PubMed] [Google Scholar]
  • 4.Naidoo D. Traumatic brain injury: The South African landscape. South Afr. Med. Journal. 2013;103(9):613–614. doi: 10.7196/SAMJ.7325. [DOI] [PubMed] [Google Scholar]
  • 5.Boland, G.W., Guimaraes, A.S. & Mueller, P.R. The radiologist's conundrum: Benefits and costs of increasing CT capacity and utilization. European radiology 19(1):9–11 (2009); discussion 2. [DOI] [PubMed]
  • 6.Waganekar A, Sadasivan J, Prabhu AS, Harichandrakumar KT. Computed tomography profile and its utilization in head injury patients in emergency department: A prospective observational study. J. Emerg. Trauma Shock. 2018;11(1):25–30. doi: 10.4103/JETS.JETS_112_17. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Alfaro D, Levitt MA, English DK, Williams V, Eisenberg R. Accuracy of interpretation of cranial computed tomography scans in an emergency medicine residency program. Ann. Emerg. Med. 1995;25(2):169–174. doi: 10.1016/S0196-0644(95)70319-5. [DOI] [PubMed] [Google Scholar]
  • 8.Al-Reesi A, Stiell IG, Al-Zadjali N, Cwinn AA. Comparison of CT head interpretation between emergency physicians and neuroradiologists. Eur. J. Emerg. Med. 2010;17(5):280–282. doi: 10.1097/MEJ.0b013e32833483ed. [DOI] [PubMed] [Google Scholar]
  • 9.Liu X, Faes L, Kale AU, Wagner SK, Fu DJ, Bruynseels A, et al. A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: A systematic review and meta-analysis. The Lancet Digit. Health. 2019;1(6):e271–e297. doi: 10.1016/S2589-7500(19)30123-2. [DOI] [PubMed] [Google Scholar]
  • 10.Nir G, Karimi D, Goldenberg SL, Fazli L, Skinnider BF, Tavassoli P, et al. Comparison of artificial intelligence techniques to evaluate performance of a classifier for automatic grading of prostate cancer from digitized histopathologic images. JAMA Netw. Open. 2019;2(3):e190442. doi: 10.1001/jamanetworkopen.2019.0442. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Wong RK, Tandan V, De Silva S, Figueredo A. Pre-operative radiotherapy and curative surgery for the management of localized rectal carcinoma. Cochrane Database Syst. Rev. 2007 doi: 10.1002/14651858.CD002102.pub2. [DOI] [PubMed] [Google Scholar]
  • 12.Ting DSW, Cheung CY, Lim G, Tan GSW, Quang ND, Gan A, et al. Development and validation of a deep learning system for diabetic retinopathy and related eye diseases using retinal images from multiethnic populations with diabetes. JAMA. 2017;318(22):2211–2223. doi: 10.1001/jama.2017.18152. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Hssayeni MD, Croock MS, Salman AD, Al-khafaji HF, Yahya ZA, Ghoraani B. Intracranial hemorrhage segmentation using a deep convolutional model. Data. 2020;5(1):14. doi: 10.3390/data5010014. [DOI] [Google Scholar]
  • 14.Xiao F, Liao CC, Huang KC, Chiang IJ, Wong JM. Automated assessment of midline shift in head injury patients. Clin. Neurol. Neurosurg. 2010;112(9):785–790. doi: 10.1016/j.clineuro.2010.06.020. [DOI] [PubMed] [Google Scholar]
  • 15.Ye H, Gao F, Yin Y, Guo D, Zhao P, Lu Y, et al. Precise diagnosis of intracranial hemorrhage and subtypes using a three-dimensional joint convolutional and recurrent neural network. Eur. Radiol. 2019;29(11):6191–6201. doi: 10.1007/s00330-019-06163-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Arbabshirani MR, Fornwalt BK, Mongelluzzo GJ, Suever JD, Geise BD, Patel AA, et al. Advanced machine learning in action: Identification of intracranial hemorrhage on computed tomography scans of the head with clinical workflow integration. npj Digit. Med. 2018;1(1):9. doi: 10.1038/s41746-017-0015-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Majumdar, A., Brattain, L., Telfer, B., Farris, C., & Scalera, J. (eds) Detecting intracranial hemorrhage with deep learning. In 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC); 2018 18–21 July 2018. [DOI] [PubMed]
  • 18.Phaphuangwittayakul A, Guo Y, Ying F, Dawod AY, Angkurawaranon S, Angkurawaranon C. An optimal deep learning framework for multi-type hemorrhagic lesions detection and quantification in head CT images for traumatic brain injury. Appl. Intell. 2022;52(7):7320–7338. doi: 10.1007/s10489-021-02782-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Lee H, Yune S, Mansouri M, Kim M, Tajmir SH, Guerrier CE, et al. An explainable deep-learning algorithm for the detection of acute intracranial haemorrhage from small datasets. Nat. Biomed. Eng. 2019;3(3):173–182. doi: 10.1038/s41551-018-0324-9. [DOI] [PubMed] [Google Scholar]
  • 20.Cho J, Park KS, Karki M, Lee E, Ko S, Kim JK, et al. Improving sensitivity on identification and delineation of intracranial hemorrhage lesion using cascaded deep learning models. J. Digit. Imaging. 2019;32(3):450–461. doi: 10.1007/s10278-018-00172-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Chang PD, Kuoy E, Grinband J, Weinberg BD, Thompson M, Homo R, et al. Hybrid 3D/2D convolutional neural network for hemorrhage evaluation on head CT. AJNR Am. J. Neuroradiol. 2018;39(9):1609–1616. doi: 10.3174/ajnr.A5742. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Inkeaw P, Angkurawaranon S, Khumrin P, Inmutto N, Traisathit P, Chaijaruwanich J, et al. Automatic hemorrhage segmentation on head CT scan for traumatic brain injury using 3D deep learning model. Comput. Biol. Med. 2022;146:105530. doi: 10.1016/j.compbiomed.2022.105530. [DOI] [PubMed] [Google Scholar]
  • 23.Kamnitsas K, Ledig C, Newcombe VFJ, Simpson JP, Kane AD, Menon DK, et al. Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 2017;36:61–78. doi: 10.1016/j.media.2016.10.004. [DOI] [PubMed] [Google Scholar]
  • 24.Lee JY, Kim JS, Kim TY, Kim YS. Detection and classification of intracranial haemorrhage on CT images using a novel deep-learning algorithm. Sci. Rep. 2020;10(1):20546. doi: 10.1038/s41598-020-77441-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Kundisch A, Hönning A, Mutze S, Kreissl L, Spohn F, Lemcke J, et al. Deep learning algorithm in detecting intracranial hemorrhages on emergency computed tomographies. PLoS ONE. 2021;16(11):e0260560. doi: 10.1371/journal.pone.0260560. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Rao B, Zohrabian V, Cedeno P, Saha A, Pahade J, Davis MA. Utility of artificial intelligence tool as a prospective radiology peer reviewer—Detection of unreported intracranial hemorrhage. Acad. Radiol. 2021;28(1):85–93. doi: 10.1016/j.acra.2020.01.035. [DOI] [PubMed] [Google Scholar]
  • 27.Waite S, Scott J, Gale B, Fuchs T, Kolla S, Reede D. Interpretive error in radiology. Am. J. Roentgenol. 2017;208(4):739–749. doi: 10.2214/AJR.16.16963. [DOI] [PubMed] [Google Scholar]
  • 28.Bruno MA, Walker EA, Abujudeh HH. Understanding and confronting our mistakes: The epidemiology of error in radiology and strategies for error reduction. Radiographics. 2015;35(6):1668–1676. doi: 10.1148/rg.2015150023. [DOI] [PubMed] [Google Scholar]
  • 29.Hanna TN, Lamoureux C, Krupinski EA, Weber S, Johnson JO. Effect of shift, schedule, and volume on interpretive accuracy: A retrospective analysis of 2.9 million radiologic examinations. Radiology. 2018;287(1):205–212. doi: 10.1148/radiol.2017170555. [DOI] [PubMed] [Google Scholar]
  • 30.Ruutiainen AT, Durand DJ, Scanlon MH, Itri JN. Increased error rates in preliminary reports issued by radiology residents working more than 10 consecutive hours overnight. Acad. Radiol. 2013;20(3):305–311. doi: 10.1016/j.acra.2012.09.028. [DOI] [PubMed] [Google Scholar]
  • 31.Miyakoshi A, Nguyen QT, Cohen WA, Talner LB, Anzai Y. Accuracy of preliminary interpretation of neurologic CT examinations by on-call radiology residents and assessment of patient outcomes at a level I trauma center. J. Am. Coll. Radiol. 2009;6(12):864–870. doi: 10.1016/j.jacr.2009.07.021. [DOI] [PubMed] [Google Scholar]
  • 32.Briggs GM, Flynn PA, Worthington M, Rennie I, McKinstry CS. The role of specialist neuroradiology second opinion reporting: Is there added value? Clin. Radiol. 2008;63(7):791–795. doi: 10.1016/j.crad.2007.12.002. [DOI] [PubMed] [Google Scholar]
  • 33.O’Neill TJ, Xi Y, Stehel E, Browning T, Ng YS, Baker C, et al. Active reprioritization of the reading worklist using artificial intelligence has a beneficial effect on the turnaround time for interpretation of head CT with intracranial hemorrhage. Radiol. Artif. Intell. 2021;3(2):e200024. doi: 10.1148/ryai.2020200024. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Prevedello LM, Erdal BS, Ryu JL, Little KJ, Demirer M, Qian S, et al. Automated critical test findings identification and online notification system using artificial intelligence in imaging. Radiology. 2017;285(3):923–931. doi: 10.1148/radiol.2017162664. [DOI] [PubMed] [Google Scholar]
  • 35.Newgard CD, Hsia RY, Mann CN, Schmidt T, Sahni R, Bulger EM, et al. The trade-offs in field trauma triage: A multiregion assessment of accuracy metrics and volume shifts associated with different triage strategies. J. Trauma Acute Care Surg. 2013;74(5):1298–1306. doi: 10.1097/TA.0b013e31828b7848. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Jnawali, K., Arbabshirani, M.R., Rao, N. & Patel, A.A. (eds) Deep 3D Convolution Neural Network for CT Brain Hemorrhage Classification. Medical Imaging 2018: Computer-Aided Diagnosis; 2018 February 01, (2018).

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The datasets generated during and/or analysed during the current study are available from the corresponding author on reasonable request.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES