Skip to main content
BJA: British Journal of Anaesthesia logoLink to BJA: British Journal of Anaesthesia
. 2024 Mar 5;132(5):1049–1062. doi: 10.1016/j.bja.2024.01.036

Artificial intelligence for ultrasound scanning in regional anaesthesia: a scoping review of the evidence from multiple disciplines

James S Bowness 1,2,, David Metcalfe 3,4, Kariem El-Boghdadly 5,6, Neal Thurley 7, Megan Morecroft 8, Thomas Hartley 9, Joanna Krawczyk 2, J Alison Noble 10, Helen Higham 1,11
PMCID: PMC11103083  PMID: 38448269

Abstract

Background

Artificial intelligence (AI) for ultrasound scanning in regional anaesthesia is a rapidly developing interdisciplinary field. There is a risk that work could be undertaken in parallel by different elements of the community but with a lack of knowledge transfer between disciplines, leading to repetition and diverging methodologies. This scoping review aimed to identify and map the available literature on the accuracy and utility of AI systems for ultrasound scanning in regional anaesthesia.

Methods

A literature search was conducted using Medline, Embase, CINAHL, IEEE Xplore, and ACM Digital Library. Clinical trial registries, a registry of doctoral theses, regulatory authority databases, and websites of learned societies in the field were searched. Online commercial sources were also reviewed.

Results

In total, 13,014 sources were identified; 116 were included for full-text review. A marked change in AI techniques was noted in 2016–17, from which point on the predominant technique used was deep learning. Methods of evaluating accuracy are variable, meaning it is impossible to compare the performance of one model with another. Evaluations of utility are more comparable, but predominantly gained from the simulation setting with limited clinical data on efficacy or safety. Study methodology and reporting lack standardisation.

Conclusions

There is a lack of structure to the evaluation of accuracy and utility of AI for ultrasound scanning in regional anaesthesia, which hinders rigorous appraisal and clinical uptake. A framework for consistent evaluation is needed to inform model evaluation, allow comparison between approaches/models, and facilitate appropriate clinical adoption.

Keywords: artificial intelligence, evaluation, medical devices, regional anaesthesia, regulation, standardisation, ultrasound


Editor's key points.

  • Artificial intelligence (AI) for ultrasound scanning in regional anaesthesia crosses multiple disciplines, which risks limited knowledge transfer between specialty silos and unnecessary duplication of work.

  • This scoping review summarises the available evidence regarding the accuracy and utility of such technology. The literature shows that a standardised methodology for evaluation is needed, which can be adopted for academic, clinical, and regulatory purposes.

  • A structured and consistent approach, with cross-disciplinary collaboration, may enable a full understanding of AI systems in regional anaesthesia and inform clinical uptake.

Ultrasound imaging for regional anaesthesia, first described in 1989,1 is the principal method to guide the targeted delivery of local anaesthetic during peripheral nerve blockade (PNB).2,3 Abdallah and colleagues4 described its use as ‘the most important advance in regional anaesthesia practice of the new millennium’ as it has been shown to enhance both efficacy and safety3 through visualisation of the needle tip and key sono-anatomical structures. It is also used to image the spine and surrounding tissues ahead of performing central neuraxial blockade (CNB; spinal and epidural).5 As a dynamic imaging modality, the quality of the views obtained depends on the skill of the operator and the image generated by the machine.4 Whilst ultrasound machines have advanced considerably since the inception of ultrasound-guided regional anaesthesia (UGRA), it is unclear whether clinicians' knowledge and skill in utilising this technology has progressed to a similar extent.6

Artificial intelligence (AI) is a field of computer science which uses techniques that enable computers to undertake tasks associated with human intelligence.6 Recent publications present the case for AI to aid in the acquisition and interpretation of optimal ultrasound images for UGRA (PNB and CNB).6,7 However, this is an interdisciplinary field that draws on contributions from multiple areas, including computer science, engineering, and clinical medicine. As a consequence, evaluation of AI technologies for the identification of anatomical structures on ultrasound in UGRA can be heterogeneous. For example, proof-of-concept studies in preclinical experimental work7,8 typically involve different approaches to early clinically-orientated assessments of commercially-available systems.9, 10, 11, 12

Lack of knowledge transfer between disciplinary silos risks duplication of work and hindering the development, evaluation, and adoption of this technology. It is therefore important to synthesise the available literature into a single coherent summary, which identifies areas of strength and deficiency in current evaluation approaches to inform further work in this nascent and rapidly evolving field. This scoping review aims to identify and map the available evidence on the accuracy and utility of AI systems for anatomical structure identification on ultrasound in regional anaesthesia.

Methods

The protocol presented in this manuscript was developed using the guidance provided by the Joanna Briggs Institute (JBI) and Preferred Reporting Items for Systematic Reviews and Meta-Analyses—Extension for Scoping Reviews.13, 14, 15

Prior scoping reviews and registration

When checked at the commencement of this study, on May 4, 2022, no scoping or systematic reviews were registered with the JBI Database of Systematic Reviews and Implementation Reports, Cochrane Database of Systematic Reviews, or the International Prospective Register of Systematic Reviews (PROSPERO).

The objectives, inclusion criteria, and methods for this scoping review were specified in advance and documented in a protocol registered on the medRxiv preprint server for health sciences before data extraction from the identified sources.16

Scoping review question

The question addressed by this scoping review is, ‘what is the evidence supporting the use of artificial intelligence for ultrasound scanning in regional anaesthesia?’ As per published guidance,13,14 it is based on the PCC model, containing ‘population’ [human ultrasound scans], ‘concept’ [AI], and ‘context’ [UGRA] elements.

Inclusion criteria

Population

Sources were included if they pertained to human ultrasound scanning (living or cadaveric studies). Data from animal and bench studies were excluded.

Concept

Sources included relate to the accuracy of AI-based anatomical structure identification on real-time B-mode ultrasound and the utility of these systems with respect to clinical practice. For the purposes of this review, a broad definition of utility was adopted and may refer to anaesthetists' ultrasound scanning performance, patient outcomes (PNB efficacy or safety), delivery, cost-effectiveness of URGA, or both.

Context

Sources relating to ultrasound scanning in the context of CNB (e.g. spinal and epidural) and PNBs were included.

Types of sources

All publicly available data (including academic publications, clinical/specialist society information, and commercial product literature) were considered. For sources published in languages other than English, investigators contacted the author to request a translated version.

Search strategy

Published academic literature

The search strategy was developed by the lead investigator (JSB), reviewed/modified by the author team, and then executed by a medical librarian (NT) in conjunction with the lead investigator. The strategy was explicitly designed to consider sources spanning computer science, engineering, and clinical medicine, to capture the multidisciplinary contributions to this field.

Five databases were searched from inception to April 14, 2023; Medline (OvidSP; 1946 – present), Embase (OvidSP; 1974–present), CINAHL (EBSCO; 1981–present), IEEE Xplore (IEEE: 1988–present), ACM Digital Library (ACM; 1951–present). An original search was run on Medline and Embase only on March 3, 2023. This was superseded by the searches on April 14, 2023. The search was comprised of title/abstract keywords and subject headings for AI, ultrasonography, and anaesthesia. No date or language limits were applied. References were exported to EndNote 20 (Clarivate, London, UK) for de-duplication. The search terms and results for all databases are presented in Supplementary material A.

Other literature

As AI is an area of intense commercial interest, data were also sought from sources beyond the academic literature to minimise publication bias (Supplementary material A). These resources have variable search function capabilities so, to enable a consistent approach, these repositories were searched using the subject headings from the search strategy of this review and all combinations thereof (‘artificial intelligence; ultrasound; anaesthesia’).

Two authors (JSB and MM) jointly searched the International Committee of Medical Journal Editors (ICMJE)-approved clinical trial registries for eligible studies on May 12, 2023. As the protocols and data from EudraCT are available via the EU Clinical Trials Register, the latter was searched in its place. In addition, the World Health Organization (WHO) clinical trials registry platform was searched on the same date.

The Ethos online library of doctoral theses was also jointly searched by two authors (JSB and MM, May 12, 2023), as were regulatory authority registries and competent authority websites of North America (US Food & Drug Administration; MAUDE, 510k, De Novo, and Medical Device Recall databases) and the UK (Medicines and Healthcare products Regulatory Agency).

To review centralised information in the clinical domain, two investigators (JSB, July 1, 2022, and JK, July 11, 2023) reviewed websites of prominent, international learned societies in regional anaesthesia for material relating to AI in UGRA: African Society for Regional Anesthesia; American Society of Regional Anesthesia & Pain Medicine; Asian and Oceanic Society of Regional Anaesthesia and Pain Medicine; European Society of Regional Anaesthesia and Pain Therapy; Latin American Society of Regional Anesthesia; Regional Anaesthesia UK.

Material from seven commercial organisations with products in the field were retrieved through independent search of the company websites by two investigators (JSB and MM, May 24–August 13, 2023). The included companies were GE Healthcare (Chicago, IL, USA), HiCura (Singapore), Intelligent Ultrasound (Cardiff, UK), Mindray (Shenzhen, China), Rivanna Medical (Charlottesville, VA, USA), Samsung (Suwon, South Korea), and SmartAlpha (Ankara, Tukey).

Data extraction

Screening initial results

After removing duplicates, two investigators (JSB and MM) independently screened returned titles for inclusion in the study to determine whether they relate to AI, ultrasound, and regional anaesthesia. If the title was ambiguous, the source was included at this stage. In the event of disagreement between the two assessments, a third investigator (JK) reviewed to adjudicate.

Abstract review

One investigator (JSB) then reviewed the abstract/summary of all included sources for consideration of full-text review. Sources were retained if the abstract/summary related to AI, ultrasound, and regional anaesthesia (on human subjects).

Extraction of results

The full texts of all appropriate sources were reviewed (JSB) to ensure that each was still deemed suitable, then scrutinised to extract data on accuracy, utility, or both of the AI systems evaluated. The extracted data were electronically tabulated using a standardised data extraction form available in Supplementary material A. To ensure full scrutiny of each source, a software engineer (TH) initially reviewed all computer science/engineering papers. The lead investigator (JSB, clinician) then reviewed the extracted data from these sources along with the original publication, and assessed the clinically orientated sources, to aggregate all the information.

As formal assessment of study methodological quality is not recommended for scoping reviews,14 the investigators did not predefine what data would pertain to sono-anatomical structure identification accuracy or utility, or set any minimum outcome/reporting criteria. To maximise data capture, an inclusive approach was adopted when reviewing study items. As sources of commercial data may not include methodological reporting, this was not a prerequisite for inclusion.

Data were extracted from the included sources, and then the reference list for each was manually scrutinised for relevant cited literature (JSB). In addition, Google Scholar (Menlo Park, CA, USA) was used to identify pertinent citing literature (JSB). Further sources identified in this way were included for analysis.

Results

Selection of data sources

Figure 1 shows a flowchart of the search and screening process which details the number of sources considered at each stage. Overall, 116 sources were included for analysis; 84 pertaining to PNB, 30 to CNB, and two that addressed both areas. Where a source reported or summarised findings published elsewhere (e.g. review article or book chapter), data from the original source were used preferentially (n=93; 84 academic publications and nine commercial sources). Therefore, not all 116 sources have been cited in the manuscript, although additional sources are cited in Supplementary material B.

Fig 1.

Fig 1

PRISMA flowchart showing the number of sources considered at each stage of the search and screening process. PRISMA, Preferred Reporting Items for Systematic Reviews and Meta-Analyses.

Iterative analysis of included sources identified six recurring themes in the data: AI methodology in use; output of AI systems; methods of evaluating accuracy; methods of evaluating utility; standardisation of study methodology/reporting; information on commercially available systems.

Artificial intelligence methodology

Artificial intelligence and deep learning from 2016 to 2017

There is a notable change in the AI techniques used around 2016–17. Early approaches, most published before 2016–17, typically involved steps of feature extraction from the ultrasound image, followed by classification of these features, for both PNB17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28 and CNB.29, 30, 31, 32, 33, 34, 35, 36, 37 Around this point, there has been an increase in published data where the dominant technique was deep learning for both PNB9,11,12,38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73 and CNB,7,74, 75, 76, 77 though there are some exceptions.78, 79, 80 A number of sources do not explicitly state the techniques used; these are typically clinical case reports or (external) validation studies describing accuracy or efficacy81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93 and used alternative terms such as ‘automated’ or ‘intelligent’. However, these sources did identify the software in question (e.g. SpineNav3D software in the Accuro device, produced by Rivanna Medical). A full description of the AI, machine learning, and deep learning methodologies is provided in Supplementary material B.

Ground truth for model training/development

Ground truth refers to information known to be real and is used as the reference standard to train and test an AI technology. When training models in a supervised machine learning approach, raw data are presented alongside the ground truth and the model learns data patterns underlying this interpretation of the data. When evaluating models, the ground truth is used as a benchmark against which the models are assessed. It is therefore important to use a ground truth of the highest calibre to train models to the highest standard and assess them accurately.

In the available sources, a variety of human image interpretation was used as the ground truth. This included a range in the number of humans used in the ground truth from one,7,23, 24, 25,31,32,34, 35, 36,38,46,49,58,60,67,72,77 two,8,9,19,21,26,30,33,44,61 three,51, 52, 53,63,65 and up to 15.48 Some sources cited ‘multiple’ humans, used a plural term (e.g. experts), or simply did not state the number.17,18,20,25,27,29,54,62,64,66,69, 70, 71 Where multiple humans contributed to the ground truth, it was often unclear whether a single human assessed each image (with different humans assessing a different sample of the image dataset) or whether multiple humans assessed each image.

In addition to the varying number of humans contributing to the ground truth in model development, there was variation in the specialty and experience of those contributing. Sources described their ground truth human(s) as a volunteer,64 doctor/physician,48,52,63,65 sonographer,7,30, 31, 32, 33, 34, 35,60,75,77 clinical professional,27,38 or (regional) anaesthetist.8,17, 18, 19, 20, 21,23,25,26,36,44, 45, 46,49,51,53,54,58,61,62,67,69 The level of expertise of those providing the ground truth was described as trained,64 experienced,21,30, 31, 32, 33, 34,46,48,52,65 expert,7,8,17, 18, 19, 20,25, 26, 27,33,35,45,53,58,61,66,67,69,71,77 professional,38,49,51,62 specialist,63 or the level was unstated.23,36,54 However, sources did not typically define what was required to meet the definition of these levels and there is no universally recognised definition of ‘expert’ in regional anaesthesia.

For some sources, it is not clear how the ground truth for training data was obtained.9,11,12,28,29,39, 40, 41, 42, 43,45,47,50,55,57,59,60,68,70,72, 73, 74, 75, 76,78,81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91,93

Model output

Peripheral nerve blockade

The most common output for models processing PNB ultrasound images was segmentation (the outline/colour overlay of features in the image).9,11,12,19,20,22,23,25,27,36,38, 39, 40, 41, 42, 43,45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55,57, 58, 59, 60,63, 64, 65, 66, 67, 68, 69,71,72,78,79 Less common outputs included the use of bounding boxes to identify the location of the structure,8,17,18,26,61 attempting to locate the centre of a nerve and placing expanding circular rings from the centre,21 or placing name tags over structures in question.9,43,45,47,50

Alternative methods proposed to support ultrasound scanning included ‘enhancing’ the image acquired,56 applying a ‘scan success’ indicator to the image,9,43,45,47,50 or classifying ultrasound images into the region scanned70,73 or optimal vs non-optimal views.44 Additional functions included creating a three-dimensional (3D) model of the structure in question,24,28 path planning for needle insertion,20 and guidance on probe movement.22

Central neuraxial blockade

The most common output for models processing CNB ultrasound images was classification/identification of vertebral level, the intervertebral space, or both.30, 31, 32, 33, 34,75, 76, 77,81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91,93 Distance from the skin to the posterior complex/epidural space was the second most common output of models applied to CNB ultrasound.7,80,82, 83, 84, 85, 86, 87, 88,90, 91, 92, 93 Other functions included identifying the optimal needle insertion point,29,35,81,88,89,91 structure identification (via segmentation or bounding boxes),7,35,37 or 3D reconstruction of the spine.74

Accuracy evaluations

Peripheral nerve blockade

One approach to assessing accuracy of PNB model outputs involved assessing pixel-wise agreement between the ground truth and model output data, which was typically used for models that segment structures in the ultrasound image. The most common reporting metrics used to assess the overlap of enclosed areas (e.g. the outline of a blood vessel) were intersection over union (IoU; Jaccard similarity coefficient) and the Dice similarity coefficient.8,17, 18, 19, 20,23,25,27,28,38,42,46,49,51, 52, 53, 54, 55,57, 58, 59, 60,62, 63, 64, 65, 66,68,69,71,79 The Hausdorff metric/distance was used to compare the proximity of lines (e.g. a fascial plane).19,20,25,27,28,42,53,69 Whilst a pixel-wise approach clearly gives a detailed assessment of the proximity/coverage of one segmentation to another, there is no established threshold for this metric at which an AI system gives an ‘acceptable’ output. Furthermore, it gives little qualitative information on clinical relevance which is important as it may be more critical to accurately identify the exact border of a nerve or artery than a muscle.

An alternative approach takes an overall view of each whole structure by classifying the frequency of correct and incorrect prediction using true positive/negative and false positive/negative, accuracy, F-score, precision, recall, sensitivity, specificity, and area under the curve.18,19,39,44,54,55,58,59,61,67,69,72 Bowness and colleagues39 used the majority opinion from a panel of three experts to determine whether an AI structure identification was a true positive/negative etc; whilst this method would appear to incorporate clinical context, it is a pooled subjective opinion. The remaining publications used a frequency classification for each whole structure and based the decision of correct identification on a cut-off in the pixel-wise assessment. The definitions of correct prediction included IoU of >0.5,8,17,18,46,65 ≥25% pixel overlap,58,67 and subjective rating for a range of IoU values (e.g. 0.41–0.6=fairly good precision; 0.61–0.8=good precision),38 whilst others did not define the threshold of effective segmentation.54,55

Other methods of assessing/reporting accuracy were used infrequently, including subjective human assessment without providing data (‘performed as claimed’50 or ‘the identification of regional anatomy and scanning were successful’47) and a subjective 1–5 Likert scale rating of accuracy.9

It is important to note that ultrasound scans used for the assessments were collected on different subjects, using different machines, and in different body regions/structures. A limited number of accuracy assessments used multiple structure classes (e.g. nerve, artery, bone) over multiple body regions,9,39,42 whilst some covered multiple structure classes in a single region (e.g. supraclavicular level brachial plexus or thoracic paravertebral block).22,23,27,44,46,49,53,58,62,67,78 Other sources used a single structure class (e.g. nerve) from multiple regions,8,36,54 or a single structure in a single region, such as the median nerve in the forearm,17, 18, 19, 20,61,69 the femoral artery,28,72 the femoral nerve,38,65 or the sciatic nerve.21,25,48,57 Finally, some studies did not fully define the region/structures being assessed; often simply naming the ‘brachial plexus’,51,52,55,59,60,63,64,66,68,71,79 or not stating the region(s) or structure(s) at all.47,50

Ultimately, studies failed to consistently use the same metrics to compare segmentation of the same structures on the same ultrasound scans. This makes it very difficult for clinicians to effectively compare performance of one model/approach to others.

Central neuraxial blockade

Assessment of accuracy for CNB models was more aligned than the approaches for assessing PNB model accuracy, which involved a greater element of human judgement/interpretation. The most common method was correlating the distance from skin to posterior complex/epidural space as predicted by the model to that observed on manual interpretation of ultrasound,32,34,84,88,93 depth of loss of resistance during needle insertion,80,82,85,86 or both.37,83,92 All reported good correlation between human depth assessment and AI depth assessment on the ultrasound image, and loss of resistance during needle insertion to be deeper than that calculated on ultrasound (by human or AI).

Accuracy in determining the level of the intervertebral space in view was assessed by five studies. All report ≥84% agreement in AI and human identification of vertebral level based on ultrasound interpretation,30,34,75, 76, 77 whereas a single study reported 70% agreement between palpation and AI-ultrasound.75 Similarly, four studies reported ≥93% accuracy in AI identification of lumbar ultrasound images showing an intervertebral space vs a spinous process, though this approach did not determine which level was in view.29,31, 32, 33

The ultrasound body regions used for CNB approaches were more consistent than for PNB. Nineteen sources used ultrasound scans of the lumbar spine7,29, 30, 31, 32, 33, 34, 35,37,75, 76, 77,80,83,84,86,88,90,92,93 and two utilised the thoracolumbar spine,74,82 whereas one did not specifically state which region was seen (though use of the lumbar region can be deduced as it is a case report of a patient undergoing total knee arthroplasty).85

Ground truth for accuracy assessments

Many studies reported results of testing or internal validation in which a proportion of the initial dataset was partitioned from the training data and used to assess models at the end of training. External validation is the subsequent assessment of these models, using data collected at a different time point, location, population, or all. It is used to further evaluate model performance and assess generalisability to the wider population (or different populations). Only one source explicitly stated it was an external validation study.39

Of the sources reporting accuracy data for PNB models, seven described external validation studies. The standard (ground truth) used for these studies was described as a single ‘expert anaesthetist’,22 two ‘regional anaesthesia experts’,20 ‘an experienced anaesthetist and a radiologist’,9 or ‘three UGRA experts’.39 One study compared AI model performance to that of 19 ‘experts’.42 It is not stated in/clear from two sources who was providing the gold standard (ground truth) for these assessments.47,50

Twelve sources for CNB models described external validation studies of accuracy. The gold standard (ground truth) for these studies were an anaesthetist,29,30,80,82, 83, 84,86,92 clinician,88,93 ‘provider experienced with freehand ultrasound’,75 or is not stated.85 Their level of expertise was described as experienced,30,88,93 novice and senior/expert,82,84 resident/fellow/attending,83,86,92 or unstated.29,75,80,85 The number of humans contributing to this gold standard (ground truth) was one,30,75,83,86,88,92,93 two,84 or unstated.29,85

Utility evaluations

Peripheral nerve blockade

Utility evaluations for PNB models can be broadly divided into assessments based on subjective opinion (n =10) or on clinical outcomes (n=4).

In a non-clinical or simulation environment, Bowness and colleagues11 have published a number of studies in which a panel of three UGRA experts rated the AI colour overlay on recorded ultrasound videos (7.87–8.69/10); assessed that it would be helpful in determining the correct view and identifying anatomical structures of interest in >99% of cases11; determined the accuracy rate of structure identification to be 93.5%39; and reduced the risk of unwanted needle-induced trauma or block failure in 62.9–86.4% of cases.39 In other studies this group has sought feedback from 30 users to determine use of this system for UGRA experts (teaching) and non-experts (learning/clinical practice),12 and assessed 21 UGRA non-experts to demonstrate superior scanning performance (on volunteers) with the use of an AI colour overlay on real-time ultrasound.40 Cai and colleagues48 demonstrated higher self-rating and expert assessment scores amongst 40 anaesthetic trainees when performing PNBs after teaching with an AI-based nerve identification system, whereas Erdem and colleagues43 describe that 40 anaesthetists reported reduced chance of complications, increased chance of block success, and that the system was useful as a training tool in UGRA. Other studies report that anaesthetists described the assistive AI technology to be helpful in the identification of key anatomical structures and in training,45 and that all non-experts in UGRA considered assistive AI to be a desirable and useful tool.24 Other published subjective opinion includes a 20–30% reduction in time required for UGRA procedures and a reduction in needle pass attempts (with no underlying data provided in this source),50 and a report which proposed that assistive AI may improve block success, reduce the number of needle attempts, improve the satisfaction of patient/practitioner, and reduce the volume of local anaesthetic required (again without providing underlying data).47

Clinical assessments include a secondary analysis of data collected during a service evaluation by Bowness and colleagues,41 who reported an increase in the delivery of UGRA to trauma patients when assistive AI for ultrasound scanning was available. In 28 days before the introduction of the AI system, 71 PNBs were performed amongst 207 eligible cases, compared with 93 for 193 eligible cases after (P=0.036)—with no observed decrease in efficacy or increase in complication rate.41 Other clinical outcome data for PNB systems report that UGRA performed with AI-enhanced ultrasound imaging achieved a quicker performance of UGRA, superior motor block, and better analgesia in scapula fracture surgery.56 Cai and colleagues48 report reduced paraesthesia during the first month of performing sciatic nerve blocks after teaching with an AI-based nerve identification system, though there was no difference in the rate of pain during injection, perforation of blood vessels or block success. Wang and colleagues78 report reduced postoperative analgesic requirement and cognitive dysfunction/delirium after AI-assisted lumbosacral plexus block compared with systemic analgesia for hip arthroplasty surgery.

The above sources again report evaluations for PNB for varying body regions. Some studies covered PNB for three or more body regions,11,12,39, 40, 41,43 whereas one source covered the supraclavicular and adductor canal blocks,45 three covered a single PNB region (lumbar plexus,78 ‘scapula region nerve block’,56 and the sciatic nerve48), and two did not state the PNB region assessed.47,50

Central neuraxial blockade

Utility evaluations for CNB models can be broadly divided into assessments based on technical success of the procedure (n=13) or on clinical outcomes (n=4).

Technical success criteria included reporting needle insertion attempts or first pass success; three sources reported fewer attempts/higher success rates,80,90,91 two reported no statistically significant difference,87,89 and three sources simply reported rate of first pass success using AI assistance (87.23%, 79.1%, and 92.0%, respectively) without comparison to no AI assistance.86,88,93 Two sources were case reports of successful epidural catheter insertion in patients living with obesity.81,85 A second criterion for technical success was the number of needle directs, with three sources reporting fewer redirects were required with AI assistance.82,90,91 As with utility evaluations of PNB models, time taken for the procedure/to identify the appropriate view (intervertebral space) was assessed in a number of sources. This showed variable results, with some reporting increased time,87,91 no difference,82,89 a mixed picture,75,80 or reduced time required.90 Two sources report the time taken for the procedure, but did not provide a comparison with an alternative technique (28 s and 31 s, respectively).30,85

In clinical outcome studies, no significant difference in patient satisfaction was associated with use of AI-assisted ultrasound scanning for CNB,87 or failed block and adverse events,91 though a lower incidence of paraesthesia and lower back pain have been reported elsewhere.80,90

As might be expected, the body region used to evaluate AI assistance for ultrasound scanning in CNB was more consistent than for PNB. Eleven sources report using ultrasound scans of the lumbar spine30,75,80,81,86, 87, 88, 89, 90, 91,93 and one utilised the thoracolumbar spine,82 whilst one did not specifically state which region was seen though use of the lumbar region can be deduced as it is a case report of a patient undergoing total knee arthroplasty.85

Standardisation of study methodology and reporting

An additional limitation to comparing evaluations of this technologies is that reporting varies considerably; only 32/93 (34.4%) reviewed sources reported demographic data for the subjects assessed in these studies.7,9,12,21,22,29,30,35,37, 38, 39, 40, 41, 42,44,46,52,53,74,75,78,80, 81, 82, 83, 84, 85,88,90, 91, 92, 93 This is important as demographic factors such as age (muscle atrophy) and obesity affect the appearance of anatomical structures on the ultrasound image.94 These descriptions were themselves inconsistent; such as reporting age and gender, but not BMI,78 or reporting only mean age/BMI without ranges.74 Importantly, older and obese patients were underrepresented with only 18/32 (56.3%) sources reporting demographic data which included subjects aged >60 yr, BMI ≥30 kg m−2, or both.12,37, 38, 39, 40, 41, 42,52,80,81,83, 84, 85, 86,88,90, 91, 92

Models for CNB ultrasound were most commonly assessed in the obstetric population, with 14 sources reporting data in this cohort.31, 32, 33, 34,37,75,81,83,84,86,88,90,92,93 Despite CNB being a common technique throughout anaesthesia, only four sources included non-obstetric patients,80,82,85,91 whilst four utilised healthy volunteers,29,30,35,77 and three sources did not identify the subject population being scanned.7,74,76

Only six sources reported their data according to recognised reporting standards,40,42,82,86,88,92 including the CONSORT-AI,95 CONSORT,96 DECIDE-AI,97 and STROBE98 guidelines. Of these, only CONSORT-AI and DECIDE-AI are AI-specific. As they were published in 2020 and 2022, respectively, time will tell whether future studies are reported in a more consistent manner and in line with these guidelines.

Finally, it is good practice in the field of AI to make training and testing/validation data publicly available.99, 100, 101 However, only 15 sources used publicly available data or shared their data when publishing.48,49,51,55,59, 60, 61,63, 64, 65, 66,68,71,74,79

Commercially available systems

Seven AI systems are commercially available to support ultrasound scanning in UGRA, five for PNB and two for CNB (Table 1).

Table 1.

Commercial systems and accompanying information. AI, artificial intelligence; CNB, central neuraxial blockade; PNB, peripheral nerve blockade. ∗Company approached for information on regulatory approval in UK and EU but not provided.

System name
Manufacturer
Approved as a medical device Description Company claims Sources of data on performance
cNerve
GE Healthcare (Chicago, IL, USA)
UK/EU (Unknown∗)
USA (June 2022)
Colour overlay segmentation of interscalene to supraclavicular-level brachial plexus, femoral, and popliteal-level sciatic nerve ‘Helps detect and track nerves in 99% of cases’ Commercial102
Nerveblox
SmartAlpha (Ankara, Turkey)
UK/EU (May 2021) Colour overlay of blood vessels, bone, muscles, nerves, and fascia/serosa for 12 PNBs No specific claims regarding accuracy
‘ … gives anesthesiologists extra confidence … ’
‘ … help anesthesiologists practice PNB faster.’
Commercial105
Academic9,43,47,50
NerveTrack
Samsung (Suwon, South Korea)
UK/EU (February 2021)
USA (May 2021)
Bounding box around median and ulnar nerves in the forearm ‘ … can detect the median and ulnar nerve with reasonable accuracy … ’
‘ … efficiency of nerve detection … ’ is claimed to be 4.4/5 with vs 3.8/5 without (P<0.0001)
‘ … reduce scanning time significantly, from 24.7 s to 8.2 s.’
Commercial10,104
ScanNav Anatomy Peripheral Nerve Block
Intelligent Ultrasound (Cardiff, UK)
UK/EU (April 2021)
USA (October 2022)
Colour overlay of blood vessels, bone, muscles, nerves, and fascia/serosa for 10 PNBs ‘ … enhance the accuracy and standardization of ultrasound image interpretation … ’
‘ … help tip the balance of safety and confidence in favour of performing regional anaesthesia.’
Commercial106
Academic6,11,12,39, 40, 41
Smart Nerve
Mindray (Shenzhen, China)
UK/EU (Unknown∗)
USA (November 2021)
Colour overlay segmentation of interscalene to supraclavicular-level brachial plexus nerves No specific claims regarding accuracy
‘ … automatically recognise the brachial plexus and highlight the nerve, increasing clinical confidence and reducing procedure time … ’
Commercial103
Accuro
Rivanna Medical (Charlottesville, VA, USA)
UK/EU (June 2016)
USA (March 2014)
Provides guidance for CNB (spinal and epidural) in the thoracic and lumbar regions of the spine using ‘AI-enabled SpineNav3D image recognition’.
Software identifies intervertebral space on ultrasound image and hardware identifies point of needle insertion on the skin.
‘Improve the safety, speed, and efficiency of epidural and spinal anesthesia.’
‘ … accurately identifies the epidural location with success rates exceeding 94%, Accuro gives you confidence … ’
‘Clinically proven to:
- Increase first attempt success
- Reduce needle passes
- Reduce placement times
- Significantly increase patient satisfaction and pain control’
‘Collectively, these benefits significantly reduce the cost of care’
Commercial107,108
Academic81, 82, 83, 84, 85, 86, 87,89, 90, 91, 92
uSine
HiCura Medical (Singapore)
UK/EU (August 2023)
Singapore (July 2022)
Additional screen labels spinous processes and intervertebral spaces on lumbar ultrasound scan, and measures epidural depth
‘ … machine learning algorithm allows automatic identification of spinal landmarks during ultrasound scan.’
‘ … alerts anaesthetist in real-time when the right location and right angle are reached.’
‘ … achieved very high [92%] first-attempt puncture success rate.’
‘ … safe and effective and promotes a reduction of procedural time, better clinical outcomes and improves patient satisfaction.’
Commercial109
Academic88,93

Peripheral nerve blockade

cNerve (GE Healthcare)102 and Smart Nerve (Mindray)103 are systems incorporated into the manufacturers' ultrasound machines, which segment peripheral nerves to produce a (yellow) colour overlay over them. NerveTrack (Samsung Medison, Suwon, South Korea)10,104 is also integrated into the manufacturer's machine, but produces bounding boxes around the median and ulnar nerves of the forearm. Nerveblox (SmartAlpha)105 and ScanNav Anatomy Peripheral Nerve Block (Intelligent Ultrasound)106 are external devices, which can be connected to an ultrasound machine to provide an additional display with an AI-generated colour overlay of key structures for a number of PNBs. Numerous studies have been published on the latter device,12,39, 40, 41, 42 all of which are included in this review.

Central neuraxial blockade

Accuro (Rivanna Medical) is a handheld device with an integrated ultrasound machine,107,108 whereas uSine (HiCura) is an external device which is connected to the ultrasound machine.109 Both systems aim to support identification of the vertebral level, intervertebral space, and depth to target. The former system incorporates hardware to mark the patient's skin with the optimal needle insertion site.

Conflicts of interests

As in any field, declaring sources of funding and conflicts of interest is essential to transparent reporting of research. Work in AI has a potential for commercial value, thus it is particularly pertinent in this setting. Of the 23 academic studies which directly relate to a commercial product (see Table 1), 10 declare funding for research (or in-kind support), industry affiliations, or both in the conflict of interests statement.6,11,12,39, 40, 41,86,90, 91, 92 Two report these affiliations in the limitations section of the discussion.39,40

Discussion

AI in healthcare is an area of intense and growing interest;110 the worldwide AI healthcare market projected to reach $200 billion by 2030111 and the number of regulatory approvals for AI-based medical devices is steadily increasing.112 Anaesthesia is a data-rich specialty, with a heavy reliance on technology, but there are relatively few AI devices approved for use in this field of medicine compared with others such as radiology and cardiology.113

This scoping review has surveyed literature across multiple disciplines and identified key findings. Firstly, the dominant technique is deep learning, which is used in virtually all publications since 2017. In addition, model outputs show consistency; most commonly sono-anatomical structure segmentation for PNB or identification of intervertebral space and depth to target for CNB. Secondly, and conversely, methods for evaluating accuracy and utility originate in multiple disciplines and show gross heterogeneity with a notable lack of clinical data. Thirdly, reporting is inconsistent amongst academic publications and data on device performance information from commercial organisations is scarce—often produced by clinicians in subsequent studies rather than made available by the companies from their regulatory approval studies.

It is possible that heterogenous and unstructured research literature remains a barrier to AI being successfully incorporated into clinical anaesthetic practice. The uniformity in AI techniques now used, and consistency in model outputs, should facilitate further standardisation. This will engender greater understanding of this technology and support appropriate adoption in clinical practice. Despite the rapid growth in the field, this should provide cause for optimism that appropriate structure can be implemented successfully.

However, this is an interdisciplinary field, and the discordant approaches of clinical medicine, technology, and industry are striking. This is evident in the anatomical structures that groups/companies chose to identify, how models are trained (e.g. source of ground truth), how accuracy is validated (e.g. metric used), and how utility is assessed. In addition, there is still a paucity of data on patient outcomes and institutional impact. Standardisation has become a common feature of UGRA.114, 115, 116, 117 Whilst initial efforts have been made to address this in AI,118,119 it has not yet been realised and the authors believe that this limitation will continue to impede clinical implementation of AI in UGRA until it is addressed.

Enthusiasm to share novel findings in this exciting and fast-moving field is understandable, but differences in reporting also hinder reliable comparison of models/systems and limit a true understanding of the state of the art. Furthermore, as companies are not incentivised to share data from their regulatory filings, there is variability in approaches to regulatory approval and a lack of understanding of the relative performance of approved devices.

We propose that a structured framework be developed for validating accuracy and clinical utility of AI which assists ultrasound scanning in UGRA (Fig. 2). This should include which sono-anatomical structures systems identify for specific procedures, a standardised method for both pixel-wise and clinical assessment of accuracy, clearly identified measures of clinical utility, demonstration of clinical impact, and standardised reporting.120 Clinical bodies should lead the development of this framework, to ensure clinically relevant pain points are addressed, and ensure contribution from technology experts, clinical medicine, industry and regulatory bodies, and patient and public involvement. Developing an open-access dataset of representative ultrasound scans for all parties to utilise in validation studies will be a key element to facilitating standardised evaluation. Industry should be encouraged to utilise this and publish their data. Standardised reporting of these evaluations will further aid dissemination of knowledge, aided by a non-critical view of declared potential conflicts of interest.

Fig 2.

Fig 2

Conceptual overview of a structured framework for the validation of accuracy and clinical utility of AI technology which assists ultrasound scanning in UGRA. AI, artificial intelligence; UGRA, ultrasound-guided regional anaesthesia.

Ultrasound revolutionised UGRA in an unstructured manner has led to great progress, but barriers in the field remain as some techniques to assist with ultrasound image interpretation remain unvalidated (e.g. hydrolocation).4 Recent initiatives are attempting to retrospectively provide structure.121 AI will revolutionise medicine,110 and potentially UGRA; adopting this technology in a structured manner will provide the optimal opportunities to harness its full potential.

Conclusions

This scoping review has identified gross heterogeneity and poor reporting across the literature pertaining to AI for the identification of anatomical structures on ultrasound in regional anaesthesia. This is an important barrier to developing the field and implementing AI technologies within clinical anaesthetic practice. The situation can only be improved by clinicians, scientists, and industry working together to standardise our approach to understanding these systems, to optimise use in ultrasound-guided regional anaesthesia.

Authors’ contributions

Study concept and design: JSB, DM, KEB, HH

Design and conduct of literature search: JSB, NT, MM

Review of search results and search of grey literature: JSB, MM, JK

Data extraction: JSB, TH

Manuscript preparation and editing: JSB

Manuscript review & approval: all authors

Declarations of interest

JSB is a senior clinical advisor for Intelligent Ultrasound, receiving research funding and honoraria. KE has received research, honoraria, and educational funding from Fisher and Paykel Healthcare, GE Healthcare, PAION, and Ambu. MM was formerly and TH is currently employed by Intelligent Ultrasound. JAN is a senior scientific advisor for Intelligent Ultrasound.

Handling Editor: Jonathan Hardman

Footnotes

Appendix A

Supplementary data to this article can be found online at https://doi.org/10.1016/j.bja.2024.01.036.

Appendix A. Supplementary data

The following are the Supplementary data to this article:

Multimedia component 1
mmc1.docx (48.8KB, docx)
Multimedia component 2
mmc2.docx (74.4KB, docx)

Figure S1.

Figure S1

Figure S2.

Figure S2

References

  • 1.Ting P.L., Sivagnanaratnam V. Ultrasonographic study of the spread of local anaesthetic during axillary brachial plexus block. Br J Anaesth. 1989;63:326–329. doi: 10.1093/bja/63.3.326. [DOI] [PubMed] [Google Scholar]
  • 2.Helen L., O'Donnell B.D., Moore E. Nerve localization techniques for peripheral nerve block and possible future directions. Acta Anaesthesiol Scand. 2015;59:962–974. doi: 10.1111/aas.12544. [DOI] [PubMed] [Google Scholar]
  • 3.Munirama S., McLeod G. A systematic review and meta-analysis of ultrasound versus electrical stimulation for peripheral nerve location and blockade. Anaesthesia. 2015;70:1084–1091. doi: 10.1111/anae.13098. [DOI] [PubMed] [Google Scholar]
  • 4.Abdallah F.W., Macfarlane A.J., Brull R. The Requisites of needle-to-nerve proximity for ultrasound-guided regional anesthesia: a scoping review of the evidence. Reg Anesth Pain Med. 2016;41:221–228. doi: 10.1097/AAP.0000000000000201. [DOI] [PubMed] [Google Scholar]
  • 5.Kalagara H., Nair H., Kolli S., Thota G., Uppal V. Ultrasound imaging of the spine for central neuraxial blockade: a technical description and evidence update. Curr Anesthesiol Rep. 2021;11:326–339. [Google Scholar]
  • 6.Bowness J., El-Boghdadly K., Burckett-St Laurent D. Artificial intelligence for image interpretation in ultrasound-guided regional anaesthesia. Anaesthesia. 2021;76:602–607. doi: 10.1111/anae.15212. [DOI] [PubMed] [Google Scholar]
  • 7.Pesteie M., Lessoway V., Abolmaesumi P., Rohling R.N. Automatic localization of the needle target for ultrasound-guided epidural injections. IEEE Trans Med Imaging. 2018;37:81–92. doi: 10.1109/TMI.2017.2739110. [DOI] [PubMed] [Google Scholar]
  • 8.Alkhatib M., Hafiane A., Vieyres P., Delbos A. Deep visual nerve tracking in ultrasound images. Comput Med Imaging Graph. 2019;76 doi: 10.1016/j.compmedimag.2019.05.007. [DOI] [PubMed] [Google Scholar]
  • 9.Gungor I., Gunaydin B., Oktar S.O., et al. A real-time anatomy identification via tool based on artificial intelligence for ultrasound-guided peripheral nerve block procedures: an accuracy study. J Anesth. 2021;35:591–594. doi: 10.1007/s00540-021-02947-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Intel. Solution Brief - NerveTrackTM. Real-Time AI-based Technology for Nerve Detection. Available from https://device.report/m/a7ad9623877f1e7e05c30a82ef456bfa6d3a988068863bfc28f78141f451f369.pdf (Accessed 29 February 2024).
  • 11.Bowness J., Varsou O., Turbitt L., Burkett-St Laurent D. Identifying anatomical structures on ultrasound: assistive artificial intelligence in ultrasound-guided regional anesthesia. Clin Anat. 2021;34:802–809. doi: 10.1002/ca.23742. [DOI] [PubMed] [Google Scholar]
  • 12.Bowness J.S., El-Boghdadly K., Woodworth G., Noble J.A., Higham H., Burckett-St Laurent D. Exploring the utility of assistive artificial intelligence for ultrasound scanning in regional anesthesia. Reg Anesth Pain Med. 2022;47:375–379. doi: 10.1136/rapm-2021-103368. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Peters M.D., Godfrey C.M., Khalil H., McInerney P., Parker D., Soares C.B. Guidance for conducting systematic scoping reviews. Int J Evid Based Healthc. 2015;13:141–146. doi: 10.1097/XEB.0000000000000050. [DOI] [PubMed] [Google Scholar]
  • 14.Peters M.D.J., Godfrey C., McInerney P., Baldini Soares C., Khalil H., Parker D. In: Aromataris E., Munn Z., editors. Joanna Briggs Institute Reviewer’s Manual; Adelaide, Australia: 2017. (Chapter 11: scoping reviews). [Google Scholar]
  • 15.Tricco A.C., Lillie E., Zarin W., et al. PRISMA extension for scoping reviews (PRISMA-ScR): checklist and explanation. Ann Intern Med. 2018;169:467–473. doi: 10.7326/M18-0850. [DOI] [PubMed] [Google Scholar]
  • 16.Bowness J., Metcalfe D., El-Boghdadly K., et al. Artificial intelligence for anatomical structure identification on ultrasound in regional anaesthesia: a scoping review protocol. medRxiv 2023.07.04.23291560. 2023 doi: 10.1101/2023.07.04.23291560. [DOI] [Google Scholar]
  • 17.Hadjerci O., Hafiane A., Makris P., Conte D., Vieyres P., Delbos A. In: Image Analysis and Processing - International Conference on Image Analysis and Processing 2015. Murino V., Puppo E., editors. Vol. 9279. Springer; Cham: 2015. Nerve localization by machine learning framework with new feature selection algorithm; pp. 246–256. (Lecture notes in computer science Vol.). [Google Scholar]
  • 18.Hadjerci O., Hafiane A., Conte D., Makris P., Vieyres P., Delbos A. IEEE International Conference on Image Processing (ICIP) 2015. Québec; 2015. Ultrasound median nerve localization by classification based on despeckle filtering and feature selection; pp. 4155–4159. [Google Scholar]
  • 19.Hadjerci O., Hafiane A., Vieyres P., Conte D., Makris P., Delbos A. Computer-aided detection system for nerve identification using ultrasound images: a comparative study. Inform Med Unlocked. 2016;3:29–43. [Google Scholar]
  • 20.Hadjerci O., Hafiane A., Morette N., Novales C., Vieyres P., Delbos A. Assistive system based on nerve detection and needle navigation in ultrasound images for regional anesthesia. Expert Syst Appl. 2016;61:64–77. [Google Scholar]
  • 21.Philippona C., Morse J., Wehbe M., et al. 2014 IEEE 27th international Symposium on Computer-Based medical Systems (CBMS) 2014. NerveGPS: a novel decision support system for ultrasound nerve block guidance; pp. 18–21. New York. [Google Scholar]
  • 22.Smistad E., Iversen D.H., Leidig L., Lervik Bakeng J.B., Johansen K.F., Lindseth F. Automatic segmentation and probe guidance for real-time assistance of ultrasound-guided femoral nerve blocks. Ultrasound Med Biol. 2017;43:218–226. doi: 10.1016/j.ultrasmedbio.2016.08.036. [DOI] [PubMed] [Google Scholar]
  • 23.González J.G., Alvarez M.A., Orozco A.A. 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC) Milano; 2015. Automatic segmentation of nerve structures in ultrasound images using Graph Cuts and Gaussian processes; pp. 3089–3092. [DOI] [PubMed] [Google Scholar]
  • 24.O’Donnell B., Burns C., Lindseth F., et al. Proceedings of 35th Annual ESRA Congress. Innsbruck; 2016. Preliminary user evaluation of a novel software tool for ultrasonographic anatomy visualization. [Google Scholar]
  • 25.Hafiane A., Vieyres P., Delbos A. Phase-based probabilistic active contour for nerve detection in ultrasound images for regional anesthesia. Comput Biol Med. 2014;52:88–95. doi: 10.1016/j.compbiomed.2014.06.001. [DOI] [PubMed] [Google Scholar]
  • 26.Alkhatib M., Hafiane A., Tahri O., Vieyres P., Delbos A. Adaptive median binary patterns for fully automatic nerves tracking in ultrasound images. Comput Methods Programs Biomed. 2018;160:129–140. doi: 10.1016/j.cmpb.2018.03.013. [DOI] [PubMed] [Google Scholar]
  • 27.Lindseth F., Leidig L., Smistad E. Eurographics Workshop on Visual Computing for Biomedicine Chester. 2015. Automatic real-time annotation of important landmarks in ultrasound-guided femoral nerve blocks. [Google Scholar]
  • 28.Smistad E., Lindseth F. Real-time automatic artery segmentation, reconstruction and registration for ultrasound-guided regional anaesthesia of the femoral nerve. IEEE Trans Med Imaging. 2016;35:752–761. doi: 10.1109/TMI.2015.2494160. [DOI] [PubMed] [Google Scholar]
  • 29.Yusong L., Shuang Y., Kok Kiong T., Tildsley P., Sia A.T., Ban Leong S. Development of a real-time lumbar ultrasound image processing system for epidural needle entry site localization. Annu Int Conf IEEE Eng Med Biol Soc. 2016;2016:4093–4096. doi: 10.1109/EMBC.2016.7591626. [DOI] [PubMed] [Google Scholar]
  • 30.Ikhsan M., Kok Kiong T., Ting Ting O., Lew J.P., Ban Leong S. Gabor-based automatic spinal level identification in ultrasound. Annu Int Conf IEEE Eng Med Biol Soc. 2017;2017:3146–3149. doi: 10.1109/EMBC.2017.8037524. [DOI] [PubMed] [Google Scholar]
  • 31.Yu S., Tan K.K. Simulated Evolution and Learning: 10th International Conference. 2014. Classification of lumbar ultrasound images with machine learning; pp. 287–298. [Google Scholar]
  • 32.Yu S., Tan K.K., Sng B.L., Li S., Sia A.T. Automatic identification of needle insertion site in epidural anesthesia with a cascading classifier. Ultrasound Med Biol. 2014;40:1980–1990. doi: 10.1016/j.ultrasmedbio.2014.03.010. [DOI] [PubMed] [Google Scholar]
  • 33.Yu S., Tan K.K., Sng B.L., Li S., Sia A.T. Lumbar ultrasound image feature extraction and classification with support vector machine. Ultrasound Med Biol. 2015;41:2677–2689. doi: 10.1016/j.ultrasmedbio.2015.05.015. [DOI] [PubMed] [Google Scholar]
  • 34.Yu S., Tan K., Sng B., Li S., Sia A. 2015 IEEE 12th International Symposium on Biomedical Imaging (ISBI) 2015. Real-time automatic spinal level identification with ultrasound image processing; pp. 243–246. Brooklyn, NY. [Google Scholar]
  • 35.Pesteie M., Abolmaesumi P., Ashab H.A., et al. Real-time ultrasound image classification for spine anesthesia using local directional Hadamard features. Int J Comput Assist Radiol Surg. 2015;10:901–912. doi: 10.1007/s11548-015-1202-5. [DOI] [PubMed] [Google Scholar]
  • 36.Giraldo J.J., Álvarez M.A., Orozco Á.A. Peripheral nerve segmentation using Nonparametric Bayesian Hierarchical Clustering. Annu Int Conf IEEE Eng Med Biol Soc. 2015;2015:3101–3104. doi: 10.1109/EMBC.2015.7319048. [DOI] [PubMed] [Google Scholar]
  • 37.Tran D., Rohling R.N. Automatic detection of lumbar anatomy in ultrasound images of human subjects. IEEE Trans Biomed Eng. 2010;57:2248–2256. doi: 10.1109/TBME.2010.2048709. [DOI] [PubMed] [Google Scholar]
  • 38.Berggreen J., Johansson A., Jahr J., Möller S., Jansson T. Deep learning on ultrasound images visualizes the femoral nerve with good precision. Healthcare (Basel) 2023;11:184. doi: 10.3390/healthcare11020184. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Bowness J.S., Burckett-St Laurent D., Hernandez N., et al. Assistive artificial intelligence for ultrasound image interpretation in regional anaesthesia: an external validation study. Br J Anaesth. 2023;130:217–225. doi: 10.1016/j.bja.2022.06.031. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Bowness J.S., Macfarlane A.J.R., Burckett-St Laurent D., et al. Evaluation of the impact of assistive artificial intelligence on ultrasound scanning for regional anaesthesia. Br J Anaesth. 2023;130:226–233. doi: 10.1016/j.bja.2022.07.049. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Bowness J.S., James K., Yarlett L., et al. Assistive artificial intelligence for enhanced patient access to ultrasound-guided regional anaesthesia. Br J Anaesth. 2024;132:1173–1175. doi: 10.1016/j.bja.2023.08.004. [DOI] [PubMed] [Google Scholar]
  • 42.Bowness J.S., Morse R., Lewis O., et al. Variability between human experts and artificial intelligence in identification of anatomical structures by ultrasound in regional anaesthesia: a framework for evaluation of assistive artificial intelligence. Br J Anaesth. 2024;132:1063–1072. doi: 10.1016/j.bja.2023.09.023. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Erdem G., Ermiş Y., Özkan D. The effect of using ultrasonography device integrated with artificial intelligence on imaging the injection area in peripheral nerve block applications. JARSS. 2023;31:21–27. [Google Scholar]
  • 44.Jo Y., Lee D., Baek D., et al. Optimal view detection for ultrasound-guided supraclavicular block using deep learning approaches. Sci Rep. 2023;13:17209. doi: 10.1038/s41598-023-44170-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Scholzen E., Schroeder K. Use of artificial intelligence software helpful for regional anesthesia education in self-reported questionnaire in academic medical center setting. Research Square. 2023 doi: 10.21203/rs.3.rs-2790929/v1. Advance Access published on April 21. [DOI] [Google Scholar]
  • 46.Zhao Y., Zheng S., Cai N., et al. Utility of artificial intelligence for real-time anatomical landmark identification in ultrasound-guided thoracic paravertebral block. J Digit Imaging. 2023;36:2051–2059. doi: 10.1007/s10278-023-00851-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Ahiskalioglu A., Yayik A.M., Karapinar Y.E., Tulgar S., Ciftci B. From ultrasound to artificial intelligence: a new era of the regional anesthesia. Minerva Anestesiol. 2022;88:640–642. doi: 10.23736/S0375-9393.22.16456-4. [DOI] [PubMed] [Google Scholar]
  • 48.Cai N., Wang G., Xu L., et al. Examining the impact perceptual learning artificial-intelligence-based on the incidence of paresthesia when performing the ultrasound-guided popliteal sciatic block: simulation-based randomized study. BMC Anesthesiol. 2022;22:392. doi: 10.1186/s12871-022-01937-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Ding Y., Yang Q., Wang Y., Chen D., Qin Z., Zhang J. MallesNet: A multi-object assistance based network for brachial plexus segmentation in ultrasound images. Med Image Anal. 2022;80 doi: 10.1016/j.media.2022.102511. [DOI] [PubMed] [Google Scholar]
  • 50.Erdem G., Ermiş Y., Özkan D. Peripheral nerve blocks and the use of artificial intelligence-assisted ultrasonography. J Clin Anesth. 2022;78 doi: 10.1016/j.jclinane.2021.110597. [DOI] [PubMed] [Google Scholar]
  • 51.Tian D., Zhu B., Wang J., et al. Brachial plexus nerve trunk recognition from ultrasound images: a comparative study of deep learning models. IEEE Access. 2022;10:82003–82014. [Google Scholar]
  • 52.Wang Y., Zhu B., Kong L., et al. Brachial plexus nerve trunk segmentation using deep learning: a comparative study with doctors’ manual segmentation. arXiv. 2022 doi: 10.48550/arXiv.2205.08143. [DOI] [Google Scholar]
  • 53.Yang X.Y., Wang L.T., Li G.D., et al. Artificial intelligence using deep neural network learning for automatic location of the interscalene brachial plexus in ultrasound images. Eur J Anaesthesiol. 2022;39:758–765. doi: 10.1097/EJA.0000000000001720. [DOI] [PubMed] [Google Scholar]
  • 54.Díaz-Vargas H.M., Jimenez-Castaño C.A., Cárdenas-Peña D.A., Aguirre-Ospina O.D., Orozco-Gutierrez A.A. International Workshop on Artificial Intelligence and Pattern Recognition. 124–33. 2021. Peripheral nerve segmentation in ultrasound images using conditioned u-net. Havana. [Google Scholar]
  • 55.Gao S. Proceedings of the 2021 5th International Conference on Electronic Information Technology and Computer Engineering. 2021. A study on ultrasound nerve image segmentation with multi-densely-layer supervision mechanism; pp. 567–572. New York. [Google Scholar]
  • 56.Liu Y., Cheng L. Ultrasound images guided under deep learning in the anesthesia effect of the regional nerve block on scapular fracture surgery. J Healthc Eng. 2021;2021 doi: 10.1155/2021/6231116. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.McKendrick M., Yang S., McLeod G.A. The use of artificial intelligence and robotics in regional anaesthesia. Anaesthesia. 2021;76(Suppl 1):171–181. doi: 10.1111/anae.15274. [DOI] [PubMed] [Google Scholar]
  • 58.Smistad E., Lie T., Johansen K.F. 2021 IEEE International Ultrasonics Symposium (IUS) 2021. Real-time segmentation of blood vessels, nerves and bone in ultrasound-guided regional anesthesia using deep learning; pp. 1–4. Xi’an. [Google Scholar]
  • 59.Van Boxtel J., Vousten V., Pluim J., Rad N.M. 2021 29th European Signal Processing Conference (EUSIPCO) 2021. Hybrid deep neural network for brachial plexus nerve segmentation in ultrasound images; pp. 1246–1250. Dublin. [Google Scholar]
  • 60.Wang Y., Geng J., Zhou C., Zhang Y. Information System and Computer Engineering (CISCE); Beijing: 2021. Segmentation of ultrasound brachial plexus based on U-Net. 2021 International Conference on Communications; pp. 482–485. [Google Scholar]
  • 61.Alkhatib M., Hafiane A., Vieyres P. 25th International Conference on Pattern Recognition (ICPR) 2020. Merged 1D-2D deep convolutional neural networks for nerve detection in ultrasound images; pp. 4774–4780. Milano. [Google Scholar]
  • 62.Ding Y., Yang Q., Wu G., Zhang J., Qin Z. Multiple instance segmentation in brachial plexus ultrasound image using BPMSegNet. arXiv. 2012 doi: 10.48550/arXiv.2012.12012. [DOI] [Google Scholar]
  • 63.Wang R., Shen H., Zhou M. 2019 IEEE International Conference on Imaging Systems and Techniques (IST) 1–6. 2020. Ultrasound nerve segmentation of brachial plexus based on optimized ResU-Net. Abu Dhabi. [Google Scholar]
  • 64.Abraham N., Illanko K., Khan N., Androutsos D. 2019 3rd International Conference on Imaging, Signal Processing and Communication (ICISPC) 85–9. 2019. Deep learning for semantic segmentation of brachial plexus nervesin ultrasound images using U-Net and M-Net. Singapore. [Google Scholar]
  • 65.Huang C., Zhou Y., Tan W., et al. Applying deep learning in recognizing the femoral nerve block region on ultrasound images. Ann Transl Med. 2019;7:453. doi: 10.21037/atm.2019.08.61. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Kakade A., Dumbali J. 2018 International conference on communication information and Computing Technology (ICCICT) 1–6. 2018. Identification of nerve in ultrasound images using U-net architecture. Mumbai. [Google Scholar]
  • 67.Smistad E., Johansen K.F., Iversen D.H., Reinertsen I. Highlighting nerves and blood vessels for ultrasound-guided axillary nerve block procedures using neural networks. J Med Imaging (Bellingham) 2018;5 doi: 10.1117/1.JMI.5.4.044004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Baby M., Jereesh A. 2017 International Conference of Electronics, Communication and Aerospace Technology (ICECA) 107–12. 2017. Automatic nerve segmentation of ultrasound images. Coimbatore. [Google Scholar]
  • 69.Hafiane A., Vieyres P., Delbos A. Deep learning with spatiotemporal consistency for nerve segmentation in ultrasound images. arXiv. 2017 doi: 10.48550/arXiv.1706.05870. [DOI] [Google Scholar]
  • 70.Jin A., Tighe P., Miyatake M., Pawpowicz N., Rashidi P. 2017 Annual Meeting Abstracts: Pain Medicine. American Association of Physicists in Medicine; 2017. Convolutional neural network approaches to preliminary classification of ultrasound-guided regional anesthetic target regions; pp. 569–621. [Google Scholar]
  • 71.Zhao H., Sun N. Image and Graphics: 9th International Conference, ICIG 2017, Shanghai, China, September 13-15, 2017, Revised Selected Papers, Part II 9. 2017. Improved U-net model for nerve segmentation; pp. 496–504. Shanghai. [Google Scholar]
  • 72.Smistad E., Løvstakken L. International Workshop on Deep Learning in Medical Image Analysis. 30–8. 2018. Vessel detection in ultrasound images using deep convolutional neural networks. Granada. [Google Scholar]
  • 73.Simpson D., Jin A., Miyatake M., Rashidi P., Tighe P. What makes it this, and not that? Deep learning neural networks for characterisation of ultrasound guided peripheral nerve blocks: elementary hyper-parameter explorations of pilot anatomical windows. Scientific Abstracts and ePosters. ASRA 42nd Annual Regional Anesthesiology and Acute Pain Medicine Meeting April 6–8, 2017 San Francisco. Reg Anesth Pain Med. 2017;42:802. [Google Scholar]
  • 74.Cao Y., Tan C., Qian W., et al. 2022 IEEE International Conference on Unmanned Systems (ICUS) 2022. Automatic spinal ultrasound image segmentation and deployment for real-time spine volumetric reconstruction; pp. 922–927. Beijing. [Google Scholar]
  • 75.Hetherington J., Brohan J., Rohling R., et al. A novel ultrasound software system for lumbar level identification in obstetric patients. Can J Anaesth. 2022;69:1211–1219. doi: 10.1007/s12630-022-02300-6. [DOI] [PubMed] [Google Scholar]
  • 76.Hetherington J., Lessoway V., Gunka V., Abolmaesumi P., Rohling R. SLIDE: automatic spine level identification system using a deep convolutional neural network. Int J Comput Assist Radiol Surg. 2017;12:1189–1198. doi: 10.1007/s11548-017-1575-8. [DOI] [PubMed] [Google Scholar]
  • 77.Hetherington J., Pesteie M., Lessoway V.A., Abolmaesumi P., Rohling R.N. Identification and tracking of vertebrae in ultrasound using deep networks with unsupervised feature learning. Medical Imaging 2017: Image-Guided Procedures, Robotic Interventions, and Modeling. 2017:159–165. [Google Scholar]
  • 78.Wang N., Chen L., Wang H., Wang Y., Ruan B. MATrix LABoratory algorithm in ultrasound image-guided general drug anesthesia along with lumbar and sacral plexus block in hip replacement under. Sci Program. 2021;2021 [Google Scholar]
  • 79.Akkasaligar P.T., Koolagudi S.G., Biradar S., Hotagi P., Badiger S. 2020 IEEE Bangalore Humanitarian Technology Conference (B-HTC) 1–6. 2020. Vijiyapur. [Google Scholar]
  • 80.Xu H., Li Y., Zhang J., Cheng L., Pan S. The study of automatic matching and optimal location of epidural anesthesia based on ultrasonic image processing. J Intell Fuzzy Syst. 2020;38:3849–3857. [Google Scholar]
  • 81.Compagnone C., Borrini G., Calabrese A., Taddei M., Bellini V., Bignami E. Artificial intelligence enhanced ultrasound (AI-US) in a severe obese parturient: a case report. Ultrasound J. 2022;14:34. doi: 10.1186/s13089-022-00283-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82.Kimizuka M., Tokinaga Y., Taguchi M., Takahashi K., Yamakage M. Usefulness and accuracy of a handheld ultrasound device for epidurssal landmark and depth assessment by anesthesiology residents. J Anesth. 2022;36:693–697. doi: 10.1007/s00540-022-03096-x. [DOI] [PubMed] [Google Scholar]
  • 83.Baglioni S., Paolatti S., Velardo M., Capogna G. Accuracy of the SpineNav3DTM ultrasound technology in estimating the epidural space depth for epidural and spinal insertion in pregnant obese patients. Open J Anesthesiol. 2021;11:221–228. [Google Scholar]
  • 84.Capogna G., Baglioni S., Milazzo V., Vitale A. Accuracy of the SpineNav3DTM technology to measure the depth of epidural space: a comparison with the standard ultrasound technique in pregnant volunteers. Open J Anesthesiol. 2018;8:113. [Google Scholar]
  • 85.Ramteke J.H., Sahu D.K., Sonawane A., Parampill R. New technology: Handheld ultrasound-assisted localization of epidural space. Saudi J Anaesth. 2018;12:365–367. doi: 10.4103/sja.SJA_696_17. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86.Seligman K.M., Weiniger C.F., Carvalho B. The accuracy of a handheld ultrasound device for neuraxial depth and landmark assessment: a prospective cohort trial. Anesth Analg. 2018;126:1995–1998. doi: 10.1213/ANE.0000000000002407. [DOI] [PubMed] [Google Scholar]
  • 87.Delforche J., Vanderhaeghen S., Coppens M. How effective is an ultrasound-based imaging technique with automated guidance as an aid in performing spinal anesthesia in elective caesarean section patients? A prospective randomized controlled trial. Acta Anaethesiol Belg. 2021;72:123–132. [Google Scholar]
  • 88.Chan J.J., Ma J., Leng Y., et al. Machine learning approach to needle insertion site identification for spinal anesthesia in obese patients. BMC Anesthesiol. 2021;21:246. doi: 10.1186/s12871-021-01466-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89.Singla P., Dixon A.J., Sheeran J.L., Scalzo D., Mauldin F.W., Jr., Tiouririne M. Feasibility of spinal anesthesia placement using automated interpretation of lumbar ultrasound images: a prospective randomized controlled trial. J Anesth Clin Res. 2019;10:878. doi: 10.4172/2155-6148.1000878. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90.Ni X., Li M.Z., Zhou S.Q., et al. Accuro ultrasound-based system with computer-aided image interpretation compared to traditional palpation technique for neuraxial anesthesia placement in obese parturients undergoing cesarean delivery: a randomized controlled trial. J Anesth. 2021;35:475–482. doi: 10.1007/s00540-021-02922-y. [DOI] [PubMed] [Google Scholar]
  • 91.D. Ghisi, M. Tomasi, S. Giannone, et al., A randomized comparison between Accuro and palpation-guided spinal anesthesia for obese patients undergoing orthopedic surgery, Reg Anesth Pain Med, 45,2020, 63 - 66. [DOI] [PubMed]
  • 92.Carvalho B., Seligman K.M., Weiniger C.F. The comparative accuracy of a handheld and console ultrasound device for neuraxial depth and landmark assessment. Int J Obstet Anesth. 2019;39:68–73. doi: 10.1016/j.ijoa.2019.01.004. [DOI] [PubMed] [Google Scholar]
  • 93.Oh T.T., Ikhsan M., Tan K.K., et al. A novel approach to neuraxial anesthesia: application of an automated ultrasound spinal landmark identification. BMC Anesthesiol. 2019;19:57. doi: 10.1186/s12871-019-0726-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 94.Henderson M., Dolan J. Challenges, solutions, and advances in ultrasound-guided regional anaesthesia. BJA Educ. 2016;16:374–380. [Google Scholar]
  • 95.Liu X., Cruz Rivera S., Moher D., et al. Reporting guidelines for clinical trial reports for interventions involving artificial intelligence: the CONSORT-AI extension. Lancet Digit Health. 2020;2:e537–e548. doi: 10.1016/S2589-7500(20)30218-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 96.Schulz K.F., Altman D.G., Moher D. CONSORT 2010 statement: updated guidelines for reporting parallel group randomised trials. BMJ. 2010;340:c332. doi: 10.1136/bmj.c332. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 97.Vasey B., Nagendran M., Campbell B., et al. Reporting guideline for the early-stage clinical evaluation of decision support systems driven by artificial intelligence: DECIDE-AI. Nat Med. 2022;28:924–933. doi: 10.1038/s41591-022-01772-9. [DOI] [PubMed] [Google Scholar]
  • 98.von Elm E., Altman D.G., Egger M., Pocock S.J., Gøtzsche P.C., Vandenbroucke J.P. The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) statement: guidelines for reporting observational studies. Lancet. 2007;370:1453–1457. doi: 10.1016/S0140-6736(07)61602-X. [DOI] [PubMed] [Google Scholar]
  • 99.O. Russokovsky, J. Deng, H. Su, et al. ImageNet large scale visual recognition challenge International Journal of Computer Vision, 115, 2015, 211 - 252 (https://www.image-net.org/challenges/LSVRC/index.php accessed 21 September 2023).
  • 100.COCODataset, Common Objects in Context. Available from https://cocodataset.org/#home (Accessed 21 September 2023).
  • 101.Google/DeepMind. Multi-object image datasets with ground-truth segmentation masks and generative factors. Available from https://github.com/google-deepmind/multi_object_datasets (accessed February 2024, 29).
  • 102.GE Healthcare. Venus Point of Care Ultrasound for Regional Anaesthesia. Available from https://www.gehealthcare.com/products/ultrasound/venue-family/regional-anesthesia. Accessed 17.072023 - 21.09.2023.
  • 103.Mindray. Diagnostic Ultrasound System, Tex20 Series. Available from https://www.mindray.com/uk/products/ultrasound/point-of-care/tex20-series. (accessed 17 July 2023–21 September 2023).
  • 104.Samsung. News Centre - The NerveTrackTM solution received FDA clearance. Available from https://www.samsunghealthcare.com/en/about_us/news_center/170. (accessed 17 July 2023–21 September 2023).
  • 105.Nerveblox. Artificial intelligence to delivery nerve blocks faster. Available from https://www.nerveblox.com/. (accessed 17 July 2023–21 September 2023).
  • 106.Intelligent Ultrasound. ScanNav Anatomy Peripheral Nerve Block. Available from https://www.intelligentultrasound.com/scannav-anatomy-pnb/. (accessed 17 July 2023–21 September 2023).
  • 107.Rivanna Medical. Accuro is a Clinically Proven Solution for Neuraxial Anesthesia Guidance. Available from https://rivannamedical.com/clinical-value/#1566748145876-d7f7db9f-203a. (accessed 17 July 2023–21 September 2023).
  • 108.Rivanna Medical. Ultrasound Made Simple with Accuro Neuraxial Guidance. Available from https://rivannamedical.com/why-accuro/. (accessed 17 July 2023–21 September 2023).
  • 109.HiCura. uSine Technology. Available from https://hicuramedical.com/usine-technology/. (accessed 17 July 2023–21 September 2023).
  • 110.Rajpurkar P., Chen E., Banerjee O., Topol E.J. AI in health and medicine. Nat Med. 2022;28:31–38. doi: 10.1038/s41591-021-01614-0. [DOI] [PubMed] [Google Scholar]
  • 111.Grand View Research. AI in Healthcare Market Size, Share & Trends Analysis Report (Software Solutions, Hardware, Services), By Application (Virtual Assistance, Connected Machines), By Region, And Segment Forecasts, 2024 - 2030. Available https://www.grandviewresearch.com/industry-analysis/artificial-intelligence-ai-healthcare-market. (accessed 22 September 2023).
  • 112.https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-aiml-enabled-medical-devices FDA. Artificial Intelligence and Machine Learning (AI/ML)-Enabled Medical Devices. Available from.
  • 113.Davoud S.C., Kovacheva V.P. On the horizon: specific applications of automation and artificial intelligence in anesthesiology. Curr Anesthesiol Rep. 2023;13:31–40. doi: 10.1007/s40140-023-00558-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 114.Bowness J.S., Pawa A., Turbitt L., et al. International consensus on anatomical structures to identify on ultrasound for the performance of basic blocks in ultrasound-guided regional anesthesia. Reg Anesth Pain Med. 2022;47:106–112. doi: 10.1136/rapm-2021-103004. [DOI] [PubMed] [Google Scholar]
  • 115.Ashken T., Bowness J., Macfarlane A.J.R., et al. Recommendations for anatomical structures to identify on ultrasound for the performance of intermediate and advanced blocks in ultrasound-guided regional anesthesia. Reg Anesth Pain Med. 2022;47:762–772. doi: 10.1136/rapm-2022-103738. [DOI] [PubMed] [Google Scholar]
  • 116.El-Boghdadly K., Wolmarans M., Stengel A.D., et al. Standardizing nomenclature in regional anesthesia: an ASRA-ESRA Delphi consensus study of abdominal wall, paraspinal, and chest wall blocks. Reg Anesth Pain Med. 2021;46:571–580. doi: 10.1136/rapm-2020-102451. [DOI] [PubMed] [Google Scholar]
  • 117.Ahmed H.M., Atterton B.P., Crowe G.G., et al. Recommendations for effective documentation in regional anesthesia: an expert panel Delphi consensus project. Reg Anesth Pain Med. 2022;47:301–308. doi: 10.1136/rapm-2021-103136. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 118.Cohen I.G., Evgeniou T., Gerke S., Minssen T. The European artificial intelligence strategy: implications and challenges for digital health. Lancet Digit Health. 2020;2:e376–e379. doi: 10.1016/S2589-7500(20)30112-6. [DOI] [PubMed] [Google Scholar]
  • 119.Available from https://www.fda.gov/medical-devices/software-medical-device-samd/good-machine-learning-practice-medical-device-development-guiding-principles. (accessed 22 September 2023).
  • 120.van der Meijden S.L., Arbous M.S., Geerts B.F. Possibilities and challenges for artificial intelligence and machine learning in perioperative care. BJA Educ. 2023;23:288–294. doi: 10.1016/j.bjae.2023.04.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 121.Turbitt L.R., Mariano E.R., El-Boghdadly K. Future directions in regional anaesthesia: not just for the cognoscenti. Anaesthesia. 2020;75:293–297. doi: 10.1111/anae.14768. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Multimedia component 1
mmc1.docx (48.8KB, docx)
Multimedia component 2
mmc2.docx (74.4KB, docx)

Articles from BJA: British Journal of Anaesthesia are provided here courtesy of Elsevier

RESOURCES