Skip to main content
Springer logoLink to Springer
. 2022 May 12;17(8):1437–1444. doi: 10.1007/s11548-022-02642-y

Task model-specific operator skill assessment in routine fetal ultrasound scanning

Yipei Wang 1,, Qianye Yang 2, Lior Drukker 3, Aris Papageorghiou 3, Yipeng Hu 1,2, J Alison Noble 1
PMCID: PMC9307537  PMID: 35556206

Abstract

Purpose

For highly operator-dependent ultrasound scanning, skill assessment approaches evaluate operator competence given available data, such as acquired images and tracked probe movement. Operator skill level can be quantified by the completeness, speed, and precision of performing a clinical task, such as biometry. Such clinical tasks are increasingly becoming assisted or even replaced by automated machine learning models. In addition to measurement, operators need to be competent at the upstream task of acquiring images of sufficient quality. To provide computer assistance for this task requires a new definition of skill.

Methods

This paper focuses on the task of selecting ultrasound frames for biometry, for which operator skill is assessed by quantifying how well the tasks are performed with neural network-based frame classifiers. We first develop a frame classification model for each biometry task, using a novel label-efficient training strategy. Once these task models are trained, we propose a second task model-specific network to predict two skill assessment scores, based on the probability of identifying positive frames and accuracy of model classification.

Results

We present comprehensive results to demonstrate the efficacy of both the frame-classification and skill-assessment networks, using clinically acquired data from two biometry tasks for a total of 139 subjects, and compare the proposed skill assessment with metrics of operator experience.

Conclusion

Task model-specific skill assessment is feasible and can be predicted by the proposed neural networks, which provide objective assessment that is a stronger indicator of task model performance, compared to existing skill assessment methods.

Supplementary Information

The online version contains supplementary material available at 10.1007/s11548-022-02642-y.

Keywords: Skill assessment, Ultrasound, Fetal ultrasound, Deep learning

Introduction

Skill assessment of imaging specialists has long been established as an essential tool in training, continuing education, and clinical service auditing and improvement [1]. Ultrasound imaging is a particularly hard skill to learn and is known to be operator dependent. Ultrasound skill needs to capture both the ability to read and to take accurate diagnostic measurements on images (as in radiology) but also the ability to capture diagnostic images (unlike in radiology). Traditional quantitative metrics such as time-to-completion and clinical outcome have been used to provide quantitative evidence to inform, design, and deliver skill assessment methods [2, 3] but are considered simplistic.

Building on recent advances in machine learning in imaging, especially in deep learning, recent efforts have proposed neural network-based automatic skill assessment approaches [48]. Most of these learning-based methods automate skill assessment by predicting the skill level from available intra-procedure data, including imaging [4], motion [5, 6], or a combination of both [7]. The training labels that represent skill level include operator experience, such as length of practice [6] and other competence indicators such as different clinical roles [8]. Machine learning algorithms have also been proposed to measure task-specific image quality [9] that may be partially indicative of skill assessment and directly relevant to the clinical tasks to hand.

Standard plane selection [10], gestational age regression [11], and direct biometry estimation [12] are common clinical fetal ultrasound imaging tasks that have been considered for automation by machine learning approaches. Indeed, sufficient progress has been made for some solutions to appear in commercial systems. In this paper, we consider the impact that the introduction of such task models has on the skills required to acquire, interpret, and report ultrasound images and the assessment of the “new” skills. For example, a frame classifier may indicate that sufficiently accurate biometry can be reliably measured on an ultrasound frame deemed “positive”, while a “negative” frame potentially suggests the need for image re-acquisition. Such a classification model may relax the definition of a standard plane and subsequently alter the skill required to acquire frames amenable to the classifier.

In this paper, we focus on two established clinical tasks in fetal ultrasound examination, measuring the head circumference and the abdomen circumference on ultrasound video frames. Firstly, for each task, a frame classification model is trained using a novel label-efficient approach that utilises mixed manual and image-similarity-based surrogate labels from a high volume of ultrasound video frames. Secondly, we propose to measure how well these task models perform for assessing operator skill level, then develop new neural networks to predict these skill levels based on real-time ultrasound images and synchronised probe motion data. These skill assessment predictors are trained to quantify two proposed criteria that are directly informed by frame classification performance. In addition to these methodological developments, our contributions in this paper include a set of systematic results showing the efficacy of both task models and skill assessment predictors, together with detailed analysis with ablation studies evaluating the importance of a selection of network hyperparameters, training strategy options, varying input data combinations, and different skill level definitions.

Method

An overview of the proposed task model-specific skill assessment framework is shown in Fig. 1. Let Ji be the available number of ultrasound video frames from ith subject. A total of iIJi video frames {{vij|j=1,,Ji}|i=1,,I}, sampled from a set of I video clips Vi={vij|j=1,,Ji}, provide the training input for the task model which performs a clinical task of interest, developed in section “Frame classification networks”. The task labels {{gij|j=1,,Ji}|i=1,,I} are application dependent, such as the suitability of given frames for fetal biometry (see “Label-efficient mixed ground-truth generation” section). The ultrasound video clips {Vi} are then combined with synchronised probe motion data {Mi}, for training a second neural network for skill assessment, supervised by a measure of task performance {fskill(G^i)|i=1,,I}, based on the task model predictions G^i={G^ij|j=1,,Ji} at inference. This skill assessment predictor is described in section “Skill assessment predictors”, with two proposed definitions of task performance fskill(·).

Fig. 1.

Fig. 1

Overview of the task model-specific skill assessment framework

Task models

Frame classification networks

Without loss of generality, we consider an end clinical task being a binary classification. Specifically in this work, the tasks are to classify whether a sufficiently accurate measurement can be obtained on individual frames in the ultrasound video. For each clinical task described in “Clinical tasks and data acquisition” section, a deep neural network ftask(V;θ):VG, with learnable parameters θ, is trained to predict the probability g^ij=ftask(V;θ^) of a positive class for each input frame V=vij. Optimising parameters θ^ requires a ground-truth label G=gij for every input frame vij, during the supervised network training. The details of network training are summarised in “Model implementation and training” section.

Label-efficient mixed ground-truth generation

In the routine second trimester fetal ultrasound examination, the fetal head circumference (HC) and abdominal circumference (AC) are measured on biometry planes selected manually by operators following the standard plane definitions detailed in widely-adopted guidelines [13]. In order to train the frame classification networks, labels are required for all image frames in the entire ultrasound video. We therefore propose a practical approach to utilise a limited number of manually annotated labels from each subject, then use them to regress the binary labels on the unlabelled frames using a set of image similarity measures between a given frame and the one selected as the diagnostic plane vid.p. for the ith subject. Five image similarity metrics were adopted, cosine similarity in feature space [14], mean squared error, cross correlation [15], mutual information, and structural similarity index [16]. To calculate the cosine similarity, an in-house ultrasound-based pre-trained model was used to extract image features.

For any given frame vij, the percentage relative error is defined directly between the biometry fbiometry(vij) obtained on the frame and that from the diagnostic plane of the same ith subject: δij=fbiometry(vij)fbiometry(vid.p.)×100%. Then, the fitted linear model δ^ij=flinear(V=vij;vid.p.) predicts an estimation δ^ij of the percentage relative error. Thus, the surrogate binary label gij for an unlabelled frame vij can be generated using a minimum threshold δmin on the estimated δ^ij, such that: gij=0,δ^ij>δmin1,otherwise. The threshold δmin has a direct clinical interpretation, the precision of target biometry, which is of significance in controlling the quality of the mixed labels.

Skill assessment predictors

The skill assessment predictor fassess(V,M;ω):V,MS is a multi-modal neural network with trainable parameters ω, which takes τ consecutive frames V={vij|j=n,,n+τ-1} as well as the corresponding probe motion segments M={mij|j=n,,n+τ-1} and predicts a task performance score S=S^in during this time interval, where n=1,,Ji-τ+1 and each video sequence forms Ji-τ+1 training data. The network is optimised with the mean-squared-error (MSE) loss L=Ei,n[(Sin-S^in)2], where Ei,n[·] is the expectation over all training data and Sin=fskill({g^ij|j=n,,n+τ-1}) is the supervising task performance score obtained from the frame classification networks, and is considered as an estimate of fskill(G^i), for the ith subject at the nth available time point. In the following subsections, two types of task performance score are introduced for operator skill assessment.

Expected positive task prediction

For both tasks considered in this paper, positive frames enable successful biometric measurement. Therefore, we first investigate a task performance score defined as the expected value of probability E[g^ij], for the positive class predicted by the task models. Namely, for the ith subject at the nth time point:

fskillpos({g^ij};n)=1τj=nn+τ-1g^ij, 1

Expected task prediction accuracy

A second skill assessment criterion determines how accurate the task model predicts both a positive and a negative class, rather than the probability of obtaining a positive class. Correct prediction of a negative class is therefore considered as important as the correct prediction of a positive frame. This definition takes into account both risks with positive and negative frames being incorrectly classified, corresponding to Types 1 and 2 errors, rather than favouring a high positive rate only (as Eq. 1 could). Let TPin(c) and TNin(c) represent the number of true positive and true negative frames respectively. The second task performance score for assessing skill level is then defined in terms of the binary classification accuracy, at the nth time point:

fskillacc({g^ij};n,c)=TPin(c)+TNin(c)τ, 2

where c is a pre-defined cut-off on the task model-predicted class probability.

Experiments

Clinical tasks and data acquisition

Ultrasound video and probe motion data used in this study were acquired as part of the PULSE study, approved by the UK Research Ethics Committee (reference 18/WS/0051) [17]. Scans were performed from June 2018 to February 2020 at the Oxford University Hospitals NHS Foundation Trust, by sonographers and fetal medicine doctors (collectively referred to as operators in this paper) using a Voluson E8 scanner (GE Healthcare, USA), with curvilinear (C2-9-D, C1-5-D) and 3D/4D (RAB6-D) probes. Written informed consent was given by all participating operators and pregnant women. To acquire probe motion data, an inertial measurement unit (Next Generation IMU (NGIMU), x-io Technologies Limited, Bristol, UK) was fixed at the same probe position during all scans. The raw motion signal was recorded at 400 Hz and then down-sampled with the timestamp-synchronised ultrasound video frames, losslessly frame-grabbed and recorded at 30 Hz.

A subset of the second trimester scans from the PULSE dataset was used, with manual labels of head circumference and abdominal circumference obtained using the available user interface on the scanner. Up to 12 s of video before the operator stops scanning for measurement was extracted, which resulted in an average of 10.61 s of consecutive frames. A total of 294 clips, for scanning fetal head and abdomen which have consistent zoom factors and a length of at least 6 s, formed the dataset used in this study, for the purpose of controlled experiments. In summary, up to 84 clips with 27k video frames from each of the 11 individual sonographers were included in this study. Further details of the data contributions with respect to different operators are included in the supplementary materials.

Head circumference (HC) measurement is a routine clinical task that indicates a number of clinical anomaly conditions and is also used for gestational age estimation. HC measurement is typically taken on a manually selected diagnostic plane. In this experiment, we investigate a frame classification model to assist this task by differentiating individual ultrasound image frames, from which the measured HC is sufficiently accurate, using a total of 146 HC video clips. To generate the task model labels for this task, 345 frames from randomly sampled 22 clips were annotated manually. The circumferences were calculated on fitted ellipses, in an “outer-to-outer” manner [18].

Abdominal circumference (AC) measurement is another important biometry that is correlated to fetal growth parameters, such as weight and gestational age. Similar to the frame classifier defined for assisting the HC measurement task, experiments were designed to examine a frame classification model for accurate AC reporting. From 19 out of 149 available AC clips, 298 frames were annotated manually for generating the mixed labels.

Model implementation and training

The linear regression models for training label generation were fitted on 266 and 228 images from 17 and 14 clips and were tested on 79 and 70 holdout test images, for the HC and AC tasks, respectively.

All ultrasound images used in this work were resized to 224×288 pixels and normalised to zero-mean and unit-variance. A VGG-based network [19] was adopted with the same implementation details as the task models for the purpose of a reference quality benchmark in this work. For skill assessment prediction, a VGG16 network [20] was adopted for extracting features from a stack of consecutive ultrasound images, with a 1D ResNet18 network [21] for extracting features from 1D probe motion signals, here, the three-axis angular velocities formed the three-channel 1D input. The input window size is referring to the number of input images and the length of the synchronised probe motion signals. Features from the two branches were concatenated via adaptive pooling, resulting in a feature vector with a length of 32, and this feature vector was fed into linear layers which predict the task performance score, as illustrated in Fig. 1. All networks were implemented in PyTorch and trained with a mini-batch size of 64 on the NVIDIA Quadro GV100 GPUs using an Adam optimiser with a learning rate of 10-4. The task models and skill assessment predictors were trained with cross-entropy and MSE losses, for 50 and 700 epochs, respectively.

Model evaluation and ablation studies

Task models were evaluated on the acquired ultrasound video dataset as detailed in “Clinical tasks and data acquisition” section. The dataset was randomly partitioned into a development set and a holdout test set, without any operators or subjects in both sets. This partition was repeated to form the A and B “splits” for the HC task and the C and D “splits” for the AC task. Further details of dataset split are provided in the supplementary material. A further 85:15 random split of the development set resulted in a training set and a validation set during developing the models, while models were trained with different hyperparameters and those with the highest class-balanced accuracy on the validation set were selected. The results reported in this paper are based on these selected models tested on the holdout test set. All models were trained and tested using the generated surrogate labels. The reason to only test the surrogate labels is that, in real-world ultrasound scanning, we do not expect the sonographers to make manual annotations on frames that are not selected as the diagnostic plane. Hyperparameters investigated in this work include learning rate, mini-batch size and, for training the task models, multiple consecutive frames for contextualising the input (a five-gram context) were also tested. A ResNet18 network [21] for extracting ultrasound image features was also tested and the results were not found significantly different to those using VGG16, therefore they are included in supplementary material.

The accuracy, sensitivity, and specificity were computed for ablation studies to evaluate the task models, with varying precision requirements, δmin{1%,2%,3%,4%}, and different cut-off values on predicted class probabilities, such that the resulting specificity values are controlled at 0.8 and 0.9 as examples of clinically acceptable false positive rates.

Skill assessment predictors were tuned on the validation set and those that achieved the lowest root mean square error (RMSE) were used to report results on the test set. Two metrics were computed to compare ablation studies on varying window size τ, for each of the development-test splitting strategies and for each of the HC and AC tasks, (1) the RMSE between the predicted and the task model-generated performance scores and (2) the Pearson correlation coefficient (PCC) between them. The impact from different types of input data types was also compared, by using either the video data or motion data alone for training and testing the skill assessment predictors.

Results

Quality of the mixed label generation

The mixed label generator flinear described in Section “Label-efficient mixed ground-truth generation” was tested on an independently sampled subset with manual labels as ground truth. The predicted relative percentage error δ^ achieved MSE values of 0.00076±0.00046 and 0.00084±0.00048, on the HC and AC datasets, respectively. Of note, this error is with a normalised range of [0, 1], therefore the absolute MSE is reported. Further visualisation results of the mixed label generator performance are provided in the supplementary materials.

The head circumference task

Task model performance

For both datasets A and B, the task model accuracy for the HC task generally increases as the allowed percentage error δmin increases, with consistent sensitivity and specificity values observed at a default cut-off value of 0.5. The performance of the task model with HC dataset is summarised in Table 1.

Table 1.

Task model performance for different data splits as δmin varies

Task Split δmin Accuracy Sensitivity Specificity Task Split δmin Accuracy Sensitivity Specificity
HC A 0.01 0.58 0.79 0.52 AC C 0.01 0.80 0.38 0.84
0.02 0.74 0.80 0.64 0.02 0.71 0.62 0.75
0.03 0.73 0.71 0.82 0.03 0.71 0.58 0.82
0.04 0.89 0.90 0.66 0.04 0.64 0.59 0.76
B 0.01 0.68 0.34 0.82 D 0.01 0.60 0.57 0.60
0.02 0.67 0.60 0.80 0.02 0.73 0.69 0.75
0.03 0.77 0.79 0.66 0.03 0.77 0.79 0.74
0.04 0.85 0.87 0.56 0.04 0.65 0.61 0.77

Skill assessment predictor performance

As summarised in Table 2, for both criteria of fskillacc and fskillpos, the skill assessment predictors using both motion and video as the input achieved the highest accuracy, which indicates that using complementary multi-modality data significantly improved performance on fskillacc (both p values < 0.01, paired t test at α=0.05), while using video input alone can predict fskillpos well.

Table 2.

Ablation study results for different input data modalities at δmin=0.04

Criterion Data modality RMSE PCC Criterion Data modality RMSE PCC
fskillacc Both 0.164 ± 0.189 − 0.176 ± 0.201 fskillpos Both 0.171 ± 0.116 0.212 ± 0.257
Motion 0.204 ± 0.137     0.059 ± 0.161 Motion 0.209 ± 0.095 0.509 ± 0.191
Video 0.234 ± 0.075 − 0.485 ± 0.218 Video 0.156 ± 0.085 0.798 ± 0.256

Table 3 summarises the results with controlled specificity values by different cut-off values. The performance of the skill assessment predictors, therefore, demonstrates the accuracy with a controlled false positive rate which is critical in many use cases for the skill assessment predictors. Table 4 summarises the ablation results using different window sizes for the two criteria. For fskillacc, the RMSEs decrease as the window size increases, with a decreasing PCC. Notably for predicting fskillpos, the RMSEs with window size 4 are significantly lower than others (p values < 0.01), on both data splits A and B.

Table 3.

Ablation study results of fskillacc, with different cut-off values, at δmin=0.04

Split Cut off Specificity RMSE PCC Split Cut off Specificity RMSE PCC
A 0.710 0.8 0.272 ± 0.185 − 0.543 ± 0.323 B 0.702 0.8 0.318 ± 0.257 − 0.017 ± 0.296
0.799 0.9 0.457 ± 0.197 − 0.076 ± 0.165 0.713 0.9 0.474 ± 0.106 0.023 ± 0.310
Table 4.

Ablation study results of fskillacc and fskillpos, with different τ values, at δmin=0.02

Criterion τ RMSE PCC Criterion τ RMSE PCC
fskillacc 15 0.359 ± 0.146 0.373 ± 0.333 fskillpos 1 0.295 ± 0.111 0.457 ± 0.204
30 0.328 ± 0.159 0.459 ± 0.397 4 0.229 ± 0.090 0.510 ± 0.185
60 0.262 ± 0.133 0.278 ± 0.327 8 0.240 ± 0.096 0.294 ± 0.223
120 0.230 ± 0.156 0.097 ± 0.332 16 0.266 ± 0.135 0.415 ± 0.193

The abdominal circumference task

Task model performance

The performance of the AC task model on two dataset splits is presented in Table 1. Different from the HC task reported in Section “Task model performance”, we observed that the accuracy of the AC task model decreases as the allowed percentage error δmin increases. This might be because there are more anatomical landmarks required to determine the AC plane than that of HC. A higher δmin may result in higher increased variance in the positive class of the AC task and perhaps more challenging cases too, compared with those in the HC task.

Skill assessment predictor performance

The performance of two skill assessment predictors for the AC task is presented in Table 5. The RMSEs generally decrease as δmin increases for fskillacc. A slight increase in RMSE was also observable for fskillacc, perhaps a consequence of the above-discussed biased prediction from the task models.

Table 5.

Performance of skill assessment predictor for the AC task

Criterion Split δmin RMSE PCC Criterion Split δmin RMSE PCC
fskillacc C 0.01 0.299 ± 0.075 −0.602 ± 0.326 fskillpos C 0.01 0.316 ± 0.046 − 0.214 ± 0.179
0.02 0.294 ± 0.09 − 0.383 ± 0.331 0.02 0.456 ± 0.123 0.374 ± 0.170
0.03 0.380 ± 0.128 0.147 ± 0.490 0.03 0.486 ± 0.166 0.253 ± 0.257
0.04 0.389 ± 0.180 0.124 ± 0.382 0.04 0.260 ± 0.099 0.252 ± 0.168
D 0.01 0.371 ± 0.074 0.401 ± 0.166 D 0.01 0.258 ± 0.066 0.092 ± 0.151
0.02 0.374 ± 0.076 − 0.421 ± 0.299 0.02 0.458 ± 0.146 0.310 ± 0.182
0.03 0.363 ± 0.127 0.277 ± 0.444 0.03 0.359 ± 0.133 0.163 ± 0.294
0.04 0.397 ± 0.180 0.145 ± 0.448 0.04 0.312 ± 0.096 0.124 ± 0.267

Case studies: comparison with operator experience

With the above-summarised quantitative results, three real-world cases are included in Fig. 2 to demonstrate the clinical relevance and potential use scenarios for these developed skill assessment predictors. More cases are included in the supplementary materials for further reference. Case 1 was performed by an operator with 7 years of experience. Scores reflect a visible discrepancy in the middle part of the clip, where fskillpos indicates a low possibility of positive class and fskillacc still retain a high accuracy of the task; Case 2 was scanned by a newly qualified operator, yet measured stable and relatively high scores for both criteria during the tested period; Case 3 was assessing an operator with 6 years of experience, as an example of satisfactory prediction from the skill assessment predictors for both criteria.

Fig. 2.

Fig. 2

Three example scan clips plotted along the time, with the time-synchronised skill assessment scores, fskillpos in orange and fskillacc in blue, with both task model-generated scores (dotted lines) and the predicted scores (solid lines). The red boxed frames were the manually annotated ground truth for the diagnostic planes

Discussion and conclusion

It is important to note that, although the mixed label generation was found highly effective with available intra-subject manual labels (section “Quality of the mixed label generation”), it may not be able to replace the frame classifiers for generalising to unseen new subjects. These classifiers are example task models in the proposed skill assessment framework, potentially applicable to a wide range of clinical tasks.

This work compared the proposed skill assessment criteria with the operator experience. Section “Results” presented our first proof-of-concept results using clinical data, demonstrating that it is feasible to predict the task model performance and for assessing operator skill. The performance of the skill assessment predictors may be further improved with data acquired during task model-assisted procedures, albeit with potential ethical challenges. Future work will explore the benefits of training the skill assessment predictors simultaneously with the task models, similar to previously proposed image quality assessment [9, 22]. Investigating the robustness of the proposed method with respect to data acquired from different image settings would also be interesting.

In summary, this work first identified a need for new skill criteria when clinical tasks are assisted by machine learning models and, in turn, proposed to automatically predict the skill levels based on task model performance. The experimental results show that the two new skill assessment scores not only are feasible to predict, but also provide informative criteria that are different from existing experience-based metrics.

Supplementary Information

Below is the link to the electronic supplementary material.

11548_2022_2642_MOESM1_ESM.pdf (5.7MB, pdf)

Supplementary information This manuscript has one accompanying supplementary file. (5,811 KB)

Acknowledgements

We acknowledge the ERC (ERC-ADG-2015 694581, project PULSE) and the NIHR Oxford Biomedical Research Centre.

Declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Ethics approval

This study is approved by the UK Research Ethics Committee (reference 18/WS/0051).

Informed consent

Informed consent was obtained by all operators and women participated in the study.

Footnotes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Harden RM, Stevenson M, Downie WW, Wilson G. Assessment of clinical competence using objective structured examination. Br Med J. 1975;1(5955):447–451. doi: 10.1136/bmj.1.5955.447. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Tolsgaard MG, Todsen T, Sorensen JL, Ringsted C, Lorentzen T, Ottesen B, Tabor A. International multispecialty consensus on how to evaluate ultrasound competence: a Delphi consensus survey. PLoS ONE. 2013;8(2):57687. doi: 10.1371/journal.pone.0057687. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Amini R, Adhikari S, Fiorello A. Ultrasound competency assessment in emergency medicine residency programs. Acad Emerg Med. 2014;21(7):799–801. doi: 10.1111/acem.12408. [DOI] [PubMed] [Google Scholar]
  • 4.Tyrrell RE, Holden MS (2021) Ultrasound video analysis for skill level assessment in fast ultrasound. Comput Methods Biomech Biomed Eng Imaging Vis 9(3):308–312
  • 5.Ziesmann MT, Park J, Unger B, Kirkpatrick AW, Vergis A, Pham C, Kirschner D, Logestty S, Gillman LM. Validation of hand motion analysis as an objective assessment tool for the focused assessment with sonography for trauma examination. J Trauma Acute Care Surg. 2015;79(4):631–637. doi: 10.1097/TA.0000000000000813. [DOI] [PubMed] [Google Scholar]
  • 6.Wang Y, Droste R, Jiao J, Sharma H, Drukker L, Papageorghiou AT, Noble JA (2020) Differentiating operator skill during routine fetal ultrasound scanning using probe motion tracking. In: ASMUS, pp 180–188 [DOI] [PMC free article] [PubMed]
  • 7.Sharma H, Drukker L, Papageorghiou AT, Noble JA (2021) Multi-modal learning from video, eye tracking, and pupillometry for operator skill characterization in clinical fetal ultrasound. In: ISBI, pp 1646–1649 [DOI] [PMC free article] [PubMed]
  • 8.Le Lous M, Despinoy F, Klein M, Fustec E, Lavoué V, Jannin P. Impact of physician expertise on probe trajectory during obstetric ultrasound: a quantitative approach for skill assessment. Simul Healthc. 2021;16(1):67–72. doi: 10.1097/SIH.0000000000000465. [DOI] [PubMed] [Google Scholar]
  • 9.Saeed SU, Fu Y, Stavrinides V, Baum Z, Yang Q, Rusu M, Fan RE, Sonn GA, Noble JA, Barratt DC, Hu Y (2021) Adaptable image quality assessment using meta-reinforcement learning of task amenability. In: ASMUS, pp. 191–201 [DOI] [PubMed]
  • 10.Rahmatullah B, Papageorghiou A, Noble JA (2011) Automated selection of standardized planes from ultrasound volume. In: MLMI, pp 35–42
  • 11.Lee LH, Bradburn E, Papageorghiou AT, Noble JA (2020) Calibrated Bayesian neural networks to estimate gestational age and its uncertainty on fetal brain ultrasound images. In: ASMUS, pp 13–22
  • 12.Bano S, Dromey B, Vasconcelos F, Napolitano R, David AL, Peebles DM, Stoyanov D (2021) AutoFB: Automating fetal biometry estimation from standard ultrasound planes. In: MICCAI, pp 228–238
  • 13.Salomon LJ, Alfirevic Z, Berghella V, Bilardo C, Hernandez-Andrade E, Johnsen S, Kalache K, Leung K-Y, Malinger G, Munoz H, PREFUMO F, TOI A, LEE W (2011) Practice guidelines for performance of the routine mid-trimester fetal ultrasound scan. Ultrasound Obstetr Gynecol 37(1):116–126 [DOI] [PubMed]
  • 14.Jiao J, Cai Y, Alsharid M, Drukker L, Papageorghiou AT, Noble JA (2020) Self-supervised contrastive video-speech representation learning for ultrasound. In: MICCAI, pp 534–543 [DOI] [PMC free article] [PubMed]
  • 15.Yoo J-C, Han TH. Fast normalized cross-correlation. Circuits Syst Signal Process. 2009;28(6):819–843. doi: 10.1007/s00034-009-9130-7. [DOI] [Google Scholar]
  • 16.Wang Z, Bovik AC, Sheikh HR, Simoncelli EP. Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process. 2004;13(4):600–612. doi: 10.1109/TIP.2003.819861. [DOI] [PubMed] [Google Scholar]
  • 17.Drukker L, Sharma H, Droste R, Alsharid M, Chatelain P, Noble JA, Papageorghiou AT (2021) Transforming obstetric ultrasound into data science using eye tracking, voice recording, transducer motion and ultrasound video. Sci Rep 10.1038/s41598-021-92829-1 [DOI] [PMC free article] [PubMed]
  • 18.Sarris I, Ioannou C, Chamberlain P, Ohuma E, Roseman F, Hoch L, Altman D, Papageorghiou A, Fetal I (2012) for the 21st Century (INTERGROWTH-21st), N.G.C.: Intra-and interobserver variability in fetal ultrasound measurements. Ultrasound Obstetr Gynecol 39(3):266–273 [DOI] [PubMed]
  • 19.Baumgartner CF, Kamnitsas K, Matthew J, Fletcher TP, Smith S, Koch LM, Kainz B, Rueckert D. SonoNET: real-time detection and localisation of fetal standard scan planes in freehand ultrasound. IEEE Trans Med Imag. 2017;36(11):2204–2215. doi: 10.1109/TMI.2017.2712367. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556
  • 21.He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: CVPR, pp 770–778
  • 22.Saeed SU, Fu Y, Baum Z, Yang Q, Rusu M, Fan RE, Sonn GA, Barratt DC, Hu Y (2021) Learning image quality assessment by reinforcing task amenable data selection. In: IPMI, pp 755–766

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

11548_2022_2642_MOESM1_ESM.pdf (5.7MB, pdf)

Supplementary information This manuscript has one accompanying supplementary file. (5,811 KB)


Articles from International Journal of Computer Assisted Radiology and Surgery are provided here courtesy of Springer

RESOURCES