Abstract
AI is becoming more prevalent in healthcare and is predicted to be further integrated into workflows to ease the pressure on an already stretched service. The National Health Service in the UK has prioritised AI and Digital health as part of its Long-Term Plan. Few studies have examined the human interaction with such systems in healthcare, despite reports of biases being present with the use of AI in other technologically advanced fields, such as finance and aviation. Understanding is needed of how certain user characteristics may impact how radiographers engage with AI systems in use in the clinical setting to mitigate against problems before they arise. The aim of this study is to determine correlations of skills, confidence in AI and perceived knowledge amongst student and qualified radiographers in the UK healthcare system. A machine learning based AI model was built to predict if the interpreter was either a student (n = 67) or a qualified radiographer (n = 39) in advance, using important variables from a feature selection technique named Boruta. A survey, which required the participant to interpret a series of plain radiographic examinations with and without AI assistance, was created on the Qualtrics survey platform and promoted via social media (Twitter/LinkedIn), therefore adopting convenience, snowball sampling This survey was open to all UK radiographers, including students and retired radiographers.
Pearson’s correlation analysis revealed that males who were proficient in their profession were more likely than females to trust AI. Trust in AI was negatively correlated with age and with level of experience. A machine learning model was built, the best model predicted the image interpreter to be qualified radiographers with 0.93 area under curve and a prediction accuracy of 93%. Further testing in prospective validation cohorts using a larger sample size is required to determine the clinical utility of the proposed machine learning model.
Author summary
Artificial Intelligence (AI) is becoming increasingly integrated into healthcare systems. Radiology, as a technologically advanced profession, is one area where AI has proved useful for myriad tasks. Developments in computer vision have allowed for applications in computer aided diagnosis using radiographic images. The integration and development of AI has been supported by healthcare providers and government agencies, such as the NHS in the UK. With the introduction of these systems in the clinical setting it is imperative to understand the nuances of the interaction between the clinicians using the technology and the system. Trust has been cited as a potentially significant issue in the effective integration of AI in radiology. Our research pinpoints the factors which have an impact on trust in radiographers and clarifies the strength of these associations. The means of analysis (Boruta) used provides an assessment of any correlations, allowing for intervention to be made. We found that females were less likely to trust AI that males and that trust in AI was negatively correlated with age and level of experience, indicating areas were further investigation and intervention are needed to ensure balanced trust and effective integration of AI in clinical radiography.
Introduction
Artificial Intelligence systems are becoming more integrated into healthcare settings [1]. These systems have been proposed as a means to ‘free up’ clinicians’ time, increase accuracy and reduce error [2], however, there have been few studies investigating the impact of these systems on the humans who are using them. Research has been conducted into the impact of computer assistance in other technologically advanced fields, such as aviation and the financial sector [3,4]. In healthcare, studies have focused on the impact of AI on decision making and Automation Bias (AB) [5,6]. These studies have highlighted some of the potential problems which may become more prevalent with the growing integration of AI systems in healthcare, however further exploration into other factors which may impact human behaviour when using computer assisted AI is lacking.
Decision support systems and AI are being developed in conjunction with clinicians in many areas of medicine. Recent developments in computer vision have permitted AI systems to be further developed for use in radiology, with promising results reported in many studies [7,8,9,10]. Neural networks are commonly used in computer vision tasks. This type of system comes with its own unique challenges, due in large part to the mode of operation, where some of the processing takes place in ‘hidden layers’, the so-called ‘black box’ of AI. This has led to a lack of trust from clinicians, who are expected to assume ultimate responsibility for the eventual diagnosis, whether assisted by AI or not.
The rate of the development of AI systems for use in healthcare has been increasing rapidly with many systems already in place [1]. Understanding of the human interaction with such systems will permit a critical approach and allow for an awareness of the characteristics of those who may be more susceptible to either over or under reliance on such technologies.
Increasing pressure on healthcare systems may provide much of the motivation for further integration of AI in medicine [11,12,13,14,15,16,17]. Staff shortages, coupled with increased demand for services mean that many countries are training and employing new staff in order to fill the backlog. Previous studies have found that less experienced clinicians may be more likely to trust computer assisted AI, inducing errors of both commission and omission [5,6,18,19]. Reliance on the AI diagnosis can be beneficial when the AI feedback is correct but can be detrimental when error exists in the advice given. Conversely, experienced clinicians report a lack of trust in AI systems and may, therefore, not derive the benefit of the assistance, choosing to rely on their own knowledge and experience. Hence, this study sought to identify potential user characteristics which might predispose them to certain behaviours when using AI e.g., lack of trust.
This paper answers the following research questions in relation to clinicians interacting with AI:
What are the clinical variables (features) that are significant when predicting student and qualified radiographers’ interactions.
What is the best predictive performance of a machine learning (ML) model in predicting performance of student and qualified radiographers.
What are the clinical correlations between student and qualified radiographers.
Using ML to discover knowledge in a radiographer dataset
Models built using ML algorithms require constant monitoring of their predictive performance when used in prediction or classification tasks [20] due to the probabilistic nature of risk prediction tools [21]. Monitoring tests such as stability of features are crucial in ML. There are different ML algorithms to address different types of analytical problems: from detection, localisation, assessment to prediction, classification, regression or density estimation tasks [22]. The predictive performance of a ML model can be enhanced by investigating the features that drive the prediction task [23]. This can lead to knowledge discovery, especially when we identify features or variables that stand out as important in prediction and classification tasks.
To assess the performance of the ML models, feature selection is often used. In most bioinformatics analysis [24], feature selection has become a pre-requisite for model building. Feature selection techniques select a subset of variables; in the case of projection, principal component analysis is evident and in the case of compression, information theory is used. The main objectives of feature selection are: a) To select the most predictive features, b) To improve model performance and to avoid over-fitting, and c) To provide more cost-effective and faster models.
There are three main types of feature selection methods, two of which are used here: (i) The filter-based feature selection method looks at the relevance of the features by using a scoring system. For example, Chi-square test is a non-parametric statistical test that assesses the dependence between the class and the feature as a measure of Chi-square statistics [25]. (ii) The wrapper-based feature selection method looks for a specific subset of features by training and testing the classification algorithm. For example, Boruta is a wrapper-based ML algorithm that calculates and displays the Z-score distribution of the input features [26].
Results
The correlation between continuous variables for all the participants of this study were investigated. Males who perceived themselves as proficient in their profession were more likely to trust AI as the male gender positively correlates with trust and with proficiency with a correlation coefficient (correlation = 0.67) >0. Trust was found to be negatively correlated with age and with level of experience with a correlation coefficient (correlation = -0.67 and -0.82 respectively) <0, meaning that the younger, less experienced participants reported less trust in AI. Age positively correlated with level of experience, i.e., the older respondents had more clinical experience. Age negatively correlated with practice, trust, males and proficiency, i.e., the younger participants were found to have less trust in AI, were more often females and reported themselves to be less professionally proficient. Based on work location of the country where the participants were located, females who had less practice were less likely to trust AI as the work location negatively correlates with female gender, trust and practice with a correlation coefficient (correlation = -0.42, -0.34 and -0.34, respectively) <0, meaning that there may be locations which have a great population of female radiographers and that there exists a variability of trust in AI in different geographical regions. However, we do note that the males and female ratio relative to average workforce hasn’t been recorded in our study. The variables: work location, age and level of experience were clustered together into one subgroup, while practice, trust, males, and proficiency were clustered together into another subgroup. (See Fig 1)
Further analysis
compared the importance of continuous and categorical clinical variables for predicting whether the interpreter was a student or qualified radiographers.
The distribution of Z-scores boxplots were ranked by the Boruta algorithm and revealed level of experience, age and work location to be the significant features (above the shadow max attribute). See Fig 2
High performing ML models useful to predict student and qualified radiographers
Five ML algorithms (Support Vector Machines, Naïve Bayes, Logistic Regression, k-Nearest Neighbour and Random Forest) were trained and tested for prediction of student and qualified radiographer groups. A detailed overview of the five comparative ML models’ performance evaluation metrics, consisting of Area Under the Receiver Operating Characteristic Curve (AUC), Classification Accuracy (CA), Mathews Correlation Coefficient (MCC), precision (sensitivity), recall (specificity) and F1 score built for predicting students and qualified are represented in Table 1. When classifying the student and qualified radiographers the Naïve Bayes model shows maximal sensitivity of 93%, specificity of 93%, a MCC score of 0.85 and an AUC of 0.93; outperforming other ML models in Table 1. The Boruta significant variables from Fig 2 were used to build these ML models.
Table 1. Evaluation metrics of the ML models performance built for student and qualified radiographer groups.
Comparison type | ML Model | Area Under Curve (AUC) | Classification Accuracy (CA) | Matthew’s Correlation Coefficient (MCC) | Positive Predictive Value (PPV) | Negative Predictive Value (NPV) | F1 score |
---|---|---|---|---|---|---|---|
Student versus qualified radiographers | SVM | 0.91 | 92.09±3.01% | 0.83 | 0.92 | 0.91 | 0.91 |
NB | 0.93 | 93.43±3.51% | 0.85 | 0.93 | 0.93 | 0.92 | |
K-NN | 0.91 | 92.09±3.01% | 0.83 | 0.92 | 0.91 | 0.91 | |
LR | 0.91 | 92.09±3.01% | 0.83 | 0.92 | 0.91 | 0.91 | |
RF | 0.91 | 92.09±3.01% | 0.83 | 0.92 | 0.91 | 0.91 |
ROC curve comparisons of the top ML algorithms using significant Boruta variables were performed for student versus qualified radiographers. The x-axis in Fig 3 denotes False Positive Rate (FPR) prediction and y-axis denotes True Positive Rate (TPR) prediction. The dotted lines in the figures represents the ROC curve for a random classification model (random performance). Legend denotes the Area Under Curve (AUC) values obtained with different ML algorithms colour coded for differentiation.
Discussion
In the present study, we have investigated the correlations of continuous variables for all the participants of this study using a novel ML approach for this task. Previous studies have identified the negative correlation between trust and age [27].
Males who were proficient in their profession were likely to trust AI. Literature suggests that in a male dominated respondents of a study, their trust level was statistically lower in usage of AI applications [28]. However, other studies have reported that radiologists who were more trained and informed in AI and have a higher trust in AI would be among those who would have higher adoption probability of AI applications [29]. This corroborates with our findings that males who were more proficient in their profession would likely trust AI.
In our findings, trust was negatively correlated with age and with level of experience. This could most likely be the case because expert radiologists would likely trust and rely on their own skills instead of trusting AI for applications [30]. This is supported by existing literature on the subject indicating that more experienced professionals are less likely to show preference for the input of AI in their diagnosis [5,6]. This is despite radiographers being accustomed to, and rely on, the use of technology in their day-to-day work. Furthermore, age was positively correlated with level of experience and negatively correlated with practice, trust, male and proficiency. Age and level of experience were found to be positively correlated, indicating that there is a greater proportion of young, newly qualified radiographers included in this study, rather than professionals who have entered the profession as mature students, as expert radiographers are more confident in AI in general with limitations in the ability to explain AI terminologies. However, younger radiographers miss out on the benefits of AI due to their inexperience and age.
Work location negatively correlated with females, trust and practice. The literature review suggests that female radiographers were likely to be unaware of using AI in clinical practice, due to their lack of trust in AI. Mistrust may be also due to the sparce uptake of technology in the UK healthcare system, despite of funding and focus from the government through the NHS [31]. Our findings indicate that female radiographers are less likely to use a highly complex AI system due to their lower trust.
The variables work location, age and level of experience were clustered together into one subgroup, while practice specialism (‘practice’), trust, males and proficiency were clustered together into another subgroup. The clustering of variables indicates similarity among variables.
The present results of ML AI predictive models are significant in at least two major respects. Firstly, the ML model built using algorithm Naive Bayes, has an excellent prediction accuracy of 93% and ROC curve area of 0.93 using the important Boruta ranked variables from student versus qualified radiographers. Furthermore, this approach may be used to test prospectively in clinical settings to predict qualified radiographer status, potentially earlier than is presently possible. The ML models built using the important Boruta ranked variables from student versus qualified radiographers can be improvised further with larger sample size and experimenting with other feature selection methods. Secondly, the identified important Boruta ranked variables can aid knowledge discovery and can be studied further in research settings. So far, ours is the first research design to have employed the usage of traditional ML based algorithms to predict student versus qualified radiographers using continuous variables. With the increasing use of new technologies in the clinical setting a simple and reliable means to identify any important corelations in how the human is predicted to interact with the system, and indeed, how the interaction may result in a negative patient outcome, will be useful to target education and training interventions. This may mean that those who are engaged as clinical experts in the new systems, as suggested by Rainey et al. (2022) and Strohm et al. (2020) [32,33] can have a more detailed knowledge of the specific intervention and support needed, both before implementation and during clinical use.
Materials and methods
Data collection and sample size of this study
Participants data were collected via the Qualtrics survey platform. The study was promoted via Twitter and LinkedIn to student, retired and practising radiographers. As a result of this, a total of 106 participants’ data was collected (students n = 67, radiographers n = 39). Participants were asked to provide binary diagnosis (pathology/no pathology) on three, randomly allocated, plain radiographic examinations from a dataset of 21 full examinations. Each image was presented with and without AI assistance in the form of heatmaps (Grad-CAM) and binary diagnosis. The AI model was based on a ResNet-152 architecture, with the arithmetic mean of the output of the AI used to determine pathology (threshold 0.5). Preliminary findings from this initial study were presented at the European Congress of Radiology [34].
The study was designed by the authors of this paper and piloted on a group of experienced radiographers (n = 3) and student radiographers from each year group of the Diagnostic Radiography and Imaging programme at Ulster University (n = 3). Face and content validity were ensured by request for feedback on the accessibility of the platform and comprehension of the content. Additionally, participants were asked to indicate the time commitment required to complete the survey. Three examinations were chosen as this was deemed an acceptable time sacrifice, supported by other studies [35]. No changes were made as a result of the pilot study as all participants were content that the survey was accessible, understood and measured the aim of the research.
The radiographic images used in the study were previously used in other studies [36]. The sim of the initial study was to clarify the impact of different forms of AI on student and qualified radiographers’ interaction with the technology.
Demographic information such as age, level of experience, work location, area of practice (or specialism) and proficiency in the use of technology in day-to-day life was collected. Trust in AI was determined on a scale of 0 to 5 ranging from not at all to absolute trust in AI for each examination. The survey was openly available from 2nd March to 2nd November 2021.
Correlation analysis
Pearson’s correlation coefficient was used to find correlations between the variables, and these are presented as heatmaps. Pearson’s correlation coefficient was also used to determine the direction and strength of association between variables; it measures if any linear association could occur between the variables under study [37]. Consequently, the variables (continuous and categoric) that were used in the correlation analysis were imputed using simple imputer function with the mean of the column, normalised using normalizer function and Pearson’s correlation matrix was calculated for all the participants of this study. Results were presented as heatmaps.
Clustering analysis
The Euclidean distance between each variable was computed, followed by hierarchical clustering, to check for any specific sub-clusters of participants in this study. A Euclidean distance metric was used to determine the distance between an existing datapoint and a new data point [37], which forms the basis of measure of similarity and dissimilarity between the two data points. Consequently, the continuous clinical variables and the categorical variables were imputed using simple imputer function, normalised using normalizer function with the mean of the column and Euclidean distance metric computed for all the participants of this study. Results were presented as heatmaps.
Boruta feature selection
The Boruta algorithm in R interface, which is a Random Forest based wrapper method, was used for feature selection [26]. Here, continuous and categorical variables were imputed using MICE imputation function for all the missing data, split based on 80% training and 20% test set. MICE imputation performs multiple regressions on random samples of the data and aggregates for imputing the missing values. Scaling of variables was performed using standard scaler function and the Boruta algorithm was applied for all the participants of this study.
Machine learning model development and evaluation
The Boruta significant variables were used as input features to develop five ML based models using following algorithms: Support Vector Machine (SVM), Naive Bayes (NB), K-Nearest Neighbour (K-NN), Logistic Regression (LR) and Random Forest (RF). The comparative model performance was assessed by Receiver Operator Characteristic (ROC) curves. Here, the gold standard refers to the most accurate and reliable method available for determining the student versus qualified radiographers. The ML models were built to assess the efficiency of the Boruta significant variables between student and qualified radiographers. Cross validation was used to enable a robust estimation of the performance of the ML model. Here, the ML models were built using a 3-fold stratified split cross validation and they were assessed based on their predictive performance using model evaluation metrics such as Area Under Curve (AUC), Classification Accuracy (CA), Matthew’s Correlation Coefficient (MCC), Positive Predictive Value (PPV), Negative Predictive Value (NPV) and F1 score.
To enable a robust estimation of the performance of a ML model, cross validation is performed. Depending upon the size of the dataset, different types of cross validation methods can be used, e.g. n-fold, random sampling, leave-one-out, etc. In the presented work, an n-fold cross validation method was performed, wherein at each step of cross validation, n-1 (i.e. total number of subsets—1) subsets are merged as the training set to train the model; while the remaining 1 subset is used as the test set to validate the trained model. This procedure is repeated n times such that each subset is used in training and then validated in testing [38].
Several metrics are used to evaluate the performance of a ML based model. The commonly used metrics for ML model evaluation are:
-
Positive Predictive Value (PPV): This test identifies the proportion of true positives out of the total of true (TP) and false (FP) positives [39].
PPV = TP / TP+FP
-
Negative Predictive Value (NPV): This test identifies the proportion of true negatives out of the total of true (TN) and false (FN) negatives [39].
NPV = TN / TN+FN
-
Classification Accuracy (CA): This determines the overall predictive accuracy of the ML model that has been trained [40].
CA = TP+TN / TP+TN+FP+FN
-
F1 score: Measures the accuracy of a test by calculating the true positives to the arithmetic mean of real positives (precision) and predicted positives (recall), wherein F1 score of 1 depicts best accuracy and 0 represents lowest accuracy [41].
F1 = (TN/(TN+FP)x (TP/(TP+FN))/(TN/(TN+FP))+(TP/(TP+FN))
-
Mathews Correlation Coefficient (MCC): It is widely used as a performance measure to validate predictive models. The MCC metric calculation uses four quantities (TP, TN, FP, FN) and gives in score ranging from -1 to 1, with 1 depicting complete agreement, -1 depicting complete disagreement and 0 depicting that the prediction is uncorrelated [42].
MCC = ((TPxTN)(FPxFN))/√((TP+FP)x(TP+FN)x(TN+FP)x(TN+FN))
-
Area Under Curve (AUC): The area under the Receiver Operating Characteristic (ROC) curve helps to compare the performance of different ML classifier models [43].
TPR = ∫ (T to ∞) f1(x)dx; FPR = ∫ (T to ∞) f0(x)dx; Area = P(X1>X0); where X1 is the score for positive instance and X0 is the score for negative instance.
Usage of software and libraries
All analyses were performed in Jupyter Notebook [44] using python [45] Version 2 and packages: pandas [46]; numpy [47]; SimpleImputer [48]; Normalizer [48]; seaborn [49]; matplotlib [50]; math [51]; mice [52]; train_test_split [53]; StandardScaler [54]; DataFrame [55]; and RStudio [56] Version 2022 and statistical functions and libraries: ranger [43] and Boruta package [57] was used for feature importance ranking and feature selection.
Acknowledgments
Ethical permission for this study has been granted by Ulster University Nursing and Health Research Ethics Filter Committee FCNUR-20-035. Ethical permission for the use of the clinical dataset images has been granted previously to use the images for research purposes (Monash University, Clayton, Australia, 2011).
Data Availability
Data cannot be shared publicly because of ethical restrictions on sharing sensitive data. Access to the data can be provided following a successful application to Ulster University’s Nursing and Health Research Ethics Filter Committee. Ulster University’s Research Portal contains metadata on the dataset and instructions on how to request access to this dataset. This information can be accessed at https://doi.org/10.21251/50890091-4b54-4644-b980-ea9da646aa0e.
Funding Statement
This work was supported by the College of Radiographers Research Industry Partnership award Scheme (CoRIPS: Grant number 213) (CR (P.I.), SMF, CH, RB, JMcC). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Mehrizi MH, van Ooijen P, Homan M. (2021) Applications of artificial intelligence (AI) in diagnostic radiology: a technography study. European radiology. Apr;31:1805–11. doi: 10.1007/s00330-020-07230-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.HEE (2019) The Topol review. Preparing the healthcare workforce to deliver the digital future 1–48.
- 3.Lui A, Lamb GW. (2018) Artificial intelligence and augmented intelligence collaboration: regaining trust and confidence in the financial sector. Information & Communications Technology Law. Sep 2;27(3):267–83. [Google Scholar]
- 4.Parasuraman R, Manzey DH. (2010) Complacency and bias in human use of automation: An attentional integration. Human factors. Jun;52(3):381–410. doi: 10.1177/0018720810376055 [DOI] [PubMed] [Google Scholar]
- 5.Goddard K, Roudsari A, Wyatt JC. (2014) Automation bias: empirical results assessing influencing factors. International Journal of Medical Informatics.1;83(5):368–75. doi: 10.1016/j.ijmedinf.2014.01.001 [DOI] [PubMed] [Google Scholar]
- 6.Bond RR, Novotny T, Andrsova I, Koc L, Sisakova M, Finlay D, Guldenring D, McLaughlin J, Peace A. McGilligan V. (2018). Automation bias in medicine: The influence of automated diagnoses on interpreter accuracy and uncertainty when reading electrocardiograms. Journal of Electrocardiology.;51:S6–11. doi: 10.1016/j.jelectrocard.2018.08.007 [DOI] [PubMed] [Google Scholar]
- 7.Qin C. Yao D., Shi Y., Song Z. (2018) Computer-aided detection in chest radiography based on artificial intelligence: a survey. Biomedical Engineering Online 17(1). doi: 10.1186/s12938-018-0544-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Tang Y-X, Tang Y-B, Peng Y., Yan K, Bagheri M., Redd B.A., Brandon C.J., Lu Z., Han M, Xiao J, Summers R.M. (2020) Automated abnormality classification of chest radiographs using deep convolutional neural networks. npj Digital Medicine. 3 (70). doi: 10.1038/s41746-020-0273-z [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Rainey C., McConnell J., Hughes C., Bond R., McFadden S., (2021) Artificial intelligence for diagnosis of fractures on plain radiographs: A scoping review of current literature. Intelligence-Based Medicine (5) 10.1016/j.ibmed.2021.100033 [DOI] [Google Scholar]
- 10.Lamb L., Lehman C.D., Gastounioti A., Conant E.F., Bahl M. (2022). Artificial Intelligence (AI) for Screening Mammography. AJR. https://www.ajronline.org/doi/10.2214/AJR.21.27071 [DOI] [PubMed] [Google Scholar]
- 11.Malamateniou C, Knapp KM, Pergola M, Woznitza N, Hardy M. (2021) Artificial intelligence in radiography: Where are we now and what does the future hold? Radiography. (Supplement 1). doi: 10.1016/j.radi.2021.07.015 [DOI] [PubMed] [Google Scholar]
- 12.Botwe B.O.; Akudjedu T.N.; Antwi W.K.; Rockson P.; Mkoloma S.S.; Balogun E.O.; Elshami W.; Bwambale J.; Barare C.; MdletsheB S. Yao S. Arkoh S. (2021)The integration of artificial intelligence in medical imaging practice: Perspectives of African radiographers. Radiography doi: 10.1016/j.radi.2021.01.008 [DOI] [PubMed] [Google Scholar]
- 13.Botwe B.O., Antwi W.K., Arkoh S. Akudjedu T.N. (2021) Radiographers’ perspectives on the emerging integration of artificial intelligence into diagnostic imaging: The Ghana study. Journal of Medical Radiation Sciences. doi: 10.1002/jmrs.460 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Antwi WK, Akudjedu TN, Botwe BO. (2021) Artificial intelligence in medical imaging practice in Africa: a qualitative content analysis study of radiographers’ perspectives. Insights Imaging. 16;12(1):80. doi: 10.1186/s13244-021-01028-z [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Malamateniou C, McFadden S., McQuinlan Y., England A., Woznitza N., Goldsworthy S., Currie C., Skelton E., Chu K-Y., Alware, N., Matthews, P., Hawkesford, R., Tucker, R., Town, W., Matthew, J., Kalinka, C., O’Regan, T. (2021a) Artificial Intelligence: Guidance for clinical imaging and therapeutic radiography professionals, a summary by the Society of Radiographers AI working group Radiography. Available at: https://www.sor.org/learning-advice/professional-body-guidance-and-publications/documents-and-publications/policy-guidance-document-library/artificial-intelligence-guidance-for-clinical-imag. [accessed 3rd December 2021]. [DOI] [PubMed] [Google Scholar]
- 16.Akudjedu TN, Torre S, Khine R, Katsifarakis D, Newman D, Malamateniou C(2022) Knowledge, perceptions, and expectations of Artificial intelligence in radiography practice: A global radiography workforce survey. 10.1016/j.jmir.2022.11.016 [DOI] [PubMed] [Google Scholar]
- 17.Rainey C., O’Regan T., Matthew J., Skelton E., Woznitza N., Chu K-Y, Goodman, S., McConnell, J., Hughes, C., Bond, R. Malamateniou, C., McFadden, S (2022) UK reporting radiographers’ perceptions of AI in radiographic image interpretation: Current perspectives and future developments. Radiography. Volume 28 doi: 10.1016/j.radi.2022.06.006 [DOI] [PubMed] [Google Scholar]
- 18.Logg JM, Minson JA, Moore DA. Algorithm appreciation: People prefer algorithmic to human judgment. Organizational Behavior and Human Decision Processes. 2019. Mar 1;151:90–103. [Google Scholar]
- 19.Dijkstra JJ, Liebrand WB, Timminga E. Persuasiveness of expert systems. Behaviour & Information Technology. 1998. Jan 1;17(3):155–63. [Google Scholar]
- 20.Breck E, Cai S, Nielsen E, Salib M, Sculley D. What’s your ML test score? A rubric for ML production systems. [Google Scholar]
- 21.Bello GA, Teitelbaum SL, Lucchini RG, Dasaro CR, Shapiro M, Kaplan JR, Crane MA, Harrison DJ, Luft BJ, Moline JM, Udasin IG. Assessment of cumulative health risk in the World Trade Center general responder cohort. American journal of industrial medicine. 2018. Jan;61(1):63–76. doi: 10.1002/ajim.22786 [DOI] [PubMed] [Google Scholar]
- 22.Worden K, Manson G. The application of machine learning to structural health monitoring. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences. 2007. Feb 15;365(1851):515–37. doi: 10.1098/rsta.2006.1938 [DOI] [PubMed] [Google Scholar]
- 23.Das R, Wales DJ. Machine learning landscapes and predictions for patient outcomes. Royal Society open science. 2017. Jul 26;4(7):170175. doi: 10.1098/rsos.170175 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Saeys Y, Inza I, Larranaga P. A review of feature selection techniques in bioinformatics. bioinformatics. 2007. Oct 1;23(19):2507–17. doi: 10.1093/bioinformatics/btm344 [DOI] [PubMed] [Google Scholar]
- 25.McHugh ML. The chi-square test of independence. Biochemia medica. 2013. Jun 15;23(2):143–9. doi: 10.11613/bm.2013.018 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Kursa MB, Rudnicki WR. Feature selection with the Boruta package. Journal of statistical software. 2010. Sep 16;36:1–3. [Google Scholar]
- 27.Rainey C, O’Regan T, Matthew J, Skelton E, Woznitza N, Chu KY, Goodman S, McConnell J, Hughes C, Bond R, McFadden S. Beauty is in the AI of the beholder: are we ready for the clinical integration of artificial intelligence in radiography? An exploratory analysis of perceived AI knowledge, skills, confidence, and education perspectives of UK radiographers. Frontiers in digital health. 2021. Nov 11;3:739327. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Qurashi AA, Alanazi RK, Alhazmi YM, Almohammadi AS, Alsharif WM, Alshamrani KM. Saudi radiology personnel’s perceptions of artificial intelligence implementation: a cross-sectional study. Journal of Multidisciplinary Healthcare. 2021. Nov 23:3225–31. doi: 10.2147/JMDH.S340786 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Von Wedel P, Hagist C. Physicians’ preferences and willingness to pay for artificial intelligence-based assistance tools: a discrete choice experiment among german radiologists. BMC Health Services Research. 2022. Dec;22(1):1–4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Jungmann F, Ziegelmayer S, Lohoefer FK, Metz S, Müller-Leisse C, Englmaier M, Makowski MR, Kaissis GA, Braren RF. Algorithmic transparency and interpretability measures improve radiologists’ performance in BI-RADS 4 classification. European Radiology. 2023. Mar;33(3):1844–51. doi: 10.1007/s00330-022-09165-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.York TJ, Raj S, Ashdown T, Jones G. Clinician and computer: a study on doctors’ perceptions of artificial intelligence in skeletal radiography. BMC Medical Education. 2023. Dec;23(1):1–0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Rainey C., O’Regan T., Matthew J., Skelton E., Woznitza N., Chu K-Y, Goodman S., McConnell J., Hughes C., Bond R. Malamateniou C., McFadden S. (2022. a) An insight into the current perceptions of UK radiographers on the future impact of AI on the profession: A cross-sectional survey. Journal of Medical Imaging and Radiation Sciences. 53 (3) doi: 10.1016/j.jmir.2022.05.010 [DOI] [PubMed] [Google Scholar]
- 33.Strohm L., Hehakaya C., Ranschaert E.R., Boon W.P.C., Moors E.H.M. (2020) Implementation of artificial intelligence (AI) applications in radiology: hindering and facilitating factors. European Radiology. 2020; 30. pp. 5525–5532. doi: 10.1007/s00330-020-06946-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Rainey C., McConnell J., Hughes C., Bond R., Kumar D., McFadden S (2023) The impact of AI feedback on the accuracy of diagnosis, decision switching and trust in Radiography. The European Congress of Radiology, Vienna, March, 2023. [Google Scholar]
- 35.Revilla M., Ochoa C. Ideal and Maximum Length for a Web Survey. International Journal of Market Research, 59(5), 557–565. 10.2501/IJMR-2017-039 (2017)) [DOI] [Google Scholar]
- 36.McConnell J., Devaney C., Gordon M. (2013) Queensland radiographer clinical descriptions of adult appendicular musculo-skeletal trauma following a condensed education programme. Radiography. 19(1) 48–55 10.1016/j.radi.2012.09.002 [DOI] [Google Scholar]
- 37.Schober P, Boer C, Schwarte LA. Correlation coefficients: appropriate use and interpretation. Anesthesia & analgesia. 2018. May 1;126(5):1763–8. doi: 10.1213/ANE.0000000000002864 [DOI] [PubMed] [Google Scholar]
- 38.Song J, Wang H, Wang J, Leier A, Marquez-Lago T, Yang B, Zhang Z, Akutsu T, Webb GI, Daly RJ. PhosphoPredict: A bioinformatics tool for prediction of human kinase-specific phosphorylation substrates and sites by integrating heterogeneous feature selection. Scientific reports. 2017. Jul 31;7(1):6862. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Loong TW. Understanding sensitivity and specificity with the right side of the brain. Bmj. 2003. Sep 25;327(7417):716–9. doi: 10.1136/bmj.327.7417.716 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Kim YK, Na KS. Application of machine learning classification for structural brain MRI in mood disorders: Critical review from a clinical perspective. Progress in Neuro-Psychopharmacology and Biological Psychiatry. 2018. Jan 3;80:71–80. doi: 10.1016/j.pnpbp.2017.06.024 [DOI] [PubMed] [Google Scholar]
- 41.Powers DM. Evaluation: from precision, recall and F-measure to ROC, informedness, markedness and correlation. arXiv preprint arXiv:2010.16061. 2020. Oct 11. [Google Scholar]
- 42.Allyn J, Allou N, Augustin P, Philip I, Martinet O, Belghiti M, Provenchere S, Montravers P, Ferdynus C. A comparison of a machine learning model with EuroSCORE II in predicting mortality after elective cardiac surgery: a decision curve analysis. PloS one. 2017. Jan 6;12(1):e0169772. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Boughorbel S, Jarray F, El-Anbari M. Optimal classifier for imbalanced data using Matthews Correlation Coefficient metric. PloS one. 2017. Jun 2;12(6):e0177678. doi: 10.1371/journal.pone.0177678 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Kluyver T, Ragan-Kelley B, Pérez F, Granger BE, Bussonnier M, Frederic J, Kelley K, Hamrick JB, Grout J, Corlay S, Ivanov P. Jupyter Notebooks-a publishing format for reproducible computational workflows. 2016. May 26. [Google Scholar]
- 45.Menczer F, Fortunato S, Davis CA. Python tutorial. A First Course in Network Science. 2020:221–37. [Google Scholar]
- 46.McKinney W. pandas: a foundational Python library for data analysis and statistics. Python for high performance and scientific computing. 2011. Nov 18;14(9):1–9. [Google Scholar]
- 47.Ascher D, Dubois PF, Hinsen K, Hugunin J, Oliphant T. Numerical python. [Google Scholar]
- 48.Subasi A. Practical machine learning for data analysis using python. Academic Press; 2020. Jun 5. [Google Scholar]
- 49.Waskom ML. Seaborn: statistical data visualization. Journal of Open Source Software. 2021. Apr 6;6(60):3021. [Google Scholar]
- 50.Barrett P, Hunter J, Miller JT, Hsu JC, Greenfield P. matplotlib—A Portable Python Plotting Package. In Astronomical data analysis software and systems XIV 2005. Dec (Vol. 347, p. 91). [Google Scholar]
- 51.Oliphant TE. Python for scientific computing. Computing in science & engineering. 2007. Jun 18;9(3):10–20. [Google Scholar]
- 52.Penone C, Davidson AD, Shoemaker KT, Di Marco M, Rondinini C, Brooks TM, Young BE, Graham CH, Costa GC. Imputation of missing data in life-history trait datasets: Which approach performs the best?. Methods in Ecology and Evolution. 2014. Sep;5(9):961–70. [Google Scholar]
- 53.De Vazelhes W, Carey CJ, Tang Y, Vauquier N, Bellet A. metric-learn: Metric learning algorithms in python. The Journal of Machine Learning Research. 2020. Jan 1;21(1):5447–52. [Google Scholar]
- 54.Brownlee J. Machine learning mastery with Python: understand your data, create accurate models, and work projects end-to-end. Machine Learning Mastery; 2016. Apr 8. [Google Scholar]
- 55.McKinney W. Data structures for statistical computing in python. InProceedings of the 9th Python in Science Conference 2010. Jun 28 (Vol. 445, No. 1, pp. 51–56). [Google Scholar]
- 56.Braun WJ, Murdoch DJ. A first course in statistical programming with R. Cambridge University Press; 2021. May 20. [Google Scholar]
- 57.Wright MN, Ziegler A. ranger: A fast implementation of random forests for high dimensional data in C++ and R. arXiv preprint arXiv:1508.04409. 2015. Aug 18. [Google Scholar]