Abstract
Imaging is important in cancer diagnostics. It takes a long period of medical training and clinical experience for radiologists to be able to accurately interpret diagnostic images. With the advance of big data analysis, machine learning and AI-based devices are currently under development and taking a role in imaging diagnostics. If an AI-based imaging device can read the image as accurately as experienced radiologists, it may be able to help radiologists increase the accuracy of their reading and manage their workloads. In this paper, we consider two potential study objectives of a clinical trial to evaluate an AI-based device for breast cancer diagnosis by comparing its concordance with human radiologists. We propose statistical design and analysis methods for each study objective. Extensive numerical studies are conducted to show that the proposed statistical testing methods control the type I error rate accurately and the design methods provide required sample sizes with statistical powers close to pre-specified nominal levels. The proposed methods were successfully used to design and analyze a real device trial.
Keywords: artificial intelligence (AI), breast cancer, clinical device trial, concordance rate, generalized estimating equation, sample size calculation, statistical test
1. Introduction
There are different types of device trials depending on the use of device and the study objectives. In this paper, we introduce statistical design and analysis methods for a trial on an artificial intelligence (AI)-based device for the diagnosis of breast cancer.
Imaging technologies play a major role in the diagnosis of breast cancer. The reading and interpretation of imaging requires intensive medical training and significant clinical experience. With the advance of big data analysis methods, machine learning and AI-based imaging systems are currently under active development [1]. If an AI-based imaging device can read the image as accurately as experienced radiologists, it may be able to help radiologists increase the accuracy of their reading, manage their workloads, or possibly replace radiologists in remote clinics that would not have an experienced radiologist available for consultation.
In the assessment of breast lesions, the BI-RADS reporting system and classification are widely used [2]. This system includes categories between 1 and 5 (benign to malignant) with a key diagnostic transition subdivided into categories 4a (low suspicion of malignancy), 4b (moderate suspicion) and 4c (high suspicion, greater than 50% likelihood but less than 95% likelihood of malignancy). Furthermore, the BI-RADS lexicon covers radiological descriptive features that are important in diagnostic assessments, and these vary by modality. Examples of ultrasound lexicon used in AI-based classifications is given in Table A1. The earlier approaches to breast ultrasound technology concentrated on the extraction of features of lesions such as size, shape, texture, and boundaries within a clustering or classification or rule-based decision making algorithms [3,4,5,6]. More recent developments in AI, machine learning, and deep learning systems have utilized layers of convolution neural network models, a variety of approaches and extensive training sets to produce differentiated output classifications [7,8].
In this paper, we consider the requirements for a clinical device trial to evaluate the performance of an AI-based imaging device using BI-RADS reporting system for the diagnosis of breast cancer. Since BI-RADS reporting system does not have a gold standard, we evaluate the performance of the device by how well its reading aligns with those of radiologists. We propose design and analysis methods for two different types of study objectives that can be used for such a trial. The first objective is to test if the reading of the AI-based device concurs with those of radiologists as much as the readings concord among radiologists. The second objective is to test if the reading of the AI-based device is more concordant with those of experienced radiologists than with those of junior radiologists. For each objective, we propose a statistical testing method and its sample size calculation formula. The proposed testing methods will be used to analyze the data for each of the five BI-RADS lexicon classification category listed in Table A1, but the sample size calculation for a trial may be conducted only for the most important one. The performance of these methods are evaluated using simulations.
2. Materials and Methods
We consider two types of study objectives to evaluate the performance of an AI-based device for the diagnosis of breast cancer. For each study objective, we propose a testing method and its sample size formula. Suppose that we have images from n subjects.
2.1. Objective 1: Is the Concordance Rate between the AI-Based Device and Radiologists as High as That among Radiologists?
The image of each subject is read by m radiologists and the AI-based device. BI-RADS lexicon does not have a gold standard. So, in order to validate a device with an AI-based algorithm, we should show that the reading of the device concurs with those with radiologists. For example, in Table A1, for BI-RADS lexicon classification, Shape, two radiologists will be declared to be concurrent for an image if they both read oval, round or irregular. The question is how high the concordance rate should be between the device and the radiologists. The concordance rate among radiologists is used as a reference.
For each category of BI-RADS lexicon classifications, let and denote the concordance rate among radiologists and that between radiologists and the device, respectively. Since the latter can not be higher than the former, we specify a similarity margin . That is, we will not be interested in the AI-based device if and will be highly interested in it if . So, we want to test a null hypothesis against the alternative hypothesis .
2.1.1. Statistical Testing Method
Suppose that there are n patients, and the image of each patient is read by the AI-based device and m radiologists. For subject and radiologist , let if radiologists j and concur and otherwise, and let if radiologist j and the device concur and otherwise. Note that we have and . Since and for , the number of informative concordance scores among m radiologists is for each image, the concordance rate among radiologists for subject i is estimated by
On the other hand, for subject i, the concordance rate between the device and m radiologists is estimated by
Using the images from n subjects, concordance rate among radiologists is estimated by
and that between the device and radiologists is estimated by
Those estimates are unbiased because
and
Since are independent random variables with mean , for large n by the central limit theorem,
is asymptotically normal with mean 0 and variance that can be consistently estimated by
Similarly, are independent random variables with mean , so that for large n,
is asymptotically normal with mean 0 and variance that can be consistently estimated by
Since each subject’s image is read by the device and radiologists, and are correlated. However, are independent, with mean 0 under the null hypothesis . Hence, by the central limit theorem under ,
is asymptotically normal with mean 0 and variance that can be consistently estimated by
Hence, we reject the null hypothesis if , where
and is the percentile of the standard normal distribution. Note that we use a 1-sided test because the hypotheses are 1-sided and to avoid too large a sample size with a small .
Note that and are the generalized estimating Equation [9] (GEE) estimators of and , respectively, using the working independent correlation. Furthermore, the robust estimator of by the GEE method is given as
Since is a consistent estimator of , is asymptotically identical to under . Hence, can be counted as a test statistic based on the GEE method with the working independent correlation.
2.1.2. Power and Sample Size Calculation
We calculate the sample size for the test statistic under a specific alternative hypothesis . An accurate sample size calculation for the statistical test requires specification of correlation coefficients between and , between and , and between and . The dependency between and is expected to be higher than that between and for because the former pair includes the same reader while the latter pair contains four different readers. Similarly, we expect that the dependency between and is expected to be higher than that between and for .
For a simplified sample size formula, we just specify . We define the correlation coefficients among the concordance scores , , , and , and . Appendix A.1 shows that we have is expressed as
Under , are independent random variables with mean 0, so that is asymptotically normal with mean 0 and variance that can be consistently estimated by . Since is asymptotically identical to under , it converges to . Hence, the power for a given sample size n is
| (1) |
where is the survivor function of the standard normal distribution and is the limit of .
By solving the power Equation (1) with respect to n, we obtain the required sample size for power
| (2) |
where, as shown in the Appendix A.2,
| (3) |
and
The process of calculating the required sample size is summarized as follows:
-
(1)
Specify (), expected concordance rate among radiologists , similarity margin and hypothetical correlation coefficients and .
-
(2)
Calculate using (3).
-
(3)
Obtain sample size using (2).
It may be difficult to specify the correlation coefficients and . If pilot data are available, we may estimate them from the pilot data. Otherwise, we may conduct a two-stage trial to estimate these correlation coefficients from the first stage data and recalculate the sample size for the whole trial based on the estimated correlation coefficients.
2.2. Objective 2: Is the AI-Based Device More Concordant with Experienced Radiologists Than with Junior Radiologists?
As another study objective, we may want to test if the reading of the AI-based device agrees more with those of experienced radiologists than with those of junior radiologists for each BI-RADS lexicon classification category.
Let and denote the concordance rate between the AI-based device and highly experienced radiologists and that between the AI-based device and less experienced radiologists, respectively. We want to test the null hypothesis against the alternative hypothesis .
2.2.1. Statistical Testing Method
Let m (= 5, say) denote the number of radiologists in each group (highly experienced group and less experienced group). For subject and senior radiologist , let if the reading by senior radiologist j and that by the AI-based device agree and =0 otherwise, and let if the reading by less experienced radiologist and that by the AI-based device agree and =0 otherwise. Then, we have and . Using the data from subject i,, is estimated by
and is estimated by
Using the whole data, we estimate and by
and
respectively. Note that those estimates are unbiased.
For large n under , is approximately normal with mean 0 and variance that can be estimated by
Hence, we reject if , where
Note that we use a standard 2-sided test since usually there is no small effect size issue in this case.
2.2.2. Power and Sample Size Calculation
We calculate the sample size under a specific alternative hypothesis . Since each subject’s image is read by all experienced and inexperienced radiologists as well as the device, are correlated.
Let denote the correlation coefficient between the concordance score between the AI-based device and a highly experienced radiologist and the concordance score between the device and another highly experienced radiologist, and denote the correlation coefficient between the AI-based device and a less experienced radiologist and the concordance score between the device and another less experienced radiologist. Furthermore, let for denote the correlation coefficient between the concordance score between the device and a highly experienced radiologist and that between the device and a less experienced radiologist. Let . As shown in Appendix A.3. is a function of , and .
Under , are independent random variables with mean 0, so that is asymptotically normal with mean 0 and variance that can be consistently estimated by . Note that is asymptotically identical to under , so that it converges to . Hence, the power for a given sample size n is
| (4) |
since is under , where is the limit of under .
By solving (4) with respect to n, we obtain the required sample size for power
| (5) |
Appendix A.4 shows that
| (6) |
where
and
under .
The process of calculating the required sample size is summarized as follows:
-
1.
Specify (), expected concordance rate between the AI-based device and a highly experienced radiologist , clinically meaningful difference in concordance rates and correlation coefficients and .
-
2.
Calculate using (6).
-
3.
Obtain the required sample size using (5).
It will be difficult to specify the correlation coefficients and . If pilot data are available, we may estimate them from the pilot data. Otherwise, we may use a two-stage design to estimate these correlation coefficients from the first stage data and calculate the sample size for the whole trial based on the estimated correlation coefficients.
3. Numerical Studies and Results
Note that our test statistics and sample size formulas are derived based on large sample approximations. We want to conduct simulation studies to evaluate their finite sample performance.
We consider the first type of study objective to test if the concordance rate between an AI-based device and radiologists is as high as that among radiologists. Suppose that each subject’s image is read by the AI-based device and radiologists. We set , , , and . Assuming , we calculate the correlation coefficients for a given . That is, we obtain , , and for , respectively.
For each design setting, we calculate the required sample size n using our proposed formula (2) and generate 10,000 simulation data sets of size n under the design setting and or . Then, we apply the statistical testing using to each simulation data set, and compute the empirical type I error () and power () by the proportion of samples that reject among the 10,000 samples simulated under and , respectively. The correlated concordance (binary) data are generated by first generating multivariate normal data and then dichotomizing them with corresponding proportion level [10].
Table 1 reports the sample size n, empirical type I error rate , and power under each design setting. We observe that the required sample size increases in and decreases in and . With other design parameters fixed, we have the same sample sizes for and . We have this result because, from (2), the sample size depends on only through . Since the empirical type I errors are very close to the nominal overall, our test statistic controls the type I error rate accurately. On the other hand, the empirical powers are close to the corresponding nominal level overall, so that we conclude that our sample size formula is accurate too.
Table 1.
Sample size (empirical type I error rate, empirical power), , under various design settings of () for the first type of study objective.
| 0.3 | 0.05 | 0.1 | ||
| 0.3 | ||||
| 0.5 | ||||
| 0.7 | ||||
| 0.1 | 0.1 | |||
| 0.3 | ||||
| 0.5 | ||||
| 0.7 | ||||
| 0.5 | 0.05 | 0.1 | ||
| 0.3 | ||||
| 0.5 | ||||
| 0.7 | ||||
| 0.1 | 0.1 | |||
| 0.3 | ||||
| 0.5 | ||||
| 0.7 | ||||
| 0.7 | 0.05 | 0.1 | ||
| 0.3 | ||||
| 0.5 | ||||
| 0.7 | ||||
| 0.1 | 0.1 | |||
| 0.3 | ||||
| 0.5 | ||||
| 0.7 |
Now we conduct simulations for the second type of study objective to test if the AI-based device is more concordant with experienced radiologists than with junior radiologists. We assume that each subject’s image is read by experienced radiologists and junior radiologists. We set , , , , and . Assuming , we solve the corresponding correlation coefficients for given . So, we have , , , and for , and , respectively. For each design setting, we calculate sample size n using (5), and generate 10,000 samples of size n under the design setting and or . We apply the test statistic to each sample, and calculate the empirical type I error rate and power under and , respectively.
Table 2 summarizes the required sample size n, and empirical type I error rate and power, , under each design setting. We observe that the required sample size increases in and decreases in and . Since the empirical type I errors are very close to the nominal overall, our test statistic controls the type I error accurately. On the other hand, the empirical powers are close to the corresponding nominal level overall, so that we conclude that our sample size formula is accurate too.
Table 2.
Sample size (empirical type I error rate, empirical power), , under various design settings of () for the second type of study objective.
| 0.3 | 0.05 | 0.1 | ||
| 0.3 | ||||
| 0.5 | ||||
| 0.7 | ||||
| 0.1 | 0.1 | |||
| 0.3 | ||||
| 0.5 | ||||
| 0.7 | ||||
| 0.5 | 0.05 | 0.1 | ||
| 0.3 | ||||
| 0.5 | ||||
| 0.7 | ||||
| 0.1 | 0.1 | |||
| 0.3 | ||||
| 0.5 | ||||
| 0.7 | ||||
| 0.7 | 0.05 | 0.1 | ||
| 0.3 | ||||
| 0.5 | ||||
| 0.7 | ||||
| 0.1 | 0.1 | |||
| 0.3 | ||||
| 0.5 | ||||
| 0.7 |
4. Discussion and Conclusions
Existing papers on comparing correlated concordance rates mainly focus on comparing two (or more) competitive diagnosis methods using their concordance rates with a gold standard on multiple sites [11]. In this paper, there is no gold standard and we compare the concordance rate between an AI-based diagnostic device and human radiologists and that among radiologists. We also compare the concordance rate between an AI-based diagnostic device and highly experienced radiologists and that between AI-based device and less experienced radiologists. In our design setting, each study subject has single site but is rated by the AI-based device and multiple human radiologists. We extend existing methods to perform design and analysis in this new setting.
We provide design and analysis plan for two types of study objectives to perform different comparisons of concordance between the AI-based diagnostic device and human radiologists. For each type of study objective, we propose a test statistic using GEE method with independent working correlation to account for the dependency in the observations from the device and the radiologists for each study subject, and derive its sample size formula based on large sample theory. Through extensive simulations, we show that the test statistics control the type I error accurately and the sample size formulas estimate sample sizes with powers close to the specified ones accounting for the dependency of images read by radiologists and device.
Since each subject’s image is read by the device and many radiologists, the concordance scores have complicated dependency structure, while the test statistics do not require specification of the multiple correlation coefficients by using the GEE method, the sample size formulas require specification of these correlation coefficients. Since it is difficult to accurately specify the correlation coefficients, we propose to conduct a two-stage device trial to estimate these correlation coefficients from the first stage data and recalculate the required sample size for the whole trial based on the estimated correlation coefficients.
We use concordance rate as a measure of agreement among multiple raters. Cohen’s kappa is another measure of agreement that is popularly used, e.g., Qureshi et al. [12]. Unlike concordance rate, however, it is not clear how similar two kappa values should be to conclude similarity of two different groups of raters. The proposed methods were successfully used by O’Connell et al. [8] to design and analyze a device trial.
Appendix A
Table A1.
Examples of ultrasound lexicon.
| Ground Truth | Lesion Type |
|---|---|
| Shape | Oval |
| Round | |
| Irregular | |
| Margin | Circumscribed |
| Indistinct | |
| Angular | |
| Microlobulated | |
| Spiculated | |
| Orientation | Parallel |
| Not parallel | |
| Echo pattern | Anechoic |
| Hypoechoic | |
| Complex cystic and solid | |
| Isoechoic | |
| Hyperechoic | |
| Heterogeneous | |
| Posterior features | No features |
| Enhancement | |
| Shadowing | |
| Combined pattern |
Appendix A.1. Derivation of ρ1
Since and , we have
Here
and
Hence,
Appendix A.2. The Limit of under
The limit of is its expected value . Since under , where ,
and
since and .
Appendix A.3. Derivation of ρ2
Since and ,
Here,
and, similarly,
Hence,
Appendix A.4. The Limit of under
The limit of is . Since under , . Here, ,
and, similarly,
since .
Author Contributions
Conceptualization, L.L. and S.-H.J.; methodology, L.L.; software, L.L.; validation, L.L. and S.-H.J.; formal analysis, L.L.; investigation, L.L. and S.-H.J.; resources, K.J.P. and S.-H.J.; data curation, L.L.; writing—original draft preparation, L.L.; writing—review and editing, S.-H.J.; visualization, L.L.; supervision, K.J.P. and S.-H.J. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Conflicts of Interest
The authors declare no conflict of interest.
Footnotes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Zhang Z., Sejdic E. Radiological images and machine learning: Trends, perspectives, and prospects. Comput. Biol. Med. 2019;108:354–370. doi: 10.1016/j.compbiomed.2019.02.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.DSickles E.A., D’Orsi C.J., Bassett L.W., Appleton C.M., Berg W.A., Burnside E.S. ACR BI-RADS®Atlas, Breast Imaging Reporting and Data System. American College of Radiology; Reston, VA, USA: 2013. [Google Scholar]
- 3.Wu W.J., Lin S.W., Moon W.K. Combining support vector machine with genetic algorithm to classify ultrasound breast tumor images. Comput. Med. Imaging Graph. 2012;36:627–633. doi: 10.1016/j.compmedimag.2012.07.004. [DOI] [PubMed] [Google Scholar]
- 4.Liu B., Cheng H.D., Huang J., Tian J., Tang X., Liu J. Fully automatic and segmentation-robust classification of breast tumors based on local texture analysis of ultrasound images. Pattern Recogn. 2010;43:280–298. doi: 10.1016/j.patcog.2009.06.002. [DOI] [Google Scholar]
- 5.Shan J., Cheng H.D., Wang Y. Completely automated segmentation approach for breast ultrasound images using multiple-domain features. Ultrasound Med. Biol. 2012;38:262–275. doi: 10.1016/j.ultrasmedbio.2011.10.022. [DOI] [PubMed] [Google Scholar]
- 6.Cheng H.D., Shan J., Ju W., Guo Y., Zhang L. Automated breast cancer detection and classification using ultrasound images: A survey. Pattern Recogn. 2010;43:299–317. doi: 10.1016/j.patcog.2009.05.012. [DOI] [Google Scholar]
- 7.Wu G.G., Zhou L.Q., Xu J.W., Wang J.Y., Wei Q., Deng Y.B., Cui X.W., Dietrich C.F. Artificial intelligence in breast ultrasound. World J. Radiol. 2019;11:19–26. doi: 10.4329/wjr.v11.i2.19. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.O’Connell A.M. Diagnostic Performance of An Artificial Intelligence System in Breast Ultrasound. J. Ultrasound Med. 2021 doi: 10.1002/jum.15684. [DOI] [PubMed] [Google Scholar]
- 9.Liang K.Y., Zeger S. Longitudinal data analysis using generalized linear models. Biometrika. 1986;73:13–22. doi: 10.1093/biomet/73.1.13. [DOI] [Google Scholar]
- 10.Emrich I.J., Piedmonte M.R. A method for generating high dimensional multivariate binary variables. Am. Stat. 1991;45:302–304. [Google Scholar]
- 11.Jung S.H., Barnhart H.X., Sohn I., Stinnett S.S., Wallace D.K. Sample Size for Comparing Correlated Concordance Rates. J. Biopharm. Stat. 2008;18:359–369. doi: 10.1080/10543400701697216. [DOI] [PubMed] [Google Scholar]
- 12.Qureshi A., Lakhtakia R., Bahri M.A., Al Haddabi I., Saparamadu A., Shalaby A., Al Riyami M., Rizvi G. Gleason’s Grading of Prostatic Adenocarcinoma: Inter-Observer Variation Among Seven Pathologists at a Tertiary Care Center in Oman. Asian Pac. J. Cancer Prev. 2016;17:4867–4868. doi: 10.22034/APJCP.2016.17.11.4867. [DOI] [PMC free article] [PubMed] [Google Scholar]
