Abstract
Background
The application of artificial intelligence (AI) in the ultrasound (US) diagnosis of breast cancer (BCa) is increasingly prevalent. However, the impact of US-probe frequencies on the diagnostic efficacy of AI models has not been clearly established.
Objectives
To explore the impact of using US-video of variable frequencies on the diagnostic efficacy of AI in breast US screening.
Methods
This study utilized different frequency US-probes (L14: frequency range: 3.0-14.0 MHz, central frequency 9 MHz, L9: frequency range: 2.5-9.0 MHz, central frequency 6.5 MHz and L13: frequency range: 3.6-13.5 MHz, central frequency 8 MHz, L7: frequency range: 3-7 MHz, central frequency 4.0 MHz, linear arrays) to collect breast-video and applied an entropy-based deep learning approach for evaluation. We analyzed the average two-dimensional image entropy (2-DIE) of these videos and the performance of AI models in processing videos from these different frequencies to assess how probe frequency affects AI diagnostic performance.
Results
The study found that in testing set 1, L9 was higher than L14 in average 2-DIE; in testing set 2, L13 was higher in average 2-DIE than L7. The diagnostic efficacy of US-data, utilized in AI model analysis, varied across different frequencies (AUC: L9 > L14: 0.849 vs. 0.784; L13 > L7: 0.920 vs. 0.887).
Conclusion
This study indicate that US-data acquired using probes with varying frequencies exhibit diverse average 2-DIE values, and datasets characterized by higher average 2-DIE demonstrate enhanced diagnostic outcomes in AI-driven BCa diagnosis. Unlike other studies, our research emphasizes the importance of US-probe frequency selection on AI model diagnostic performance, rather than focusing solely on the AI algorithms themselves. These insights offer a new perspective for early BCa screening and diagnosis and are of significant for future choices of US equipment and optimization of AI algorithms.
Supplementary Information
The online version contains supplementary material available at 10.1186/s12911-023-02404-z.
Key points
1. The study explored the impact of ultrasound images with different frequencies on the diagnostic efficacy of artificial intelligence.
2. Ultrasound images obtained with different frequency probes exhibited variable levels of average two-dimensional image entropy, influencing the diagnostic performance of artificial intelligence models in nuanced ways.
3. Datasets with higher average two-dimensional image entropy were associated with superior artificial intelligence breast diagnostic efficacy.
Abstract
The research on artificial intelligence-assisted breast diagnosis often relies on static images or dynamic videos obtained from ultrasound probes with different frequencies. However, the effect of frequency-induced image variations on the diagnostic performance of artificial intelligence models remains unclear. In this study, we aimed to explore the impact of using ultrasound images with variable frequencies on AI’s diagnostic efficacy in breast ultrasound screening. Our approach involved employing a video and entropy-based feature breast network to compare the diagnostic efficiency and average two-dimensional image entropy of the L14 (frequency range: 3.0-14.0 MHz, central frequency 9 MHz), L9 (frequency range: 2.5-9.0 MHz, central frequency 6.5 MHz) linear array probe and L13 (frequency range: 3.6-13.5 MHz, central frequency 8 MHz), and L7 (frequency range: 3-7 MHz, central frequency 4.0 MHz) linear array probes. The results revealed that the diagnostic efficiency of AI models differed based on the frequency of the ultrasound probe. It is noteworthy that ultrasound images acquired with different frequency probes exhibit different average two-dimensional image entropy, while higher average two-dimensional image entropy positively affect the diagnostic performance of the AI model. We concluded that a dataset with higher average two-dimensional image entropy is associated with superior diagnostic efficacy for AI-based breast diagnosis. These findings contribute to a better understanding of how ultrasound image variations impact AI-assisted breast diagnosis, potentially leading to improved breast cancer screening outcomes.
Introduction
Breast cancer (BCa) is the most prevalent cancer and the leading cause of cancer mortality in females worldwide [1, 2]. Early identification and intervention of BCa lead to significant improvements in 5-year relative survival rates [3, 4]. Ultrasound(US) is the imaging method of choice for the evaluation of breast disease because it is technically simple, cost-effective, and safe [5]. Also, the US is seen as a primary measure of BCa detection and mortality reduction [3]. However, US-image-based diagnosis of the breast greatly relies on the experiences of the sonographers [6], so it is significant to further explore the information carried by US images to enhance the detection rate and diagnostic accuracy of BCa in its early stages.
In recent years, artificial intelligence (AI) has brought opportunities for the advancement of medical imaging [7–10]. The algorithm is enabled to extract a large amount of information from medical images that cannot be observed by the naked eye for diagnosis and improve the computer detection rate of nodules [11–13]. US-based AI studies rely on the sonographers-selected images during the scanning process or partially on responsibility frames selected from video [14–21]. Therefore, the US image selection is particularly crucial in BCa AI diagnosis.
In information theory [22], entropy is the average amount of information contained in each received “message”. Image entropy (IE) is a statistical form of image features, reflecting the average amount of information in the image, which can reflect the distribution complexity of each pixel point of the image [23]. Most previous studies were image-based that required high-frequency probe acquisition as a dataset [24–26]. The principle of choosing a US probe is to ensure sufficient detection depth while maximizing the frequency to ensure the resolution of the US image [25]. Although the high-frequency probe images may aid the sonographer in making a diagnosis. Whether they are favorable for the training and diagnosis of AI models is not known yet. Earlier works [21] demonstrated that the richer the average information content of an image, the better its tumor classification. Thus, based on the principle that US low frequency corresponds to high penetration [27], US data obtained at different frequencies may carry different levels of information, thereby impacting the diagnosis of the AI model.
Therefore, this study introduces the feature entropy of breast US to calculate the magnitude of the average two-dimensional image entropy(2-DIE) at different frequencies. Further, to investigate whether the US images obtained at lower frequencies have higher average 2-DIE and are more beneficial to improve AI diagnosis.
Materials and methods
Participants
This retrospective research collected US videos examined at Shenzhen People’s Hospital from June 2021 to December 2021. As a retrospective study, informed consent was waived by the Medical Ethics Committee of Shenzhen People’s Hospital. All patient information was handled with strict confidentiality in compliance with ethical guidelines. The benignity and malignancy of the nodules obtained by the US were confirmed by pathology.
Inclusion criteria
(a) Simultaneous acquisition of US video images of tumors in the same patient with two different frequency probes. (b) US-detected nodules must be classified as 0, 2, 3, 4a, 4b, 4c, or 5 following the BI-RADS. (c) No biopsy or surgical treatment of the breast nodules is to be evaluated before the US scan. (d) Patients were biopsied or surgically treated within 1 week of US data acquisition, while pathological results were obtainable.
Exclusion criteria
(a) BI-RADS 6 in the US, DR (Digital mammography), or MRI (magnetic resonance imaging). (b) BI-RADS 1 in the US. (c) History of breast surgery. (d) Single-frequency US video data. (e) Missing pathological results. (f) Poor image quality.
In this study, a total of 668 breast tumors (260 malignancy and 408 benign) of US videos from 167 female patients were included and divided into 2 testing sets: (1) In testing set 1, breast US videos were obtained from Resona I9 (Mindray, China) with L14 (frequency range: 3.0-14.0 MHz, central frequency 9 MHz), L9 (frequency range: 2.5-9.0 MHz, central frequency 6.5 MHz) linear array probe. (2) Testing set 2, the data were obtained from DC-65 (Mindray, China) with L13 (frequency range: 3.6-13.5 MHz, central frequency 8 MHz), and L7 (frequency range: 3.0-7.0 MHz, central frequency 4.0 MHz) linear array probe, which was aimed to further evaluate our theory and discoveries.
Ultrasound examination and video acquisition
In this study, all US videos were acquired by the same radiologist with more than 10 years of experience. The researchers utilized 3 markers to localize the location of the mass during the collection process. Specifically, the largest section of the target tumor was first located using one of the frequency probes and marked on the body surface. Then a complete sweep was made horizontally along the largest section of the tumor to find two more markers ≥ 2 cm from the tumor margin, respectively. Finally, the whole tumor is swept along the marked direction, and the operation is repeated, keeping the direction and position of the probe consistent each time, until the US video acquisition of the four different frequencies is completed.
Processing of US-video and use of AI model
First, we conducted this study based on the constructed feature entropy breast network (FEBrNet), which inherits the pre-trained backbone of the fully connected layer and the weight-optimal model [21]. We use the AI model to select responsibility frames to reduce subjective dependence. Our method of selecting pivotal frames draws inspiration from established applications of entropy in information theory, such as in decision trees. Specifically, the Iterative Dichotomiser 3 (ID3) decision tree algorithm utilizes entropy to ascertain the most suitable parent node and its division. In our methodology, we aim to minimize the discrepancy between the FScore of the video and that of the chosen frame collection, where a smaller disparity indicates that the information content of the chosen frames closely mirrors that of the entire video. By incrementally adding frames to this collection, starting from one and increasing to n, and at each increment selecting the frame that least differs, we gradually form an optimally representative set of frames, each contributing unique features. Subsequently, for the final collection of these optimal frames, our study computes the two-dimensional image entropy for each frame using the FEBrNet model. We then determine the video’s image entropy by calculating the average two-dimensional image entropy (2-DIE) of all the chosen frames. Finally, pathological results were used as the gold standard to compare the ability of using image entropy of different frequencies in the differential diagnosis of benign and malignant breast tumors. The processing and validation of the data are based on the pre-trained entropy-based model (FEBrNet). For specific information about the model refer to this literature [21] and supplementary materials. We investigate and verify the effect of entropy on the diagnostic performance of AI models from the perspective of IE. US images obtained from US probes of different frequencies are various. The researchers compared the variations by collecting US data from the same patient at different frequencies simultaneously. This is used to research the difference in diagnostic efficacy of US images obtained at different frequencies for AI models. The flow chart is shown in Fig. 1.
Statistical analysis
Statistical analysis was performed using R 3.6.3 (Copyright (C) 2020 The R Foundation for Statistical Computing). The significance level was set at P < 0.05. A normality test was performed for each variable. T-test is used for the normally distributed numerical variables, the rank sum test is used for the non-normally distributed numerical variables, and the Chi-square test is used for the disordered classification variables. The paired sample t-test was used to compare the differences within the group. The specificity, sensitivity, accuracy, receiver operating characteristic curve (ROC), and area under the curve (AUC) were used to evaluate models.
Results
Participant characteristics
According to the inclusion and exclusion criteria, A total of 668 tumor videos from 167 patients were included in this study, including 260 videos of cancerous masses and 408 videos of benign tumors. There are 280 videos in the testing set 1 and 388 videos in testing set 2. Table 1 show the baseline distribution characteristics of the collected patients, respectively.
Table 1.
Testing set 1 | ||||
Variables |
Total (n = 70) |
Benign (n = 44) |
Malignant (n = 26) |
p |
Age, Mean ± SD | 43.29 ± 12.9 | 38.89 ± 11.42 | 50.73 ± 11.99 | < 0.01 |
Height, Mean ± SD | 159.17 ± 3.86 | 159.57 ± 3.86 | 158.5 ± 3.85 | 0.27 |
Weight, Mean ± SD | 57.49 ± 6.97 | 56.27 ± 6.42 | 59.54 ± 7.5 | 0.07 |
BI-RADS, n (%) | < 0.001 | |||
2 | 14 (20) | 14 (32) | 0 (0) | |
3 | 14 (20) | 14 (32) | 0 (0) | |
4 A | 11 (16) | 9 (20) | 2 (8) | |
4B | 14 (20) | 7 (16) | 7 (27) | |
4 C | 8 (11) | 0 (0) | 8 (31) | |
5 | 9 (13) | 0 (0) | 9 (35) | |
Max. size, Median (Q1, Q3) | 8 (6, 14) | 7 (5.75, 11) | 12.5 (8, 15) | 0.001 |
Testing set 2 | ||||
Variables |
Total (n = 97) |
Benign (n = 58) |
Malignant (n = 39) |
p |
Age, Median (Q1, Q3) | 40 (31, 50) | 36 (28.5, 41) | 49 (43, 58.5) | < 0.01 |
Height, Median (Q1, Q3) | 158 (155, 162) | 159 (156, 2.75) | 158 (155, 162) | 0.87 |
Weight, Median (Q1, Q3) | 57 (52, 62) | 55.5 (51, 60) | 60 (56.5, 64.5) | < 0.01 |
BI-RADS, n (%) | < 0.01 | |||
2 | 20 (21) | 20 (34) | 0 (0) | |
3 | 19 (20) | 19 (33) | 0 (0) | |
4 A | 14 (14) | 13 (22) | 1 (3) | |
4B | 15 (15) | 6 (10) | 9 (23) | |
4 C | 15 (15) | 0 (0) | 15 (38) | |
5 | 14 (14) | 0 (0) | 14 (36) | |
Max. size, Median (Q1, Q3) | 13 (9, 23) | 11.5 (7, 18.75) | 17 (12.5, 24) | < 0.01 |
Note: BI-RADS: Breast Imaging-Reporting and Data System
Distribution of 2-DIE in various frequencies
The values of the average 2-DIE obtained at different frequencies are variable. For testing set 1, the 2-DIE of the L9 linear probe was higher than that of L14 (Mean ± SD,11.49 ± 0.769 vs. 10.94 ± 0.835); For the testing set 2, the 2-DIE of the L13 linear probe was higher than that of L7 (Mean ± SD,11.82 ± 0.356 vs. 12.27 ± 0.476). This result is summarized in Fig. 2.
Diagnosis performance of AI models
The diagnostic efficacy of US data for AI models varies at distinct frequencies. For the testing set 1, L9 attained the best AUC (0.849), with a sensitivity of 76.9%, specificity of 93.2%, and accuracy of 87.1%. For the testing set 2, L13 reached the best AUC (0.920), sensitivity 89.7%, specificity 93.8%, as well as accuracy 91.0%. The detailed results are shown in Table 2; Fig. 3.
Table 2.
Model | AUC (95%CI) |
Sensitivity (%) |
Specificity (%) |
Accuracy (%) |
P value |
---|---|---|---|---|---|
L9 |
0.849 (0.735–0.962) |
76.9 | 93.2 | 87.1 | 0.4185# |
L14 |
0.784 (0.655–0.913) |
65.4 | 90.9 | 81.4 | |
L7 |
0.887 (0.820–0.953) |
71.8 | 93.4 | 85.0 | 0.0383* |
L13 |
0.920 (0.851–0.989) |
89.7 | 91.8 | 91.0 |
Note: AUC: area under the curve; 95% CI: 95% confidence interval; L9: L9 linear array probe, frequency range: 2.5-9.0 MHz, central frequency 6.5 MHz; L14: L14 linear array probe, frequency range: 3.0-14.0 MHz, central frequency 9 MHz; L7: L7 linear array probe, frequency range: 3-7 MHz, central frequency 4.0 MHz; L13: L13 linear array probe, frequency range: 3.6-13.5 MHz, central frequency 8 MHz; P value #: L9 vs. L14; *: L7 vs. L13
Discussion
In this study, we used a video and entropy-based deep learning model [21] to compare the diagnosis performance of breast US. The assessment effect of variable frequencies on the AI model diagnosis validity was based on two retrospective data sets (Mindray L7/L13 and L9/L14). In testing set 1, compared to L14 (frequency range: 3.0-14.0 MHz, central frequency 9 MHz), the L9-had better diagnosis performance and 2-DIE. However, in testing set 2, compared to L7 (frequency range: 3-7 MHz, central frequency 4.0 MHz), the L13-had better diagnosis performance and 2-DIE. This observation suggests that US-data derived from probes operating at varying frequencies can significantly impact the diagnostic effectiveness of AI models. Another finding is that higher 2-DIE is accompanied by increased diagnostic efficacy.
In recent years, many AI-based studies have investigated the benign and malignant categorization of US breast nodules [18, 28–37]. The accuracy of their models fluctuates from 80 to 95%. While the literature recommends a frequency range of 5–17 [38] for breast US screening, it does not specify which one to use. Also, there is no literature examining the distinction in the diagnostic utility of AI for images acquired by various probes. The probe frequencies used in the studies in the literature mentioned above ranged from 1 to 42 MHz. Therefore, one reason for the variation in accuracy between these surveys may be the variance in the frequency of the probes utilized. So, we did this experience and discovered that the L9 (frequency range: 2.5-9.0 MHz, central frequency 6.5 MHz) had better diagnosis performance and higher 2-DIE. This may be contradicted by our clinical experience but offers another probability. While high-frequency US probes typically require more sophisticated technology and may be costlier, their application might not always correspond to improved diagnostic performance in AI models. Furthermore, primary hospitals may be unable to afford the purchase and maintenance of high frequency probe. The results of this experiment may now solve this challenge – using AI to aid diagnosis and compensate for the low accuracy of clinicians when using low frequency probe. On the other hand, excessively low-frequency probes do not enhance the diagnostic performance of AI models. In the testing set 2, we found that the L13 (frequency range: 3.6-13.5 MHz, central frequency 8 MHz) had better diagnosis performance. This is inconsistent with the results of our testing set 1. Possible reasons for this result include (a. The penetration is excessive, resulting in images that contain more confounding information unrelated to the lesion. b. The high frequency probe provides excellent spatial and soft-tissue resolution, greatly improving the differentiation of lesion saliency. However, the results for the 2-DIE of the L13 are higher than those of the low-frequency ones, which remains consistent with our previous findings.
Also, previous studies [21] by our team demonstrated that the richer the 2-DIE contained in US images, the more favorable the prediction of breast tumor benignity-malignancy. That is, a high 2-DIE in US images corresponds to rich image information. Meanwhile, US features varied depending on the pathological heterogeneity of the breast tumor [39]. The richer the information contained in the US image, the more comprehensive the information it may contain about the tumor characteristics. Moon et al. [34] also indicated that images with more information would help improve the diagnostic efficacy of the model. The results of this experiment are consistent with previous studies – both in the two sets, the higher the 2-DIE, the better the diagnostic performance. Because of the higher penetration of the low frequency probe, visualization of deep posterior tissues is made easy. More information related to the nodules may be captured. This information may not be recognized by the naked eye but facilitates machine learning. Accompanying the development of AI and the concept of medical-industrial integration, the application of AI-assisted diagnosis may become more extensive. However, previous research has focused more on the innovation and refinement of algorithms and hardware, ignoring the differences in images of different frequencies. Therefore, it is necessary to investigate the diagnostic efficacy of different frequency datasets on AI models. Images acquired at more appropriate frequencies will help improve diagnostic performance and provide a reference for future US image acquisition for AI models.
There are some limitations in this study. First, the study was a retrospective single-center study with smaller sample size and uneven image quality. Second, Lack of comparison of diagnostic efficacy of different frequencies for AI models. Therefore, we will further investigate the effect of other frequencies on AI diagnostic efficacy in the next research plan. Finally, Variations in sensitivity and inter-machine variability of various US devices were not considered.
Conclusion
This study indicate that US-data acquired using probes with varying frequencies exhibit diverse average 2-DIE values, and datasets characterized by higher average 2-DIE demonstrate enhanced diagnostic outcomes in AI-driven BCa diagnosis. Unlike other studies, our research emphasizes the importance of US-probe frequency selection on AI model diagnostic performance, rather than focusing solely on the AI algorithms themselves. These insights offer a new perspective for early BCa screening and diagnosis and are of significant guidance for future choices of US equipment and optimization of AI algorithms.
Electronic supplementary material
Below is the link to the electronic supplementary material.
Acknowledgements
Not applicable.
Abbreviations
- BCa
breast cancer
- 2-DIE
two-dimensional image entropy
- US
ultrasound
- AI
artificial intelligence
- IE
Image entropy
- DR
Digital mammography
- MRI
magnetic resonance imaging
- L9
L9 linear array probe (frequency range:2.5-9.0 MHz, central frequency 6.5 MHz)
- L14
L14 linear array probe (frequency range:3.0-14.0 MHz, central frequency 9 MHz)
- L7
L7 linear array probe (frequency range:3-7 MHz, central frequency 4.0 MHz)
- L13
L13 linear array probe (frequency range:3.6-13.5 MHz, central frequency 8 MHz)
- BI-RADS
Breast Imaging-Reporting and Data System
- ROC
receiver operating characteristic curve
- CI
confidence interval
- AUC
area under the curve
- IQR
interquartile range
Author contributions
Study concept and design: ZH, KY. Data extraction: HW, JC. Data quality: ST, HT. Data analysis and interpretation: CC, SS, YJ. Manuscript preparation: ZH, KY. Manuscript review: JX, FD. Manuscript editing: ZH. All authors approved the final manuscript for submission.
Funding
This work is supported by the following grant: Clinical Scientist Training Program of Shenzhen People’s Hospital, SYWGSCGZH202202. The funding unit assumes the function of the research sponsor in this research.
Data Availability
The datasets used and analyzed during the current study are available from the corresponding author on reasonable request.
Data Availability
The datasets used and analyzed during the current study are available from the corresponding author on reasonable request.
Declarations
Ethics approval and consent to participate
This retrospective study was approved by the Institutional Review Board of the Shenzhen People’s Hospital, specifically the Medical Ethics Committee of Shenzhen People ‘s Hospital. Due to the retrospective nature of this study, the need for informed consent was waived by the Institutional Review Board of Shenzhen People’s Hospital, specifically by the Medical Ethics Committee of Shenzhen People’s Hospital. All methods were carried out in accordance with relevant guidelines and regulations.
Informed consent
This study was conducted in accordance with the Declaration of Helsinki and was approved by the institutional review board of the Shenzhen People’s Hospital. Informed consent was waived by the same ethics committee that approved the study.
Consent for publication
Not applicable.
Competing interests
The authors of this manuscript declare no relationships with any companies, whose products or services may be related to the subject matter of the article.
Footnotes
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Zhibin Huang and Keen Yang contributed equally to this work.
Contributor Information
Jinfeng Xu, Email: xujinfeng@yahoo.com.
Fajin Dong, Email: dongfajin@szhospital.com.
References
- 1.Xia C, Dong X, Li H, Cao M, Sun D, He S, Yang F, Yan X, Zhang S, Li N, et al. Cancer statistics in China and United States, 2022: profiles, trends, and determinants. Chin Med J (Engl) 2022;135(5):584–90. doi: 10.1097/CM9.0000000000002108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Cao W, Chen H-D, Yu Y-W, Li N, Chen W-Q. Changing profiles of cancer burden worldwide and in China: a secondary analysis of the global cancer statistics 2020. Chin Med J. 2021;134(7):783–91. doi: 10.1097/CM9.0000000000001474. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Cedolini C, Bertozzi S, Londero AP, Bernardi S, Seriau L, Concina S, Cattin F, Risaliti A. Type of Breast cancer diagnosis, screening, and survival. Clin Breast Cancer. 2014;14(4):235–40. doi: 10.1016/j.clbc.2014.02.004. [DOI] [PubMed] [Google Scholar]
- 4.Islami F, Ward EM, Sung H, Cronin KA, Tangka FKL, Sherman RL, Zhao J, Anderson RN, Henley SJ, Yabroff KR, et al. Annual Report to the Nation on the Status of Cancer, Part 1: National Cancer statistics. J Natl Cancer Inst. 2021;113(12):1648–69. doi: 10.1093/jnci/djab131. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Berg WA. Current status of Supplemental Screening in dense breasts. J Clin Oncology: Official J Am Soc Clin Oncol. 2016;34(16):1840–3. doi: 10.1200/JCO.2015.65.8674. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Sakamoto N, Tozaki M, Higa K, Abe S, Ozaki S, Fukuma E. False-negative ultrasound-guided vacuum-assisted biopsy of the breast: difference with US-detected and MRI-detected lesions. Breast cancer (Tokyo Japan) 2010;17(2):110–7. doi: 10.1007/s12282-009-0112-1. [DOI] [PubMed] [Google Scholar]
- 7.Sechopoulos I, Teuwen J, Mann R: Artificial intelligence for breast cancer detection in mammography and digital breast tomosynthesis: State of the art. SEMINARS IN CANCER BIOLOGY 2021, 72:214-225. [DOI] [PubMed]
- 8.Yadav N, Dass R, Virmani J. Despeckling filters applied to thyroid ultrasound images: a comparative analysis. MULTIMEDIA TOOLS AND APPLICATIONS. 2022;81(6):8905–37. doi: 10.1007/s11042-022-11965-6. [DOI] [Google Scholar]
- 9.Dass R, Yadav N. Image quality assessment parameters for despeckling filters. Procedia Comput Sci. 2020;167:2382–92. doi: 10.1016/j.procs.2020.03.291. [DOI] [Google Scholar]
- 10.Yadav N, Dass R, Virmani J. Deep learning-based CAD system design for thyroid Tumor characterization using ultrasound images. MULTIMEDIA TOOLS AND APPLICATIONS 2023:1–43.
- 11.Hinton G. Deep Learning-A Technology with the potential to Transform Health Care. JAMA. 2018;320(11):1101–2. doi: 10.1001/jama.2018.11100. [DOI] [PubMed] [Google Scholar]
- 12.Huang YQ, Liang CH, He L, Tian J, Liang CS, Chen X, Ma ZL, Liu ZY. Development and validation of a Radiomics Nomogram for Preoperative Prediction of Lymph Node Metastasis in Colorectal Cancer. J Clin Oncol. 2016;34(18):2157–64. doi: 10.1200/JCO.2015.65.9128. [DOI] [PubMed] [Google Scholar]
- 13.Ouyang D, He B, Ghorbani A, Yuan N, Ebinger J, Langlotz CP, Heidenreich PA, Harrington RA, Liang DH, Ashley EA, et al. Video-based AI for beat-to-beat assessment of cardiac function. Nature. 2020;580(7802):252–6. doi: 10.1038/s41586-020-2145-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Aljuaid H, Alturki N, Alsubaie N, Cavallaro L, Liotta A. Computer-aided diagnosis for Breast cancer classification using deep neural networks and transfer learning. Comput Methods Programs Biomed. 2022;223:106951. doi: 10.1016/j.cmpb.2022.106951. [DOI] [PubMed] [Google Scholar]
- 15.Bae MS, Kim HG. Breast Cancer Risk Prediction using deep learning. Radiology. 2021;301(3):559–60. doi: 10.1148/radiol.2021211446. [DOI] [PubMed] [Google Scholar]
- 16.Hayashida T, Odani E, Kikuchi M, Nagayama A, Seki T, Takahashi M, Futatsugi N, Matsumoto A, Murata T, Watanuki R, et al. Establishment of a deep-learning system to diagnose BI-RADS4a or higher using breast ultrasound for clinical application. Cancer Sci. 2022;113(10):3528–34. doi: 10.1111/cas.15511. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Jiang M, Li CL, Luo XM, Chuan ZR, Lv WZ, Li X, Cui XW, Dietrich CF. Ultrasound-based deep learning radiomics in the assessment of pathological complete response to neoadjuvant chemotherapy in locally advanced Breast cancer. Eur J Cancer. 2021;147:95–105. doi: 10.1016/j.ejca.2021.01.028. [DOI] [PubMed] [Google Scholar]
- 18.Qi X, Zhang L, Chen Y, Pi Y, Chen Y, Lv Q, Yi Z. Automated diagnosis of breast ultrasonography images using deep neural networks. Med Image Anal. 2019;52:185–98. doi: 10.1016/j.media.2018.12.006. [DOI] [PubMed] [Google Scholar]
- 19.Qian X, Pei J, Zheng H, Xie X, Yan L, Zhang H, Han C, Gao X, Zhang H, Zheng W, et al. Prospective assessment of Breast cancer risk from multimodal multiview ultrasound images via clinically applicable deep learning. Nat Biomed Eng. 2021;5(6):522–32. doi: 10.1038/s41551-021-00711-2. [DOI] [PubMed] [Google Scholar]
- 20.Zheng X, Yao Z, Huang Y, Yu Y, Wang Y, Liu Y, Mao R, Li F, Xiao Y, Wang Y, et al. Deep learning radiomics can predict axillary lymph node status in early-stage Breast cancer. Nat Commun. 2020;11(1):1236. doi: 10.1038/s41467-020-15027-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Chen J, Jiang Y, Yang K, Ye X, Cui C, Shi S, Wu H, Tian H, Song D, Yao J, et al. Feasibility of using AI to auto-catch responsible frames in ultrasound screening for Breast cancer diagnosis. iScience. 2023;26(1):105692. doi: 10.1016/j.isci.2022.105692. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Shannon CE. A mathematical theory of communication. Bell Syst Tech J. 1948;27(3):379–423. doi: 10.1002/j.1538-7305.1948.tb01338.x. [DOI] [Google Scholar]
- 23.Chen X, Zhang Q, Lin M, Yang G, He C. No-reference color image quality assessment: from entropy to perceptual quality. EURASIP J Image Video Process. 2019;2019(1):1–14. doi: 10.1186/s13640-019-0479-7. [DOI] [Google Scholar]
- 24.Berg WA, Mendelson EB. Technologist-performed handheld screening breast US imaging: how is it performed and what are the outcomes to date? Radiology. 2014;272(1):12–27. doi: 10.1148/radiol.14132628. [DOI] [PubMed] [Google Scholar]
- 25.Berg WA, Bandos AI, Mendelson EB, Lehrer D, Jong RA, Pisano ED. Ultrasound as the primary screening test for Breast Cancer: analysis from ACRIN 6666. J Natl Cancer Inst 2016, 108(4). [DOI] [PMC free article] [PubMed]
- 26.Bevers TB, Helvie M, Bonaccio E, Calhoun KE, Daly MB, Farrar WB, Garber JE, Gray R, Greenberg CC, Greenup R, et al. Breast Cancer screening and diagnosis, Version 3.2018, NCCN Clinical Practice guidelines in Oncology. J Natl Compr Canc Netw. 2018;16(11):1362–89. doi: 10.6004/jnccn.2018.0083. [DOI] [PubMed] [Google Scholar]
- 27.Popp RL, Macovski A. Ultrasonic diagnostic instruments. Sci (New York NY) 1980;210(4467):268–73. doi: 10.1126/science.7423186. [DOI] [PubMed] [Google Scholar]
- 28.Cheng J-Z, Ni D, Chou Y-H, Qin J, Tiu C-M, Chang Y-C, Huang C-S, Shen D, Chen C-M. Computer-aided diagnosis with Deep Learning Architecture: applications to breast lesions in US images and pulmonary nodules in CT scans. Sci Rep. 2016;6:24454. doi: 10.1038/srep24454. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Seung Yeon S, Soochahn L, Il Dong Y, Sun Mi K, Kyoung Mu L. Joint weakly and semi-supervised deep learning for localization and classification of masses in breast ultrasound images. IEEE Trans Med Imaging. 2019;38(3):762–74. doi: 10.1109/TMI.2018.2872031. [DOI] [PubMed] [Google Scholar]
- 30.Shan J, Alam SK, Garra B, Zhang Y, Ahmed T. Computer-aided diagnosis for breast Ultrasound using computerized BI-RADS features and machine learning methods. Ultrasound Med Biol. 2016;42(4):980–8. doi: 10.1016/j.ultrasmedbio.2015.11.016. [DOI] [PubMed] [Google Scholar]
- 31.Han S, Kang HK, Jeong JY, Park MH, Kim W, Bang WC, Seong YK. A deep learning framework for supporting the classification of breast lesions in ultrasound images. Phys Med Biol. 2017;62(19):7714–28. doi: 10.1088/1361-6560/aa82ec. [DOI] [PubMed] [Google Scholar]
- 32.Tanaka H, Chiu S-W, Watanabe T, Kaoku S, Yamaguchi T. Computer-aided diagnosis system for breast ultrasound images using deep learning. Phys Med Biol. 2019;64(23):235013. doi: 10.1088/1361-6560/ab5093. [DOI] [PubMed] [Google Scholar]
- 33.Zhang E, Seiler S, Chen M, Lu W, Gu X. BIRADS features-oriented semi-supervised deep learning for breast ultrasound computer-aided diagnosis. Phys Med Biol. 2020;65(12):125005. doi: 10.1088/1361-6560/ab7e7d. [DOI] [PubMed] [Google Scholar]
- 34.Moon WK, Lee Y-W, Ke H-H, Lee SH, Huang C-S, Chang R-F. Computer-aided diagnosis of breast ultrasound images using ensemble learning from convolutional neural networks. Comput Methods Programs Biomed. 2020;190:105361. doi: 10.1016/j.cmpb.2020.105361. [DOI] [PubMed] [Google Scholar]
- 35.Zhang Q, Xiao Y, Dai W, Suo J, Wang C, Shi J, Zheng H. Deep learning based classification of breast tumors with shear-wave elastography. Ultrasonics. 2016;72:150–7. doi: 10.1016/j.ultras.2016.08.004. [DOI] [PubMed] [Google Scholar]
- 36.Qian X, Zhang B, Liu S, Wang Y, Chen X, Liu J, Yang Y, Chen X, Wei Y, Xiao Q, et al. A combined ultrasonic B-mode and color doppler system for the classification of breast masses using neural network. Eur Radiol. 2020;30(5):3023–33. doi: 10.1007/s00330-019-06610-0. [DOI] [PubMed] [Google Scholar]
- 37.Zhou Y, Xu J, Liu Q, Li C, Liu Z, Wang M, Zheng H, Wang S. A Radiomics Approach with CNN for Shear-Wave Elastography breast Tumor classification. IEEE Trans Bio Med Eng. 2018;65(9):1935–42. doi: 10.1109/TBME.2018.2844188. [DOI] [PubMed] [Google Scholar]
- 38.Hooley RJ, Scoutt LM, Philpotts LE. Breast ultrasonography: state of the art. Radiology. 2013;268(3):642–59. doi: 10.1148/radiol.13121606. [DOI] [PubMed] [Google Scholar]
- 39.Wang LC, Sullivan M, Du H, Feldman MI, Mendelson EB. US appearance of ductal carcinoma in situ. Radiographics: A Review Publication of the Radiological Society of North America Inc. 2013;33(1):213–28. doi: 10.1148/rg.331125092. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The datasets used and analyzed during the current study are available from the corresponding author on reasonable request.
The datasets used and analyzed during the current study are available from the corresponding author on reasonable request.