Summary
The research of AI-assisted breast diagnosis has primarily been based on static images. It is unclear whether it represents the best diagnosis image.To explore the method of capturing complementary responsible frames from breast ultrasound screening by using artificial intelligence. We used feature entropy breast network (FEBrNet) to select responsible frames from breast ultrasound screenings and compared the diagnostic performance of AI models based on FEBrNet-recommended frames, physician-selected frames, 5-frame interval-selected frames, all frames of video, as well as that of ultrasound and mammography specialists. The AUROC of AI model based on FEBrNet-recommended frames outperformed other frame set based AI models, as well as ultrasound and mammography physicians, indicating that FEBrNet can reach level of medical specialists in frame selection.FEBrNet model can extract video responsible frames for breast nodule diagnosis, whose performance is equivalent to the doctors selected responsible frames.
Subject areas: Computer-aided diagnosis method, Cancer, Artificial intelligence
Graphical abstract
Highlights
-
•
Feature entropy reduction can capture complementary responsibility frames from video
-
•
Extracted responsibility frames can be realized by reducing feature entropy
-
•
Pre-trained imaged AI model can be screened for the best model to nodule’s diagnose
-
•
Best AI model can diagnose breast nodules with accuracy up to a senior doctor
Computer-aided diagnosis method; Cancer; Artificial intelligence
Introduction
Breast cancer is now the most common malignancy in the world.1,2,3 It is the leading cause of death among females.4,5 Over 2.3 million new cases of breast cancer are diagnosed every year, according to the global cancer burden report for 2021,3 accounting for 30% of cancers in women and 11.7% of all patients with cancer worldwide, ranking first in cancer and endangering the lives and health of females. Effective screening can detect breast cancer early, reduce the local and long-term recurrence rates, and improve the five-year survival rate.6 Domestic and international guidelines recommend annual mammography for females aged 40 to 74 years, but the false-positive and false-negative rates for those with dense breasts are relatively high, making miss diagnoses more likely. As a supplementary diagnostic method, ultrasound is not limited by breast glandular tissue types and is especially appropriate for dense breast in Asian women, increasing the overall breast cancer detection rate by 17% and reducing unnecessary biopsies by 40%.7,8 Thus, breast ultrasound, along with mammography, clinical examination, and needle biopsy, plays an important role in the evaluation of breast disease. However, diagnosis based on breast ultrasound images heavily depends on the sonographer’s experience.
Artificial intelligence (AI) has developed rapidly in recent years. AI can provide individualized analysis and assist doctors in clinical decision-making.9,10,11 As a branch of AI, deep learning is able to recognize patterns in images without human intervention, facilitating high-throughput correlation between images and clinical data.12 Various AI techniques have been used to extract useful information from images, including convolutional neural networks (CNNs) and variable autoencoders.13,14,15 In the field of medical imaging, AI has made the most significant contributions to imaging medicine in disease severity classification and argan segmentation (or separation). Liver ultrasound images for fatty liver detection,16,17 ultrasound images for ovarian cancer detection and risk stratification,18,19 and chaotic ultrasound for stroke risk stratification are examples of disease classification applications.20
AI research of breast nodules has mostly relied on static images selected by doctors during scanning. Due to the operator dependence of ultrasound scans, doctors at various expertise levels may have different interpretations of the image and disease judgment. Thus, the responsible frames selected for decision-making can have inter-observer variation and junior physicians or non-professionals who lack necessary knowledge of breast ultrasound cannot choose a set of responsible frames for diagnosis. Since videos contain complete lesion information, whole videos can be used as inputs for AI without manual frame selection.21,22,23 It can broaden the scope of AI applications, especially in assisting less experienced operators that conduct breast diagnosis. However, ultrasound screening is a dynamic process, which records not only the important frames containing nodules but also the frames showing irrelevant tissue and ultrasound artifacts that can mislead AI. AI performance can be improved, if it can self-identify a set of representative frames to describe the whole lesion, just like the frames senior physicians selected.
Currently, video classification models primarily employed the fixed interval time-based method to subsample the frames. Although it reduced the number of frames for analysis, it fails to solve the existence of noisy frames and the problem of repeatedly choosing frames sharing very similar features. To address the issues, we combine greedy algorithm and the idea of information entropy to propose a model called FEBrNet for responsible frame selection. Information entropy (IE) is the average amount of the information contained in each “message” received in information field.24 Video is considered a collection of images, and each image’s information can be projected to multiple feature dimensions. Greedy algorithm is used to select a set of frames, which best summarizes the feature distribution of the whole video. Meanwhile, IE minimizing method is applied to decide when to stop picking a new responsible frame. In this study, we compared the performance of AI based on FEBrNet-recommended frames (R Frames) with AI based on all frames of video (All Frames), AI based on physician-selected images (Phy Images), AI based on fixed interval sampled frames (Fix Frames), ultrasound specialists, and mammography specialists. Area under receiver operating characteristics curve (AUROC), sensitivity, specificity, and accuracy are used as primary evaluating matrices. In 3-fold cross-validation and testing, AI can reach an equivalent performance using FEBrNet-recommended frames as using physician-selected frames and outperforming both ultrasound and mammography specialists.
Results
Comparison of clinical data
A total of 974 patients with breast nodules were included in this study. They were randomly assigned according to the training set and testing set with the ratio of 4:6 since random forest only requires limited training samples. Specifically, the training set contained 387 cases (174 malignant nodules), and the independent testing set contained 587 cases (238 malignant nodules). All patients underwent biopsy or surgery and obtained pathological diagnoses (Figure 1). There was no significant difference in age and nodule size between the patients included (p > 0.05).
Comparison of the effectiveness of FEBrNet’s responsible frame and others in training set by 3-fold cross-validation
The AUC (0.901 [95% CI: 0.877–0.925]) of FEBrNet’s responsible frame method is higher than that of the doctor frame selection method, fixed interval frame selection method, and all frames of video, p < 0.05. The cutoff value is 0.402, sensitivity: 84.5%, specificity: 80.6%, and accuracy: 82.3% (Table 1, Figure 2).
Table 1.
Modality | AUC (95%CI) | Cut-off | Sensitivity (%) | Specificity (%) | Accuracy (%) | p value |
---|---|---|---|---|---|---|
R_Frames | 0.901 (0.877–0.925) | 0.402 | 84.5 | 80.6 | 82.3 | NA |
Phy_Images | 0.857 (0.825–0.889) | 0.411 | 71.8 | 87.2 | 80.3 | 0.009 |
Fix_Frames | 0.815 (0.793–0.837) | 0.401 | 78.7 | 73.0 | 75.6 | 0.000 |
All_Frames | 0.889 (0.864–0.914) | 0.521 | 75.3 | 89.1 | 82.9 | 0.002 |
Note: AUC: area under the curve; 95% CI: 95% confidence interval; R_Frames: responsible frame; Phy_Images: doctor frame selection; Fix_Frames: fixed interval frame selection; All_Frames: all frames of video; p: R_Frames vs. others; NA: Not Applicable.
Comparison of the effectiveness of FEBrNet’s responsible frame and others AI models in the independent testing set
-
(1)
The AUC (0.912 [95% CI: 0.888–0.936]) of the responsible frame method was higher than that of the fixed interval frame selection and all video frames (p < 0.05). The cutoff value was 0.416, sensitivity: 84.4%, specificity: 87.4%, and accuracy: 86.2%.
-
(2)
The AUC of the responsible frame method is slightly higher than that of the model based on the frame selected by doctors (0.909 [95% CI: 0.884–0.935]) (p = 0.715 > 0.05), indicating that the frame selection level of AI model can reach the senior experts (Figures 3A–3D).
Comparison of the effectiveness of FEBrNet’s responsible frame and ultrasound and mammography specialists in the independent testing set
In this section, we reviewed the 383 cases where both ultrasound and mammography could be found.
-
(1)
The AUC of responsible frame method was better than that diagnosed by senior ultrasound doctors. The AUC was (0.912 [95% CI: 0.888–0.936] vs. 0.820 [95% CI: 0.788–0.853]) (p < 0.05).
-
(2)
The AUC diagnosed by FEBrNet model is better than that diagnosed by senior mammography doctors, and the AUC is (0.912 [95% CI: 0.888–0.936] vs. 0.796 [95% CI: 0.759–0.833]) (p < 0.05) (Figures 3E, 3F, and Table 2).
Table 2.
Modality | AUC (95%CI) | Cut-off | Sensitivity (%) | Specificity (%) | Accuracy (%) | p value |
---|---|---|---|---|---|---|
R_Frames | 0.912 (0.888–0.936) | 0.416 | 84.4 | 87.4 | 86.2 | NA |
Phy_Images | 0.909 (0.884–0.935) | 0.401 | 87.8 | 84.0 | 85.3 | 0.715 |
Fix_Frames | 0.852 (0.821–0.883) | 0.490 | 75.6 | 79.9 | 78.2 | 0.000 |
All_Frames | 0.881 (0.852–0.910) | 0.436 | 76.4 | 89.7 | 84.3 | 0.002 |
Ultrasound | 0.820 (0.788–0.853) | NA | 98.7 | 65.3 | 79.6 | 0.000 |
Mammography | 0.796 (0.759–0.833) | NA | 92.1 | 67.1 | 77.8 | 0.000 |
Note: AUC: area under the curve; 95% CI: 95% confidence interval; R_Frames: responsible frame; Phy_Images: doctor frame selection; Fix_Frames: fixed interval frame selection; All_Frames: all frames of video; Ultrasound: Diagnosis by senior ultrasound doctors; Mammography: Diagnosis by senior mammography doctors; p: R Frames vs. others; NA: Not Applicable.
Discussion
Based on our analysis of breast nodule static ultrasound images, a new model was constructed for this study: FEBrNet, which is an AI method for automatically capturing responsible frames in breast ultrasound videos.
In this study, FEBrNet was developed to analyze and extract responsible frames from ultrasound video data via defining the feature score (FS) and calculating entropy without adding additional training parameters, self-extracting multiple responsible frames with complementary information for the differential diagnosis of benign and malignant breast nodules. The results show that the AUC (0.912) of the AI model based on FEBrNet’s frames is higher than that of the fixed interval frames (0.852) and all video frames (0.881), indicating that the method proposed in this paper is feasible and highly reliable to assist breast diagnosis.
As one major contribution of this work, the FEBrNet addressed the issue of choosing responsible frames while avoiding selecting visually identical frames repetitively. As shown in Figure 4, we use a video taken from a 45-year-old female patient with BIRADS 4c and pathologically confirmed invasive breast cancer as a simple example to demonstrate the capacity of FEBrNet. When frames are sorted by FScore, the top three frames are frame 26, 39, and 27, with FScore of 21.94, 21.29, and 21.03. These frames seem to be relatively similar in Figure 4A and very close in time sequence. After using principal component analysis to compress and display the feature matrices into two dimensions in Figure 4B, it is obvious that the distance in feature dimensions between the three frames is rather close. In Figures 4C and 4D, the same approach is used to assess the top three responsible frames (frame 26, 111, and 96) chosen by FEBrNet. While the FScore of frame 111 is low, it is considerably distant from the first responsible frame (frame 26) in the feature dimensions. The top three frames selected by the FEBrNet are scattered, echoing their various visual attributes in Figure 4D.
Internationally recognized standard BIRADS classification for breast cancer risk is based on the characteristics of malignancy displayed in images.25,26 In this paper, the diagnostic efficiency of FEBrNet exceeds that of senior ultrasound and mammography doctors, which proves the reliable value of this technology in diagnosing breast cancer. The performance of AI algorithms depends on the training of large, labeled datasets.27,28,29 At present, a common problem is that it is difficult to find sufficient training data in the face of specific problems in a certain field. The emergence of transfer learning can alleviate the problem of insufficient data sources.29,30 Several CNN network models launched in recent years, such as Alexnet,31 VGG,32 GoogLeNet,27 ResNet,33 and Inception,34,35 can solve the visual classification task in natural images. This study used the static image pre-training and screening model and then inherits the weight of its backbone and full connection layer through the method of transfer learning (Figure 5). FEBrNet model was constructed to process video data and automatically extract responsible frames from video data and make the diagnosis (Figures 6 and 7); the results show that the model directly predicts the video responsible frame, which is very close to the AUC of the doctor’s frame selection method (0.909 [95% CI: 0.884–0.935]) (p = 0.715 > 0.05), which shows that AI model can reach the level of senior experts in frame selection. Mammography screening can detect early breast cancer and reduce breast cancer mortality by more than 30% in women over 40 years of age.36 Although mammography is the gold standard for breast cancer screening37 and has been proven to reduce mortality, the accuracy of results suffers among women with dense breasts.38 In the independent testing set, the AUC of the FEBrNet model responsible frame method was better than that of senior ultrasound and mammography doctors (p < 0.05).
Limitations of the study
-
(1)
The amount of breast ultrasound data in this study is limited. In the future, multi-center clinical research and video data of various instruments should be carried out to improve the effectiveness and robustness of training.
-
(2)
This study only focuses on the classification of benign and malignant breast nodules. In the future, it should be expanded to include image tasks such as infiltration range prediction, subtype classification, molecular phenotype prediction, and distant metastasis prediction.
-
(3)
The FEBrNet architecture can be used for ultrasound screening of various diseases, and more diseases should be evaluated in the future.
STAR★Methods
Key resources table
REAGENT or RESOURCE | SOURCE | IDENTIFIER |
---|---|---|
Software and algorithms | ||
Python | van Rossum,39 | https://www.python.org/ |
Tensorflow | Abadi et al.40 | https://www.tensorflow.org/ |
OpenCV | OpenCV team | https://opencv.org/ |
Matplotlib | Hunter,41 | https://matplotlib.org/ |
NumPy | Harris et al.42 | https://numpy.org/ |
Code for this paper | Jiang,43 | Code for FEBrNet - Mendeley Data |
Resource availability
Lead contact
Further information and requests for resources and should be directed to and will be fulfilled by the lead contact, FaJin Dong(dongfajin@szhospital.com).
Materials availability
This study did not generate new unique reagents.
Experimental model and subject details
Ethic statement
This study retrospectively analyzed the female patients between the ages of 24 and 75 who underwent breast ultrasound screening at Shenzhen People’s Hospital from July 2015 to December 2021. The study was approved by the Ethics Committee of Shenzhen People’s Hospital, and the requirement for informed consent was waived because of its retrospective design.
Inclusion criteria
(1)Breast nodules were detected by ultrasound, which were classified as 0, 3, 4a, 4b, 4c or 5 according to BIRADS; (2) At least 3.0 mm breast tissue can be displayed around the nodule; (3) No intervention or operation was performed on the nodule to be evaluated before ultrasonic examination; (4) Patients underwent surgery or biopsy within one week of ultrasonic data collection and obtained pathological results.
Exclusion criteria
(1) BIRADS 1 and 2; (2) Have a history of breast surgery or intervention; (3) Poor image quality; (4) The clinical data of cases are incomplete, and the pathological results are not tracked.
Patient anonymize and video preprocessing
Firstly, we cut the original video, remove the equipment information and patient sensitive information around the ultrasound image, and only keep the image window. Zoom video to 224 × 224 pixels, and then expand them frame by frame. These scaled frames will be used as the input data for the responsibility frame selection and comparison test of the new model (Figure 5).
Prior work
In our previous work,44 we had completed training and testing four AI models (Xception,45 MobileNet,46 ResNet50,47 and DenseNet12148) with three input image resolutions (224 × 224, 320 × 320, and 448 × 448 pixels). According to the inclusion and exclusion criteria, 3447 females breast nodules were finally included. Finally, we selected MobileNet in 224 × 224 pixels is the optimal model and image resolution, AUC: 0.893 (95% CI: 0.864-0.922), sensitivity: 81.6%, specificity: 82.1%, accuracy: 81.8%.
Method details
FEBrNet algorithm overview
In this study, we established the feature entropy breast network (FEBrNet), inheriting the backbone of the pretrained optimal models as feature extractor and adding a feature calculation layer at the last layer of models. The feature extractor processes each frame image and generates the corresponding feature matrix. We performed maxpooling on the feature matrix of all frames to get the reference feature matrix of entire video. A variable called FScore is defined by summing the values of malignant feature dimensions, indicating the total information of malignancy in feature matrix. The frame sharing most similar FScore to the reference feature matrix is selected as the first responsible frame and added to the responsible frame set. When adding a new frame into the responsible frame set, we use greedy algorithm to search all possible frames, which can minimize the differences in FScore of responsible frame set and the reference feature matrix. Meanwhile, for each added responsible frame, we calculate the malignant probability and the entropy of probability; once the entropy increases at a certain frame, the model’s certainty of prediction decreases, which is the stop signal for frame selection.
Identifying responsible frames using FScore
In our method, we focus on minimizing the difference between the FScore of the video and the FScore of the responsible frame collection, with a smaller difference suggesting that the information of the responsible frame collection is closer to the whole video. By gradually increasing the number of frames in the responsible frame collection from 1 to n, and in each step selecting the frame with the smallest difference and adding it to the responsible frame collection, we eventually obtain the local optimal responsible frame collection, with each frame contributing various features (Figure 6).
The figure below shows a basic example of picking the top two responsible frames from a video with three frames (Figure 7). The video feature matrix is constructed in step 1 by MaxPooling the three frame feature matrices, and the FScorevideo is 25. In step 2, we specify the number of frames in responsible frame collection as one, and there are three options with FScores of 16, 15, and 11. Since the difference between FScorevideo and FScoreframe1 is the lowest when selecting frame1, we choose frame1 as the first responsible frame. Step 3 increases the number of frames in the responsible frame collection to two and the first responsible frame (frame1) has already been chosen. FScore[frame1,frame2] is 17 and FScore[frame1,frame3] is 24, thus frame3 should be selected as the second responsible frame. Despite the fact that FScoreframe2 is larger, we will not choose frame2 as the second responsible frame for the benefit of adding frame2 to the responsible frame collection are few. Frame2 provides almost the same features as the already selected frame1, implying that they may also look similar.
In math, we defined FScore to quantify the contribution of a frame in certain feature dimension. We revised the equation of final prediction in neural network. The output of the CNN in the basic CNN architecture could be written as , where Y0 denotes the benign probability and Y1 is the malignant probability. The number of in the preceding equation represents the malignancy contribution of a single feature dimension, it could be separated into two parts: representing the amount of information that feature dimention i could contribute to final possibility of malignant and representing the intensity of the frame in feature dimention i. Meanwhile we hope to concentrate on the features indicating malignancy, therefore we use the equation to describe the contribution of framei in jth feature dimension, and the whole fearure matrix of framei can be represented as . To concentrate on the most representative characteristics of video, we used MaxPooling to alter the feature matrices of all frames and constructed the video feature matrix, where
Mathematically, we define the FScore of framei as . When it comes to a collection of frames A (A = [framea, frameb, … framec]), we use to express its FScore.
The process of finding the ith responsible frame can be described as the following equation:
Automatically stop frame selection using entropy
To set the auto-stop strategy for frame selection, we borrow the idea of entropy in information theory which is proposed by Shannon in 1948.24 In information theory, entropy is a measure of the uncertainty of random variables, and mathematically described as , where is the probability density of random variable X, and measure the amount of information which is also called self-information amount. We update the feature matrix every time the responsibility frame is added and calculate the entropy using malignant prediction value. With the increase of the number of responsibility frames, the prediction changes. When the certainty of malignant prediction value decreases, the entropy will increase, which is the stop signal (example Figure 1).
Experiment design
To evaluate whether FEBrNet-selected frames can be used as diagnostic frames, we trained AI models based on FEBrNet-recommended frames, all frames, fixed interval frames and physician-selected frames. Detailly, AI models based on different inputs have the same architecture, where pre-trained feature extractor process the inputted set of frames to generate feature matrix. Then, each model uses a random forest with same parameters (1024 estimators, max depth equals 10) to analyze the feature matrix of the frames set and generate predictions. Since the feature extractor is inherited, it requires no retraining process and only the random forests need training. Videos of all enrolled patients are divided into training and testing sets, where pathological results of breast nodules are used as ground truth. To ensure no crossover, the same patient can only appear in the same set. The results of models are compared, and an inference is that the model based on better-inputted frame set can achieve higher performance of AUROC.
First, we conducted three-fold cross-validation to evaluate models’ performances on the training. Then we compare the diagnostic performances of all models on testing set and the performance of real-world ultrasound and mammography doctors. The real-world ultrasound and mammography doctors’ performances are obtained by tracing back the recordings of patient’s final examination before biopsy and re-evaluated by senior physicians.
Quantification and statistical analysis
Continuous variable data are expressed as mean ± standard deviation. Categorical variable data is expressed as a percentage. The paired sample t-test was used to compare the differences within the group. R 3.6.3 was used for the statistical analysis. Firstly, draw the receiver operating characteristic curve (ROC), calculate and output the area under the curve (AUC), and 95% confidence interval (95% CI), p < 0.05 is statistically significant. Output the optimal cut-off value, specificity, sensitivity, and accuracy.
Acknowledgments
This project was supported by Commission of Science and Technology of Shenzhen (GJHZ20200731095401004). Our D.team provided all the data for the experiments. Thanks for our D.team of Ultrasound Department of Shenzhen People’s Hospital.
Author contributions
J.X.: Conceptualization, Methodology, Investigation, Funding acquisition. D.X.: Conceptualization, Methodology, Investigation, Supervision. F.D.: Conceptualization, Methodology, Investigation, Analyzed the data, Supervision. J.C.: Methodology, Writing – original draft, Writing – review & editing, Analyzed the data. Y.J.: Software, Writing – review & editing, Analyzed the data, Methodology. K.Y.: Resources, Analyzed the data, Methodology. X.Y.: Analyzed the data, Resources, Analyzed the data. S.S.: Software, Supervision, Analyzed the data. C.C.: Software, Supervision, Analyzed the data, Funding acquisition. H.W., H.T., D.S., J.Y., and S.H.: Analyzed the data, Resources.
Declaration of interests
The authors of this manuscript declare no relationships with any companies, whose products or services may be related to the subject matter of the article.
Inclusion and diversity
We support inclusive, diverse, and equitable conduct of research.
Published: January 20, 2023
Contributor Information
Jinfeng Xu, Email: xujinfeng@yahoo.com.
Dong Xu, Email: xudong@zjcc.org.cn.
Fajin Dong, Email: dongfajin@szhospital.com.
Data and code availability
Data is not publicly shared but available upon reasonable request from the lead contact.
Code of FEBrNet is available on Code for FEBrNet - Mendeley Data. Any additional information required to reanalyze the data reported in this work paper is available from the lead contact upon request.
References
- 1.Sung H., Ferlay J., Siegel R.L., Laversanne M., Soerjomataram I., Jemal A., Bray F. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA. Cancer J. Clin. 2021;71:209–249. doi: 10.3322/caac.21660. [DOI] [PubMed] [Google Scholar]
- 2.Siegel R.L., Miller K.D., Fuchs H.E., Jemal A. Cancer statistics, 2021. CA. Cancer J. Clin. 2021;71:7–33. doi: 10.3322/caac.21654. [DOI] [PubMed] [Google Scholar]
- 3.Siegel R.L., Miller K.D., Fuchs H.E., Jemal A. Cancer statistics, 2022. CA. Cancer J. Clin. 2022;72:7–33. doi: 10.3322/caac.21708. [DOI] [PubMed] [Google Scholar]
- 4.Maliszewska M., Maciążek-Jurczyk M., Pożycka J., Szkudlarek A., Chudzik M., Sułkowska A. Fluorometric investigation on the binding of letrozole and resveratrol with serum albumin. Protein Pept. Lett. 2016;23:867–877. doi: 10.2174/0929866523666160816153610. [DOI] [PubMed] [Google Scholar]
- 5.Chen W., Zheng R., Zhang S., Zeng H., Xia C., Zuo T., Yang Z., Zou X., He J. Cancer incidence and mortality in China, 2013. Cancer Lett. 2017;401:63–71. doi: 10.1016/j.canlet.2017.04.024. [DOI] [PubMed] [Google Scholar]
- 6.Cedolini C., Bertozzi S., Londero A.P., Bernardi S., Seriau L., Concina S., Cattin F., Risaliti A. Type of breast cancer diagnosis, screening, and survival. Clin. Breast Cancer. 2014;14:235–240. doi: 10.1016/j.clbc.2014.02.004. [DOI] [PubMed] [Google Scholar]
- 7.Niell B.L., Freer P.E., Weinfurtner R.J., Arleo E.K., Drukteinis J.S. Screening for breast cancer. Radiol. Clin. North Am. 2017;55:1145–1162. doi: 10.1016/j.rcl.2017.06.004. [DOI] [PubMed] [Google Scholar]
- 8.Osako T., Takahashi K., Iwase T., Iijima K., Miyagi Y., Nishimura S., Tada K., Makita M., Akiyama F., Sakamoto G., Kasumi F. Diagnostic ultrasonography and mammography for invasive and noninvasive breast cancer in women aged 30 to 39 years. Breast Cancer. 2007;14:229–233. doi: 10.2325/jbcs.891. [DOI] [PubMed] [Google Scholar]
- 9.Tonekaboni S., Joshi S., McCradden M.D., Goldenberg A. What clinicians want: contextualizing explainable machine learning for clinical end use. axRiv. 2019 doi: 10.48550/arXiv.1905.05134. Preprint at. [DOI] [Google Scholar]
- 10.Rajkomar A., Oren E., Chen K., Dai A.M., Hajaj N., Hardt M., Liu P.J., Liu X., Marcus J., Sun M., et al. Scalable and accurate deep learning with electronic health records. NPJ Digit. Med. 2018;1:18. doi: 10.1038/s41746-018-0029-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Suzuki K. Overview of deep learning in medical imaging. Radiol. Phys. Technol. 2017;10:257–273. doi: 10.1007/s12194-017-0406-5. [DOI] [PubMed] [Google Scholar]
- 12.Yasaka K., Akai H., Kunimatsu A., Kiryu S., Abe O. Deep learning with convolutional neural network in radiology. Jpn. J. Radiol. 2018;36:257–272. doi: 10.1007/s11604-018-0726-3. [DOI] [PubMed] [Google Scholar]
- 13.Hosny A., Parmar C., Quackenbush J., Schwartz L.H., Aerts H.J.W.L. Artificial intelligence in radiology. Nat. Rev. Cancer. 2018;18:500–510. doi: 10.1038/s41568-018-0016-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Mazurowski M.A., Buda M., Saha A., Bashir M.R. Deep learning in radiology: an overview of the concepts and a survey of the state of the art with focus on MRI. J. Magn. Reson. Imaging. 2019;49:939–954. doi: 10.1002/jmri.26534. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.He J., Baxter S.L., Xu J., Xu J., Zhou X., Zhang K. The practical implementation of artificial intelligence technologies in medicine. Nat. Med. 2019;25:30–36. doi: 10.1038/s41591-018-0307-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Acharya U.R., Sree S.V., Ribeiro R., Krishnamurthi G., Marinho R.T., Sanches J., Suri J.S. Data mining framework for fatty liver disease classification in ultrasound: a hybrid feature extraction paradigm. Med. Phys. 2012;39:4255–4264. doi: 10.1118/1.4725759. [DOI] [PubMed] [Google Scholar]
- 17.Saba L., Dey N., Ashour A.S., Samanta S., Nath S.S., Chakraborty S., Sanches J., Kumar D., Marinho R., Suri J.S. Automated stratification of liver disease in ultrasound: an online accurate feature classification paradigm. Comput. Methods Programs Biomed. 2016;130:118–134. doi: 10.1016/j.cmpb.2016.03.016. [DOI] [PubMed] [Google Scholar]
- 18.Acharya U.R., Sree S.V., Krishnan M.M.R., Saba L., Molinari F., Guerriero S., Suri J.S. Ovarian tumor characterization using 3D ultrasound. Technol. Cancer Res. Treat. 2012;11:543–552. doi: 10.7785/tcrt.2012.500272. [DOI] [PubMed] [Google Scholar]
- 19.Acharya U.R., Sree S.V., Kulshreshtha S., Molinari F., En Wei Koh J., Saba L., Suri J.S. GyneScan: an improved online paradigm for screening of ovarian cancer via tissue characterization. Technol. Cancer Res. Treat. 2014;13:529–539. doi: 10.7785/tcrtexpress.2013.600273. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Acharya R.U., Faust O., Alvin A.P.C., Sree S.V., Molinari F., Saba L., Nicolaides A., Suri J.S. Symptomatic vs. asymptomatic plaque classification in carotid ultrasound. J. Med. Syst. 2012;36:1861–1871. doi: 10.1007/s10916-010-9645-2. [DOI] [PubMed] [Google Scholar]
- 21.Furht B., Smoliar S.W., Zhang H. Springer Science & Business Media; 2012. Video and Image Processing in Multimedia Systems; p. 326. [Google Scholar]
- 22.Schmidhuber J. Deep learning in neural networks: an overview. Neural Netw. 2015;61:85–117. doi: 10.1016/j.neunet.2014.09.003. [DOI] [PubMed] [Google Scholar]
- 23.Golden J.A. Deep learning algorithms for detection of lymph node metastases from breast cancer: helping artificial intelligence be seen. JAMA. 2017;318:2184–2186. doi: 10.1001/jama.2017.14580. [DOI] [PubMed] [Google Scholar]
- 24.Shannon C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948;27:379–423. [Google Scholar]
- 25.Drukker L., Noble J.A., Papageorghiou A.T., et al. Introduction to artificial intelligence in ultrasound imaging in obstetrics and gynecology. Ultrasound Obstet. Gynecol. 2020;56:498–505. doi: 10.1002/uog.22122. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Muse E.D., Topol E.J. Guiding ultrasound image capture with artificial intelligence. Lancet. 2020;396:749. doi: 10.1016/S0140-6736(20)31875-4. [DOI] [PubMed] [Google Scholar]
- 27.Szegedy C., Liu W., Jia Y., Sermanet P., Reed S., Anguelov D., Erhan D., Vanhoucke V., Rabinovich A. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. IEEE; Boston, USA: 2015. Going deeper with convolutions. [Google Scholar]
- 28.Wu S., Zhong S., Liu Y. Deep residual learning for image steganalysis. Multimed. Tools Appl. 2018;77:10437–10453. [Google Scholar]
- 29.Aneja N., Aneja S. IEEE; 2019. Transfer Learning Using CNN for Handwritten Devanagari Character Recognition, 2019 1st International Conference on Advances in Information Technology (ICAIT) pp. 293–296. [Google Scholar]
- 30.Torrey L., Shavlik J. Transfer learning, Handbook of research on machine learning applications and trends: algorithms, methods, and techniques. IGI global. 2010:242–264. [Google Scholar]
- 31.Krizhevsky A., Sutskever I., Hinton G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM. 2017;60:84–90. [Google Scholar]
- 32.Simonyan K., Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv. 2014 doi: 10.48550/arXiv.1409.1556. Preprint at. [DOI] [Google Scholar]
- 33.He K., Zhang X., Ren S., Sun J. Deep residual learning for image recognition. arXiv. 2016 doi: 10.48550/arXiv.1512.03385. Preprint at. [DOI] [Google Scholar]
- 34.Xia X., Xu C., Nan B. IEEE; 2017. Inception-v3 for Flower Classification, 2017 2nd International Conference on Image, Vision and Computing (ICIVC) pp. 783–787. [Google Scholar]
- 35.Szegedy C., Vanhoucke V., Ioffe S., Shlens J., Wojna Z. Proceedings of the IEEE conference on computer vision and pattern recognition. IEEE; 2016. Rethinking the inception architecture for computer vision; pp. 2818–2826. [Google Scholar]
- 36.Tohno E., Umemoto T., Sasaki K., Morishima I., Ueno E. Effect of adding screening ultrasonography to screening mammography on patient recall and cancer detection rates: a retrospective study in Japan. Eur. J. Radiol. 2013;82:1227–1230. doi: 10.1016/j.ejrad.2013.02.007. [DOI] [PubMed] [Google Scholar]
- 37.Le-Petross H.T., Shetty M.K. Magnetic resonance imaging and breast ultrasonography as an adjunct to mammographic screening in high-risk patients. Semin. Ultrasound CT MRI. 2011;32:266–272. doi: 10.1053/j.sult.2011.03.005. [DOI] [PubMed] [Google Scholar]
- 38.Ohnuki K., Tohno E., Tsunoda H., Uematsu T., Nakajima Y. Overall assessment system of combined mammography and ultrasound for breast cancer screening in Japan. Breast Cancer. 2021;28:254–262. doi: 10.1007/s12282-020-01203-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.van Rossum G. Department of Computer Science [CS]; 1995. Python Reference Manual. [Google Scholar]
- 40.Abadi M., Barham P., Chen J., Chen Z., Davis A., Dean J., Devin M., Ghemawat S., Irving G., Isard M., Kudlur M. {TensorFlow}: a system for {Large-Scale} machine learning. In 12th USENIX symposium on operating systems design and implementation. OSDI. 2016;16:265–283. [Google Scholar]
- 41.Hunter . Computing in Science and Engineering. Vol. 9. IEEE; 2007. Matplotlib: A 2D Graphics Environment; pp. 90–95. [Google Scholar]
- 42.Harris C.R., Millman K.J., van der Walt S.J., Gommers R., Virtanen P., Cournapeau D., Wieser E., Taylor J., Berg S., Smith N.J., et al. Array programming with NumPy. Nature. 2020;585:357–362. doi: 10.1038/s41586-020-2649-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Jiang Y. Mendeley Data; 2022. Code for FEBrNet. [DOI] [Google Scholar]
- 44.Wu H., Ye X., Jiang Y., Tian H., Yang K., Cui C., Shi S., Liu Y., Huang S., Chen J., et al. A comparative study of multiple deep learning models based on multi-input resolution for breast ultrasound images. Front. Oncol. 2022;12:869421. doi: 10.3389/fonc.2022.869421. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Muhammad W., Aramvith S., Onoye T. Multi-scale Xception based depthwise separable convolution for single image super-resolution. PLoS One. 2021;16:e0249278. doi: 10.1371/journal.pone.0249278. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Wang W., Hu Y., Zou T., Liu H., Wang J., Wang X. A new image classification approach via improved MobileNet models with local receptive field expansion in shallow layers. Comput. Intell. Neurosci. 2020;2020:8817849. doi: 10.1155/2020/8817849. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Zhou J., Zhang Y., Chang K.T., Lee K.E., Wang O., Li J., Lin Y., Pan Z., Chang P., Chow D., et al. Diagnosis of benign and malignant breast lesions on DCE-MRI by using radiomics and deep learning with consideration of peritumor tissue. J. Magn. Reson. Imaging. 2020;51:798–809. doi: 10.1002/jmri.26981. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Ezzat D., Hassanien A.E., Ella H.A. An optimized deep learning architecture for the diagnosis of COVID-19 disease based on gravitational search optimization. Appl. Soft Comput. 2021;98:106742. doi: 10.1016/j.asoc.2020.106742. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Data is not publicly shared but available upon reasonable request from the lead contact.
Code of FEBrNet is available on Code for FEBrNet - Mendeley Data. Any additional information required to reanalyze the data reported in this work paper is available from the lead contact upon request.