Skip to main content
Journal of Applied Clinical Medical Physics logoLink to Journal of Applied Clinical Medical Physics
. 2008 Nov 11;9(4):181–199. doi: 10.1120/jacmp.v9i4.2741

A novel algorithm for initial lesion detection in ultrasound breast images

Moi Hoon Yap 1,, Eran A Edirisinghe 1, Helmut E Bez 1
PMCID: PMC5722351  PMID: 19020477

Abstract

This paper proposes a novel approach to initial lesion detection in ultrasound breast images. The objective is to automate the manual process of region of interest (ROI) labeling in computer‐aided diagnosis (CAD). We propose the use of hybrid filtering, multifractal processing, and thresholding segmentation in initial lesion detection and automated ROI labeling. We used 360 ultrasound breast images to evaluate the performance of the proposed approach. Images were preprocessed using histogram equalization before hybrid filtering and multifractal analysis were conducted. Subsequently, thresholding segmentation was applied on the image. Finally, the initial lesions are detected using a rule‐based approach. The accuracy of the automated ROI labeling was measured as an overlap of 0.4 with the lesion outline as compared with lesions labeled by an expert radiologist. We compared the performance of the proposed method with that of three state‐of‐the‐art methods, namely, the radial gradient index filtering technique, the local mean technique, and the fractal dimension technique. We conclude that the proposed method is more accurate and performs more effectively than do the benchmark algorithms considered.

PACS numbers: 87.57.Nk

Keywords: medical image analysis, ultrasound imaging, region‐of‐interest, hybrid filtering, multifractals

I. INTRODUCTION AND SCOPE

Breast cancer is the most common of all cancers affecting women in the developed countries. (1) More middle‐age women die of breast cancer than of any other single cause. (1) In the United Kingdom, more than 41 000 cases are diagnosed annually, and it is predicted that 1 in every 9 women will develop breast cancer at some point in life. (2) Early detection plays a significant role in the fatality of breast cancer. Technologies that aid in the early detection of cancers have therefore attracted much attention from the research community.

Mammography and ultrasound imaging are the standard technologies used in cancer screening. Mammography is accepted as the “gold standard” for breast imaging. It is widely used as the primary tool for cancer screening. However, in diagnostic workup, mammography and breast ultrasound are often used as complementary investigations. Mammography has been shown to cause high false‐positive rates in diagnosis, and the radiation dose to the breast is harmful. (3) Further, cost considerations have resulted in most countries choosing to use screen film mammography instead of a digitized version. However, the inability to change image contrast or brightness, problems in detecting subtle soft‐tissue lesions (dense glandular tissues), and difficulties with archiving have limited the application of screen film mammography.

Ultrasound studies have been shown to be good at picking up many of the cancers missed by mammography, especially in women who have dense breasts. In addition, ultrasound is noninvasive, portable, and versatile. Further, it does not use ionizing radiation, and it is relatively lower in cost. However, ultrasound images have a major disadvantage: poor quality because of multiplicative speckle noise that results in artifacts. Segmentation of lesions in ultrasound images is therefore a challenging task that remains an open problem despite many past research efforts.

In most existing breast screening approaches, the “initial lesion”—that is, a suspect region—is manually located by a trained radiologist in a pre‐processing stage by marking its topmost, leftmost, bottommost, and rightmost boundary limits with crosses. These crosses (and hence the initial lesion) are then manually encompassed within a rectangular region of interest (ROI) (4) by the radiologist and subsequently presented to a computer‐aided diagnosis (CAD) system for further analysis leading to the segmentation and classification of the tumor. Because the selection of extreme points (that is, the crosses) and the rectangular region both require human intervention, these steps are open to subjectivity and human error. As a result, a well‐trained and experienced examiner with knowledge of the normal echo anatomy of the breast and the changes caused by pathology is required for accurate breast cancer screening. A failure to include an entire lesion within the ROI or to outline ROIs containing lesions (among other errors) can severely undermine the performance of the CAD system. Further, it has also been shown that radiologists with varying training backgrounds and experiences often reach rather different results in the reading of sonograms. (5)

Our current research focus is to provide the radiologist with an automated tool that can effectively assist in the selection of the ROI and in improving the consistency of interpretation. However, it is worthwhile noting that the automatic detection of ROIs is not meant to replace the radiologist, but to provide a tool to reduce the radiologist's ROI labeling time (see Section II) and to warn of possible ROIs that might otherwise be missed because of the poor quality of the ultrasound image.

Evaluation of the performance of the proposed algorithm requires a suitable test image database, a suitable evaluation metric, and a design goal. Because of the practical difficulties in obtaining databases with ultrasound images of normal and nearly normal breasts, previous state‐of‐the art algorithms (6) , (7) of initial lesion detection have used U.S. image databases that consist solely of malignant and benign tumors. In our experiments, we used a U. S. image database of 360 malignant and nonmalignant images that additionally contains location information concerning the extreme points of tumors (marked with crosses by a number of expert radiologists). The metric used to evaluate the performance of the algorithm on the database is an “overlap” figure (6) defined as the ratio of the intersection and the union of the two lesion areas that were manually identified by the radiologists and by the computer‐based algorithm. Specifically, in work by Drukker et al. (6) and Yap et al., (7) the design goal has been to achieve an overlap value of in excess of 0.4, which can be represented as

overlap=XYXY0.4, (1)

where X is the lesion area extracted by the computer‐based algorithm, and Y is the lesion labeled manually by the radiologists. In other words, the assumption has been that, if an overlap of 0.4 or more results, then the computer‐based algorithm has been successful in accurately and automatically performing the otherwise manual task of initial lesion identification. A secondary metric, “accuracy,” can therefore be defined as the percentage of experiments obtaining an overlap value of beyond 0.4. In the present research context, we use “accuracy” as the objective metric to evaluate and compare the results of the performance of the proposed algorithm with that of state‐of‐the‐art methods. Further subjective results have been illustrated for visual comparison.

The present paper is divided into five sections for clarity of presentation. Section II discusses our research motivation and the existing solutions to the problem. Section III introduces the proposed approach and provides a detailed discussion of each stage. Section IV sets out the experimental results and discusses the results in detail. Finally, Section 5 concludes with further directions for improvement and research.

II. MOTIVATION AND RESEARCH BACKGROUND

Many ongoing ultrasound breast imaging research projects are focused on creating CAD tools that have high sensitivity, specificity, and consistency in lesion classification. Examples include Boone, (8) Gurney, (9) Boukerroui et al., (10) Chen et al., (5) and Sehgal et al. (11) Unfortunately, these systems are based on the assumption that the ROI will be pre‐selected by a radiologist and that the analysis will be performed only on the cropped ROI. This requirement improves accuracy in detecting lesion shape because the noisy, dark, poor‐quality surrounding areas can be excluded from consideration because of the manual selection of the specific ROI. It therefore follows that a fully automated CAD tool used in cancer screening will require a preprocessing stage that is capable of automatic ROI labeling. Our work is focused on providing an effective solution to that problem (see Fig. 1).

Figure 1.

Figure 1

Use of the proposed solution. US=ultrasound; ROI=region of interest.

Two existing state‐of‐the‐art techniques for automatic ROI labeling of existing breast images, used as benchmarks to compare the performance of the proposed technique, can be summarized as follows:

  • Drukker et al. (6) investigated the use of the radial gradient index (RGI) filtering technique to automatically detect lesions on breast ultrasound images. In ultrasound images, lesions are almost invariably darker than the background. Thus, in Drukker's work, the grayscale of the original ultrasound images is initially inverted. Subsequently, images are pre‐processed by a median filter to remove speckle noise. The resulting image is fed to a RGI filter. (6) In RGI filtering, the images are sub‐sampled by a factor of 4. A brightness threshold (see Subsection III.E) for the RGI‐filtered image is varied iteratively from 0.74 to 0.62 until either at least one lesion of interest is detected. The detected areas smaller than 5mm2 are discarded. Lesion candidates are segmented from the background by maximizing an average radial gradient (ARD) index for regions grown from the detected points. According to Drukker et al., maximizing the ARD index is more accurate than maximizing the RGI index. (6) At an overlap level of 0.4 with lesions outlined by a radiologist, 75% accuracy of lesion detection was reported for the test set of ultrasound images used.

  • Yap et al. (7) analyzed the use of statistical methods (for example, local mean) and values of fractal dimensions in initial lesion detection. The images were preprocessed using histogram equalization, (12) and then hybrid filtering (see Subsection III.C) and marker‐controlled watershed segmentation were applied. (A “watershed” is the ridge that divides areas drained by different river systems. A catchments basin is the geographic area draining into a river or reservoir. The watershed transform applies these ideas to grayscale image processing in a way that can be used to solve a variety of image segmentation problems.) Marker‐controlled watershed segmentation is an approach based on the concept of markers to control oversegmentation in watershed transform. (13) The minimum local mean and the minimum of the fractal dimensions (see Subsection III.D) of the identified segments were then used to identify the initial lesion. Subsequently, neighborhood segments are identified, and these are finally combined to form the ROI. The accuracy of the automated ROI labeling is measured by an overlap of 0.4 between its lesion outline and the lesions labeled by the radiologists. The accuracy of ROI detection when using local mean was reported to be 69.21%; fractal dimension was 54.21%.

The proposed approach detailed in the next section intends to further extend the accuracy of ROI detection by following an effective multistage algorithm.

III. PROPOSED APPROACH

Fig. 2 shows a modular block diagram of the proposed technique. It uniquely combines histogram equalization as a preprocessing stage, followed by hybrid filtering, multifractal analysis, thresholding segmentation, and a rule‐based approach in fully automated ROI labeling.

Figure 2.

Figure 2

Overview of the methodology.

The operation and functionality of the individual stages are described in detail in the subsections that follow. All algorithms were implemented in LTI‐Lib (14) on a Linux platform.

A. Ultrasound images

In general, ultrasound images are complex because of data composition, which can be described in terms of speckle information. Upon visual inspection, speckle noise consists of a relatively high‐level gray intensity, qualitatively ranging between hyperechoic (bright) and hypoechoic (dark) domains. (15)

Notably, any automatic system that is designed to detect abnormal lesions in ultrasound images should, in the end, be verified or compared with the judgment of a medical expert or radiologist. The test images used in our work were obtained from a professionally compiled compact disc (CD) of breast ultrasound images, (16) which consists of explanations and verifications from several qualified expert radiologists. We selected 360 images from the CD for our experiments. Of the 360 images, 20 were malignant, 76 were simple cysts, 76 were complex cysts, 58 were fibroadenomas, 38 were carcinomas, 18 were occult lesions, 15 were adenosis lesions, and 59 were a combination of other diagnoses. Each image had been manually processed by an expert radiologist, and the extreme points of the suspected lesions were already marked with crosses. Because the objective metric of the experiment (see Section I) required identification of the lesion boundary marked by a radiologist (rather than the extreme points or ROI), we obtained the services of an expert radiologist to mark these boundaries manually and to verify the extreme points given in the ultrasound CD. (16)

B. Preprocessing

As mentioned previously, the credibility of a high‐quality breast ultrasound examination depends on the scanner (that is, the quality of the original image) and the experience of the examiner. The preprocessing stage deals with the issue of guaranteeing the homogeneity of the original ultrasound images, which thus subsequently improves the chances of lesion ROI detection being more accurate. In the proposed approach, we use a histogram equalization strategy tested in earlier experiments (17) as a preprocessing stage to achieve the homogeneity guarantee.

Histogram equalization (12) is similar to contrast stretching, in that it attempts to increase the dynamic range of the pixel values in an image. However, unlike contrast stretching, no provision is made for interactivity, because applying a histogram equalization algorithm to an image with a fixed number of bins will always yield the same result. Let

pr(rj),j=1,2,,L (2)

denote the histogram associated with the intensity levels of a given image, and recall that the values in a normalized histogram are approximations to the probability of occurrence of each intensity level in the image. For discrete quantities, the equalization transformation becomes

sk=T(rk)=j=1kpr(rj)=j=1knjn (3)

for k=1,2,,L, where sk is the intensity value in the output (processed) image corresponding to value rk in the input image, n is the total number of pixels, and nj is the number of pixels in bin j.

C. Hybrid filtering

The function of the filtering stage is to remove noise, which is a major obstacle to accurate segmentation of the images. Median filtering is a popular approach used in removing speckle noise in ultrasound images. (6) , (18) , (19) However, Yap et al. (17) proved that part of the reason for the inaccuracy of the boundary detection in Drukker et al., (6) Joo et al., (18) and Kupinski et al. (19) was that, although the median filter managed to filter out the speckle noise, it also removed the important edge information—in particular, edges that belonged to the lesion.

Further, Gaussian blur (4) is a linear filtering technique that has been widely used to reduce the oversegmentation problem in ultrasound images. Gaussian blur is very effective in removing speckle noise, but it blurs and dislocates edges, (20) which may negatively affect subsequent lesion segmentation. Perona and Malik (21) proposed a nonlinear partial differential equation approach for smoothing images on a continuous domain. Anisotropic diffusion was shown to perform well for images corrupted by additive noise. However, in cases where images contain speckle noise, anisotropic diffusion enhances that noise instead of eliminating it. (22) On a more positive note, nonlinear diffusion filtering. (23) as compared with linear diffusion, has deservedly attracted much attention in the field of image processing for its ability to reduce noise while preserving (or even enhancing) important features of the image such as edges or discontinuities.

Within the context of proposed research, we make use of a hybrid filtering approach that combines the strength of nonlinear diffusion filtering to produce edge‐sensitive speckle reduction, with linear filtering (Gaussian blur) to smooth the edges and to eliminate oversegmentation. The result of hybrid filtering is visually compared in Fig. 3. Section IV sets out our experimental results and a detailed analysis to justify the use of hybrid filtering as compared with either nonlinear diffusion filtering or Gaussian blur alone.

Figure 3.

Figure 3

Comparison of filtering approaches: (a) original image; (b) Gaussian blur; (c) nonlinear diffusion filtering; (d) hybrid filtering.

Subsequent to hybrid filtering, we use multifractals (24) to further enhance the partially processed images. Section IV sets out experimental results to show that this stage allows for better segmentation of lesions than does the application of hybrid filtering only. Subsection D, next, provides a brief overview of multifractals and associated analysis methodology.

D. Multifractal dimensions

A fractal is generally “a rough or fragmented geometric shape that can be subdivided in parts, each of which is (at least approximately) a reduced‐size copy of the whole,” (25) a property called self‐similarity. In analyzing the fractal geometry of an image, an attempt is made to exploit the self‐similarity present. In fractal geometry, the term “fractal dimension” refers to a statistical quantity that indicates how completely a fractal appears to fill space, as one zooms down to finer and finer scales. (25) Multifractal analysis refers to the analysis of an image using multiple fractals—that is, not just one fractal, as in fractal analysis.

The generalized formulation for multifractal dimensions D of order q (24) can be represented as

Dq=1q1limε0log(xq(ε))log(ε)forqandq1limε0iμilogμilog(ε)forq=1, (4)

where ∊ is the linear size of the cells (in our case, because we use a 3×3‐pixel mask, ɛ=3), and q is the order for cell size ∊. Note that when q<0,Dq is sensitive to the parts where the measure is very dense. On the other hand, if q>0, information on the sparse region can be obtained. In theory, q is in the range –∞ to ∞, and Dg can have an infinite number of values. In practice, computing for all values of q is not possible. Hence, we empirically decided to use only four values of q in our experiments: 1,0, 1, and 2.

The partition function χq is defined as

xq(ε)=i=1N(ε)μiq(ε), (5)

where N(∊) is the total number of cells of size ∊, and μiθ(ɛ) is the measure that is defined on a given set. In this case, the measures are defined as the probability of the grayscale level in the images, where all the gray levels fall in the range of 0 – 1.

To investigate the effect of various q values on ultrasound images, we carried out further empirical experiments.

The graph in Fig. 4 shows that the multifractals with <q< yields four types of results. We can classify those results into q=0, q=1, 0<q<1, and (q<0 or q>1). Note that 0<q<1 is the inverted value of (q<0 or q>1). From the results illustrated in Fig. 5, it is seen that any value range 0<q<1 or (q<0 or q>1) will help improve the segmentation results. The value q=1 (that is, D1) is chosen in our work because, as compared with other values, it has the lowest associated computational complexity.

Figure 4.

Figure 4

Graph of multifractals dimension with various values of q.

Figure 5.

Figure 5

Result of multifractal analysis with (a) q<0 or q>1; (b) 0<q<1; (c) q=1; (d) threshold of (a); (e) threshold of (b); and (f) threshold of (c).

To our knowledge, hybrid filtering has not been used in previous research in breast ultrasound boundary detection or contrast enhancement, and hence its use is an additional novel aspect of our approach.

After application of the hybrid and multifractal filtering, the images are ready for lesion segmentation. Subsection E introduces the approach adopted.

E. Thresholding segmentation

In general, segmentation is a process used to divide an image into its constituent parts. Thresholding segmentation (26) is the most basic, the simplest, and the fastest algorithm in segmentation. A thresholding procedure attempts to determine an intensity value, called the “threshold,” that separates pixels into desirable classes. A parameter θ, called the “brightness threshold” is chosen and applied to the image a[m, n] as follows:

Ifa[m,n]3qthena[m.n]:=1(object)elsea[m.n]:=0(background) (6)

Within our present research context, we used a fixed threshold for segmentation—that is, a threshold chosen independently of the image data. If it is known that very high‐contrast images (where the objects are very dark and the background is homogeneous and very light) are being dealt with, then a constant threshold of 128 on a scale of 0 to 255 was experimentally found to be sufficiently accurate. That is, the number of falsely‐classified pixels is kept to a minimum. The sensitivity of the threshold selection on segmentation accuracy was further found to be low, justifying the use of a image‐independent fixed threshold.

F. Initial lesion detection

The thresholding segmentation of Subsection III.E often leads to the identification of multiple ROIs, of which generally only one or two would be of diagnostic importance (that is, belonging to abnormal lesions). Further, the location of the abnormal lesions requires the specification of both position and orientation (assuming two‐dimensional ultrasound images). We therefore propose the use of a rule‐based approach to identify these important ROIs.

The first criterion for the identification of lesions is the size of the segments. The suspect lesions are identified as the largest segments among the likely multiple segments that result from applying the single‐threshold segmentation. In addition, based on the additional guidance provided in the ultrasound image CD from which we obtained our test data, (15) we observed that 95% of tumors are located at the upper regions of the images. Hence, a reference point at (x, y), where

x=imageheight3,andy=imagewidth2 (7)

is chosen as the center of attention. The lesion that is located closest to that point and that satisfies the above mentioned size‐related criterion is selected as the final detected lesion. The size‐related criterion can be appropriately slightly relaxed if more than one lesion must be detected.

IV. RESULTS AND ANALYSIS

Fig. 6 compares the effects of using the various filtering techniques discussed in Section III with multifractal processing. When Gaussian filtering is used [see Fig. 6(b)], the smoothing effects introduced at the edges result in noisy regions that ideally should be disconnected remaining connected. This effect is a problem in the subsequent lesion detection stage, because the largest connected region may not now refer to the true lesion that should be detected. The use of nonlinear filtering results in oversegmentation and causes many problems, as illustrated by Fig. 6(c). However, with the use of hybrid filtering and multifractal processing, the single largest segmented area detected is identified as the lesion.

Figure 6.

Figure 6

Comparison of segmentation approaches: (a) original image, and use of (b) Gaussian plus multifractal filtering, (c) nonlinear plus multifractal filtering, (d) hybrid plus multifractal filtering.

Fig. 7 compares final lesion boundary detection accuracies for the use of hybrid filtering alone and for hybrid filtering followed by multifractal processing. The boundary detection accuracy in Fig. 7(e) is better than that in Fig. 7(c), justifying the positive contribution of multifractal processing to subsequent boundary detection. The boundary in Fig. 7(c) is closer to the boundary that a typical radiologist might intend to identify, because the boundary is smoother and excludes effects of noise and other artifacts more effectively.

Figure 7.

Figure 7

Justification of boundary enhancement with multifractal processing. (a) Original image; (b) hybrid filtered image; (c) boundary detected on hybrid filtered image; (d) multifractal processed image; (e) boundary detected on multifractal processed image.

Figs. 8, 9, and 10 illustrate the stage‐by‐stage operation of the algorithms from Drukker et al., (6) Yap et al., (7) and the present proposal. These three algorithms use different steps in identifying lesions. The intentions at each stage within the three algorithms are significantly different, and hence, the intermediate results are not comparable. A more comprehensive performance comparison based on the accuracy of the final detected lesion regions is given later in the present paper.

Figure 8.

Figure 8

Illustration of the operation of the intermediate stages of the Drukker et al. (6) algorithm. (a) Original image; (b) grayscale inverted image; (c) median filtered image; (d) radial gradient index (RGI) filtered image; (e) thresholded RGI image; (f) final detection.

Figure 9.

Figure 9

Illustration of the operation of the intermediate stages of the Yap et al. (7) algorithm (local mean). (a) Original image; (b) image after preprocessing (histogram equalization); (c) image after hybrid filtering; (d) watershed segmentation mapped onto the original image; (e) location of the initial lesion; (f) combination of the neighborhood segments with the initial lesion.

Figure 10.

Figure 10

Illustration of the operation of the intermediate stages of the proposed algorithm. (a) Original image; (b) image after pre‐processing (histogram equalization); (c) image after hybrid filtering; (d) image after multifractal processing; (e) image after thresholding segmentation; (f) labeling of region of interest.

Our detailed experiments, performed using the 360 test ultrasound images, revealed that the proposed method performs exceptionally well in identifying ROIs for most cyst lesions and malignant lesions, and for some fibroadenoma lesions. Because of the high degree of similarity in texture between normal and fibroadenoma regions, accurate identification of such regions is always a challenge. Fig. 11 provides a visual comparison of the results of using the proposed algorithms on various types of abnormalities. Fig. 12 illustrates two examples of fibroadenoma ROIs that are not detected accurately.

Figure 11.

Figure 11

Results of automated region of interest lesion labeling using the proposed method. (a) Malignant tumour; (b) simple cyst; (c) fibroadenoma; (d) complex cyst.

Figure 12.

Figure 12

Examples of unsuccessful lesion identification for two cases of fibroadenoma.

Based on diagnoses by radiologists as provided on the test ultrasound breast imaging CD used in our experiments, (16) Table 1 summarizes the accuracy figures obtained for each type of abnormality for the three benchmark algorithms and the proposed algorithm. Fig. 13 graphically presents ROI detection accuracy for each of the four algorithms and each type of abnormality. The results clearly show the improvements obtainable with the proposed improved approach to ROI lesion detection accuracy. The detection accuracy for fibroadenoma‐type lesions has generally been the lowest for all methods, but in this category, the proposed algorithm shows a 15% accuracy improvement as compared with the methods of Drukker et al. (6) and Yap et al. (7) Further, when using the proposed algorithm, 90% accuracy in detecting malignant‐type lesions is indicated.

Table 1.

Summary of region of interest detection accuracy for each type of abnormality under four different algorithms

    Diagnosis Total images (n) Drukker et al. (6) Accuracy (%)
Yap et al. (local mean) (7) Yap et al. (fractal) (7) Proposed method
Malignant 20 80.00 65.00 45.00 90.00
Simple cysts 76 67.11 60.53 55.26 86.84
Complex cysts 76 72.37 71.05 59.21 89.47
Fibroadenoma 58 62.07 63.79 46.55 77.59
Carcinoma 38 57.89 78.95 52.63 78.95
Occult lesions 18 94.44 88.89 66.67 88.89
Adenosis 15 80.00 73.33 46.67 93.33
Others 59 59.32 64.41 54.24 89.83
Total 360 67.78 68.05 53.89 86.11

Figure 13.

Figure 13

Graphical presentation of the region of interest (ROI) detection accuracy for each of the four algorithms and each type of abnormality.

Fig. 14 visually compares the performance of the proposed approach with that of the three benchmark algorithms. This comparison clearly illustrates the improved accuracy of lesion identification demonstrated by the proposed approach. The three benchmark algorithms can be seen to be more likely to identify non‐lesion regions—or only parts of lesions—as ROIs.

Figure 14.

Figure 14

Visual performance comparison with the benchmark algorithms. (a,f) Original images; (b,g) results of the Drukker et al. (6) algorithm; (c,h) results of the Yap et al. (7) algorithm (local mean); (d,i) results of the Yap et al. (7) algorithm (fractal dimension); (e,j) results of the proposed method.

V. SUMMARY AND CONCLUSIONS

We propose a method that is able to fully automate ultrasound CAD. We have successfully proved that the proposed algorithm achieves an improvement as compared with benchmark algorithms. The proposed method is able to very accurately label most lesions, with its best performance being the identification of malignant lesions (90%) and its worst being the identification of fibroadenomas (77.59%). We are currently considering the use of shape information and frequency domain analysis, among other techniques, to further improve the performance of the approach presented here.

A complete computer‐aided ultrasound diagnostic system (ultrasound CAD) must be able to classify the diagnosis of each ROI. In future, we will investigate the use of classification techniques such as neural networks and support vector machines, among others, to form a fully automated breast cancer detection system.

REFERENCES

  • 1. Lawson A, Lawson J. Breast cancer—can you prevent it? Sydney (Australia): McGraw–Hill; 1999. 224 p. [Google Scholar]
  • 2. News BBC. Breast cancer. BBC News Channel [Electronic version]. 30 January 2004. Retrieved 6 August 2008 from http://news.bbc.co.uk/go/pr/fr/‐/1/hi/health/medical_notes/3244315.stm.
  • 3. Freedman DA, Petiti DB, Robin JM. On the efficacy of screening for breast cancer. Int J Epidemiol. 2004;33(1):43–55. [DOI] [PubMed] [Google Scholar]
  • 4. Chen DR, Chang RF, Wu WJ, Moon WK, Wu WL. 3‐D Breast ultrasound segmentation using active contour model. Ultrasound Med Biol. 2003;29(7):1017–1026. [DOI] [PubMed] [Google Scholar]
  • 5. Chen D, Chang RF, Huang YL. Breast cancer diagnosis using self‐organizing map for sonography. Ultrasound Med Biol. 2000;26(3):405–411. [DOI] [PubMed] [Google Scholar]
  • 6. Drukker K, Giger ML, Horsch K, Kupinski MA, Vyborny CJ, Mendelson EB. Computerized lesion detection on breast ultrasound. Med Phys. 2002;29(7):1438–1446. [DOI] [PubMed] [Google Scholar]
  • 7. Yap MH, Ewe HT. Region of interest (ROI) detection in ultrasound breast images. TS11–2. In: Proceedings of MMU International Symposium on Information and Communications Technologies (M2USIC). Cyberjaya, Malaysia: Multimedia University; 2005:5–8. [Google Scholar]
  • 8. Boone JM. Neural networks at the crossroads. Radiology. 1993;189(2):357–359. [DOI] [PubMed] [Google Scholar]
  • 9. Gurney JW. Neural networks at the crossroads: caution ahead. Radiology. 1994;193(1):27–28. [DOI] [PubMed] [Google Scholar]
  • 10. Boukerroui D, Basset O, Guérin N, Baskurt A. Multiresolution texture based adaptive clustering algorithm for breast lesion segmentation. Eur J Ultrasound. 1998;8(2):135–144. [DOI] [PubMed] [Google Scholar]
  • 11. Sehgal CM, Cary TW, Kangas SA. Computer‐based margin analysis of breast sonography for differentiating malignant and benign masses. J Ultrasound Med. 2004;23(9):1201–1209. [DOI] [PubMed] [Google Scholar]
  • 12. Wintz P, Gonzalez RC. Digital image processing. 2nd ed. Boston (MA): Addison–Wesley; 1987: 146–152. [Google Scholar]
  • 13. Gonzalez RC, Woods RE, Eddins SL. Digital image processing using MATLAB. Upper Saddle River (NJ): Prentice‐Hall; 2003. 609 p. [Google Scholar]
  • 14. SourceForge.net . LTI‐LIB. Introduction [Electronic resource]. 24 November 2005. Retrieved 6 August 2008 from http://Ltilib.sourceforge.net/doc/homepage/index.shtml.
  • 15. Friedland N, Adam D. Automatic ventricular cavity boundary detection from sequential ultrasound images using simulated annealing. IEEE Trans Med Imaging. 1989;8(4):344–353. [DOI] [PubMed] [Google Scholar]
  • 16. Prapavesis ST, Fornage BD, Weismann CF, et al. Breast ultrasound and US‐guided interventional techniques [Multimedia teaching file]. Thessaloniki (Greece): authors; 2001. [Google Scholar]
  • 17. Yap MH, Edirisinghe EA, Bez HE. Object boundary detection in ultrasound images. In: Ley M. Proceedings of Third Canadian Robotics and Machine Vision Conference (CRV 2006). Quebec (QC): IEEE Computer Society; 2006: 53 Abstract retrieved 7 November 2006 from http://doi.ieeecomputersociety.org/10.1109/CRV.2006.51. [Google Scholar]
  • 18. Joo S, Yang YS, Moon WK, Kim HC. Computer‐aided diagnosis of solid breast nodules: use of an artificial neural network based on multiple sonographic features. IEEE Trans Med Imaging. 2004;23(10):1292–1300. [DOI] [PubMed] [Google Scholar]
  • 19. Kupinski M, Giger ML, Lu P, Huo ZM. Computerized detection of mammographic lesions: performance of artificial neural network with enhanced feature extraction. Proc SPIE. 1995;2434:598–605. [Google Scholar]
  • 20. Mrazek P. Nonlinear diffusion for image filtering and monotonicity enhancement [PhD thesis]. Prague (Czech Republic): Czech Technical University; 2001. 110 p. Document retrieved 6 August 2008 from ftp://cmp.felk.cvut.cz/pub/cmp/articles/mrazek/Mrazek‐phd01.pdf. [Google Scholar]
  • 21. Perona P, Malik J. Scale‐space and edge detection using anisotropic diffusion. IEEE Trans Pattern Anal Mach Intell. 1990;12(7):629–639. [Google Scholar]
  • 22. Yu Y, Acton ST. Speckle reducing anisotropic diffusion. IEEE Trans Image Process. 2002;11(11):1260–1270. [DOI] [PubMed] [Google Scholar]
  • 23. Weickert J. Nonlinear diffusion filtering. In: Jähne B, Haussecker H, Geissler P, editors. Handbook on computer vision and applications. Vol. 2 San Diego (CA): Academic Press; 1999: 423–450. [Google Scholar]
  • 24. Yeo TS, Du G. A multifractal approach for auto‐segmentation of SAR images. In: Proceedings of the IEEE 2001 International Geoscience and Remote Sensing Symposium; Sydney, Australia; 9 – 13 July 2001. Los Alamitos (CA): IEEE; 2001: 5:2301–2303. [Google Scholar]
  • 25. Mandelbrot BB. The fractal geometry of nature. Rev. ed. New York (NY): W.H. Freeman and Company; 1982. 468 p. [Google Scholar]
  • 26. Sahoo PK, Soltani S, Wong AKC, Chen YC. A survey of thresholding techniques. Comput Vis Graph Image Process. 1988;41(2):233–260. [Google Scholar]

Articles from Journal of Applied Clinical Medical Physics are provided here courtesy of Wiley

RESOURCES