Automated detection of breast masses on digital mammograms using adaptive density-weighted contrast enhancement filtering Nicholas Petrick, Heang-Ping Chan, Berkman Sahiner, Datong Wei, Mark A. Helvie, Mitchell M. Goodsitt and Dorit D Adler The University of Michigan Department of Radiology Ann Arbor, MI 48109-0030 ABSTRACT This paper presents a novel approach for segmentation of suspicious mass regions in digitized mammograms using a new adaptive Density-Weighted Contrast Enhancement (DWCE) filter in conjunction with LaplacianGaussian (LG) edge detection. The new algorithm processes a mammogram in two stages. In the first stage the entire mammogram is filtered globally using a DWCE adaptive filter which enhances the local contrast of the image based on its local mean pixel values. The enhanced image is then segmented with an LG edge detector into isolated objects. In the second stage of processing, the DWCE adaptive filter and the edge detector are applied locally to each of the segmented object regions detected in the first stage. The number of objects is then reduced based on morphological features. ROIs are selected from the remaining object set based on the centroid locations of the individual objects. The selected ROTs are then input to either a linear discriminant analysis ( LDA) classifier or a convolution neural network (CNN) to further differentiate true-positives and false-positives. In this study ROIs obtained from a set of 84 images were used to train the LDA and CNN classifiers. The DWCE algorithm was then used to extract ROTs from a set of 84 test images. The trained LDA and CNN classifiers were subsequently applied to the extracted ROTs, and the dependence of the detection system's accuracy on the feature extraction and classification techniques was analyzed. Keywords: Computer-Aided Diagnosis, Density-Weighted Contrast Enhancement, Adaptive Filter, Mammography, Edge Detection, Classification 1. DENSITY-WEIGHTED CONTRAST ENHANCEMENT SEGMENTATION We have developed a new algorithm using DWCE filtering with Laplacian-Gaussian (LG) edge detection for segmentation of low contrast objects in digital mammograms. The DWCE algorithm is used to enhance objects in the original image so that a simple edge detector can define the object boundaries. Once the object borders are known, morphological features are extracted from each object and used by a classification algorithm to differentiate mass and non-mass regions within the image. 590 ISPIE Vol. 2434 0-8194-1782-3/951$6.00 Downloaded From: http://proceedings.spiedigitallibrary.org/ on 02/17/2016 Terms of Use: http://spiedigitallibrary.org/ss/TermsOfUse.aspx Figure 1: A typical mammogram from our image database (a) and the corresponding DWCE filtered image (b). 1.1 DWCE Preprocessing Filter Edge detection applied to the original digitized mammograms has not proven effective in detecting breast masses within the image because of the low signal-to-noise ratio of the edges and the presence of complicated structured background. Fig. 1(a) shows a typical mammogram from our image database. It contains a single breast mass indicated by the arrow. This mammogram also contains dense fibroglandular tissue in the breast parenchyma. Although the mass is relatively obvious, the partially overlapping tissue makes the detection difficult. In order to detect masses of varying shapes and intensities, we propose using an aggressive adaptive filtering technique to suppress the background structures and enhance any potential signals. Fig. 1(b) shows the processed mammogram of Fig. 1(a). The background in this image is clearly reduced which allows a simple edge detector to locate the different objects within the image. The block diagram for the DWCE preprocessing filter is depicted in Fig. 2. It is an expansion of the local contrast and mean adaptive filter of Peli and Lim' designed for enhancing images degraded by cloud cover. The original image, F(x, y), is initially passed through a map rescaler. The map rescaler first determines the breast boundary (referred to as the breast map). The pixel values in the breast region are then rescaled based on the maximum and minimum values within the breast map, producing the normalized image, FN(X, y). FN(X, y) is then split into a density and a contrast image, FD(X, y) and F(x, y), respectively. The density image is produced l)y filtering FN(X, y) using a zero mean Gaussian with standard deviation, 0D. FD(X, y) thus directly correlates to a weighted average of the local optical density of the original film. The contrast image, y), is also created by filtering FN(X, y), but the low pass filter is replaced with a band pass or high-pass filter. Each pixel in the density image is then used to define a multiplication factor which modifies the corresponding pixel in the contrast SPIEVo!. 2434/591 Downloaded From: http://proceedings.spiedigitallibrary.org/ on 02/17/2016 Terms of Use: http://spiedigitallibrary.org/ss/TermsOfUse.aspx Figure 2: The block diagram of the DWCE preprocessing filter used for image enhancement. image: FKC(X,Y) = KM(FD(X,y)) x Fc(x,y). (1) is the essence of the DWCE algorithm. It allows the local density value of each pixel to weight its local contrast. Fig. 3(a)-(c) show the density, contrast and weighted contrast images, respectively, for the digitized mammogram of Fig. 1(a). This The output of the DWCE filter is a nonlinear rescaled version of the weighted contrast image, FE(X, y) = KNL(FKC(X, y)) x FKC(X, y). (2) nonlinear rescaling is used to help equalize the higher intensities and to de-emphasize very low ones in FKC(X, y). This rescaling improves the definition of the object borders by eliminating many of the low intensity The edges that cause region merging and reduces the effect of extremely large intensities on the edge detection. Fig. 1(b) shows the final enhanced image after nonlinear rescaling produced by the complete DWCE filter. 1.2 Edge Detection In this study, object edges were detected from the DWCE prefiltered images using a Laplacian-Gaussian (LG) edge detector. For a given image, I(x, y), the LG edge detector defines edges as simply the zero crossing locations of: 2G(x,y)*I(x,y), (3) where G(x, y) is a two dimensional Gaussian smoothing function.2 The degree of smoothing is controlled by a single parameter, 0E, the standard deviation of the smoothing function. This edge detector tends to produce closed regions which facilitates detection of isolated objects within the image. Once the object edges have been defined in the image, each enclosed object is filled to remove any holes that may have formed. Each region produced by the filling is defined by its edge pixels, thus forming a set of detected objects. This set of objects defines all the detected structures within the original breast image. 1.3 Morphological Object Classification Since the DWCE prefilter enhances both breast masses and normal tissue, a large number of detected objects are usually found. In order to reduce the number of objects to a manageable size, morphological features were 592 / SPIE Vol. 2434 Downloaded From: http://proceedings.spiedigitallibrary.org/ on 02/17/2016 Terms of Use: http://spiedigitallibrary.org/ss/TermsOfUse.aspx Figure 3: The density (a), contrast (b) and weighted contrast (c) images produced by the DWCE filter applied to the mammogram of Fig. 1(a) along with the corresponding LG edge map (d). SPIE Vol. 2434 / 593 Downloaded From: http://proceedings.spiedigitallibrary.org/ on 02/17/2016 Terms of Use: http://spiedigitallibrary.org/ss/TermsOfUse.aspx extracted from each object and used in conjunction with a threshold classifier to screen breast masses from normal tissue.3 The morphological features used in this classification include the number of edge pixels, area, shape, contrast and a set of normalized radial length (NRL) features. Three features, circularity, rectangularity and the ratio of the number of edge pixels (P) to the total object area (A) were used to characterize the shape of an object. Circularity is defined as the ratio of the overlapping area between an equivalent circle of radius, r= with the object to the object's total area. Rectangularity is defined as the ratio of the object area to the area of the enclosing bounding box. The radial length is defined as the Euclidean distance from the object centroid to each of the object's edge pixels and is normalized by dividing by the maximum radial length.4 The NRL features selected were the NRL mean, standard deviation, entropy, area ratio and zero crossing count as defined by Kilday ei. al.4 Using these 11 morphological features object classification was performed. The morphological classification was not meant to be the final classification of the detected objects. Instead, it was used to reduce the number of objects so that further detailed analysis could be performed on each of the detected regions. To perform this analysis ROIs of fixed sizes were extracted from the original mammograms based on the centroid locations of the objects. 2. ROl CLASSIFICATION Two different ROT classification algorithms and their combination were evaluated in their ability to reduce the number of false positive detections. The two methods were linear discriminant analysis (LDA) applied to multi-resolution texture features and a convolution neural network (CNN) applied directly to a set of texture feature images and to the original spatial domain ROT images. The multi-resolution texture features were calculated from the spatial gray level dependence (SGLD) matrices of both the original image and the corresponding low-pass wavelet images at different scales. They include eight features: correlation, energy, entropy, inertia, inverse difference moment, sum average, sum entropy and difference entropy. The definitions for each texture feature can be found in the literature.5 LDA was then used to form linear combination of these multi-resolution features by maximizing the group mean separation of the mass and the non-mass ROIs. If the features follow a multivariate normal distribution with an identical covariance matrix for both groups, then the LDA yields the optimal classification. This linear combination forms a single discriminant score for each ROT.6 Therefore, by using a single threshold the ROIs can be separated into mass and non-mass regions. A detailed description of this ROT classification method can be found in the literature.7 A CNN is a back-propagation neural network with two-dimensional weight kernels that operate on images. The input images to the CNN are derived from the original ROTs. In this application a spatial image obtained by smoothing and sub-sampling the original ROIs and texture images obtained by applying texture feature methods to small sub-regions inside the ROTs were used. The small sub-regions were defined so that the texture images had the same size as the smoothed and sub-sampled ROIs. The output of the CNN was again a single discriminant score for each ROT. The CNN classification method used in this paper was described by Sahiner et. al.8 The final classification scheme evaluated in this paper is the combination of the LDA and CNN score using a simple two input and one output back-propagation neural network (BPN). 594 / SPIE Vol. 2434 Downloaded From: http://proceedings.spiedigitallibrary.org/ on 02/17/2016 Terms of Use: http://spiedigitallibrary.org/ss/TermsOfUse.aspx Stage 1 L Filter Dection Figure 4: The block diagram of the complete DWCE segmentation method used for breast mass detection. 3. METHODS 3.1 Database The mammograms used in this study were randomly selected from the files of patients who had undergone biopsy in the Department of Radiology at the University of Michigan. The mammograms were acquired using a Kodak MinR/MRE screen/film system with extended cycle processing. The system has a 0.3mm focal spot, a molybdenum anode, 0.03mm thick molybdenum filter and a 5:1 reciprocating grid. Our selection criterion was simply that a biopsy-proven mass could be seen on the mammogram. Our data set in this preliminary study was composed of 168 mammograms. The size of the masses ranged from 5mm to 26mm with a mean size of 12.2mm, and included 85 malignant and 83 benign masses. The mammograms were digitized with a LUMISYS DIS-1000 laser film scanner with a pixel size of 100pm x 100pm and 4096 gray levels. The light transmitted through the mammographic film was logarithmically amplified before digitization, making the gray levels linearly proportional to the optical density in the range of 0.1 to 2.8 optical density units (O.D.). The O.D. range of the scanner was 0 to 3.5. The digitized images are approximately 2000 x 2000 pixels in size. Before the DWCE segmentation was applied the images were smoothed using local averaging and subsampled by a factor of 8. This resulted in images of approximately 256 x 256 pixels for processing. 3.2 DWCE Implementation Fig. 4 shows the block diagram for the DWCE implementation used to detect breast masses in the digitized mammograms. It was performed using two DWCE stages. The first stage applied the DWCE prefiltering, edge detection and threshold classification to each subsampled mammogram to identify potential mass objects. For each potential mass object, an ROl was extracted from the corresponding subsampled mammogram using the object's bounding box to define the region. Each of the object ROIs were then passed through a second DWCE stage and the final set of detected objects was defined. The parameters used in the DWCE prefiltering, edge detection and object reduction steps were identical to those of their first stage counterparts. However, object SPIE Vol. 2434 / 595 Downloaded From: http://proceedings.spiedigitallibrary.org/ on 02/17/2016 Terms of Use: http://spiedigitallibrary.org/ss/TermsOfUse.aspx extraction was more effective and accurate because it was performed in local regions. ROIs were then extracted from the full resolution mammograms centered on the centroid location of each of the remaining objects. These ROIs were fixed at 256 x 256 pixels. 3.3 ROl Classification The database was then randomly divided into 84 training and 84 test images. Multi-resolution texture features were extracted from each of the training ROTs and used to train the LDA classifier. Similarly, texture feature images along with the smoothed and subsampled ROIs were extracted and used as input to the CNN classifier. The discriminant scores obtained from both classifiers were then used to train the combined BPN network. After the classifiers were trained the corresponding inputs were extracted from the test ROTs and used to evaluate the performance of the three classifiers. 4. RESULTS Using the 84 test images, the DWCE algorithm identified 83 of the 84 true mass ROIs. The LDA and CNN classifiers were then applied to the extracted ROIs reducing the FPs to 2.5 per image and 4.6 per image, respectively, at a TP rate of 80%. When the LDA and CNN were combined the FPs were reduced to 2.4 per image at a TP rate of 80%. 5. CONCLUSIONS The results demonstrated the feasibility of using the DWCE adaptive filter with LG edge detection to automatically extract ROTs from a mammogram. This ROT extraction method in combination with a classification algorithm is a promising new approach to computer-aided detection of masses on mammograms. 6. ACKNOWLEDGMENTS This work is supported by a USPHS Grant CA 48129 and U.S. Army Grant DAMD 17-93-J-007 (through sub-grant GU RX 4300-803UM from Georgetown University). The content of this publication does not necessarily reflect the position of Georgetown University or the government, and no official endorsement of any equipment or product of any company mentioned in the publication should be inferred. 7. REFERENCES 1. T. Peli and J. S. Lim, "Adaptive filtering for image enhancement," Oplical Engineering, vol. 21, no. 1, pp. 108— 112, 1982. 2. D. Marr and E. Hildreth, "Theory of edge detection," Proceedings of the Royal Soczey of London Series B Biological Sciences, vol. 207, pp. 187—217, 1980. 3. N. Petrick, 11.-P. Chan, B. Sahiner, and D. Wei, "An adaptive density weighted contrast enhancement filter 596 / SPIE Vol. 2434 Downloaded From: http://proceedings.spiedigitallibrary.org/ on 02/17/2016 Terms of Use: http://spiedigitallibrary.org/ss/TermsOfUse.aspx for mammographic breast mass detection." Submitted to IEEE Trans. Med. 1mg., Oct. 1994. 4. J. Kilday, F. Palmieri, and M. D. Fox, "Classifying mammographic lesions using computerized image analysis," IEEE Transaciions on Medical Imaging, vol. 12, pp. 664-669, Dec. 1993. 5. A. Petrosian, H.-P. Chan, M. A. Helvie, M. M. Goodsitt, and D. D. Adler, "Computer-aided diagnosis in mammography: Classification of mass and normal tissue by texture analysis," Physics of Medicine and Biology, vol. 39, pp. 2273—2288, 1994. 6. P. Lachenbruch, Discriminani Analysis. New York: Hafner Press, 1975. 7. D. Wei, H.-P. Chan, M. A. Helvie, B. Sahiner, N. Petrick, D. D. Adler, and M. M. Goodshitt, "Classification of mass and normal breast tissue on digital mammograms: Multiresolution texture analysis." Submitted to Medical Physics, Aug. 1994. 8. B. Sahiner, H.-P. Chan, N. Petrick, D. Wei, M. A. Helvie, , D. D. Adler, and M. M. Goodshitt, "Classification of mass and normal breast tissue on digital mammograms: A convolution neural network classifier with spatial domain and texture images." Submitted to IEEE Trans. Med. 1mg., Dec. 1994. SPIE Vol. 2434 / 597 Downloaded From: http://proceedings.spiedigitallibrary.org/ on 02/17/2016 Terms of Use: http://spiedigitallibrary.org/ss/TermsOfUse.aspx
© Copyright 2026 Paperzz