Ear Recognition: A Complete System Ayman Abazaa,b and MaryAnn F. Harrisona a West Virginia High Tech Foundation, 1000 Technology Drive, Fairmont, USA; b Cairo University, Cairo, Egypt ABSTRACT Ear Recognition has recently received significant attention in the literature. Even though current ear recognition systems have reached a certain level of maturity, their success is still limited. This paper presents an efficient complete ear-based biometric system that can process five frames/sec; Hence it can be used for surveillance applications. The ear detection is achieved using Haar features arranged in a cascaded Adaboost classifier. The feature extraction is based on dividing the ear image into several blocks from which Local Binary Pattern feature distributions are extracted. These feature distributions are then fused at the feature level to represent the original ear texture in the classification stage. The contribution of this paper is three fold: (i) Applying a new technique for ear feature extraction, and studying various optimization parameters for that technique; (ii) Presenting a practical ear recognition system and a detailed analysis about error propagation in that system; (iii) Studying the occlusion effect of several ear parts. Detailed experiments show that the proposed ear recognition system achieved better performance (94.34%) compared to other shape-based systems as Scaleinvariant feature transform (67.92%). The proposed approach can also handle efficiently hair occlusion. Experimental results show that the proposed system can achieve about (78%) rank-1 identification, even in presence of 60% occlusion. Keywords: Ear recognition, haar cascade AdaBoost Classifier, Local Binary Pattern, and Hair occlusion 1. INTRODUCTION Ear trait has recently received significant attention as a powerful biometric because it avoids some of the problems inherent in facial recognition. Additionally, ear-based recognition systems can efficiently extend the compatibilities of face recognition systems. For face recognition at a high roll angle, toward the side view, the face recognition performance is very low; while the ear recognition, at that angle, is generally yielding high performance. Ear structure is a promising characteristic because the human ear has many variation among individuals,1 such as the various curves and geometric measurements of the ear. The first documented use of the ear for identification was put forth by the French Criminologist Alphonse Bertillon2 in 1890. Iannarelli3 used one of the first ear recognition systems in the literature in 1949. Burge and Burger4 present one of the most widely cited methods for ear biometrics. To the best of our knowledge, they never present a recognition experiment. Even though the concept of using the ear as a biometric has been around since the 1890’s there are currently no commercially available systems to automatically verify the identity of individuals using ear information. Generally, ear biometric systems consist of three main stages (as shown in fig. 1): • Ear Detection: Segmenting the ear region from the image. • Feature Extraction: Representing the ear structure by a feature vector. • Classification: Matching the probe and the gallery feature vectors to verify the subject claimed identity or to search a database in order to identify the admitted person. In this paper, we propose the use of the Block-based Local Binary Pattern (LBP) to generate features for ear recognition. The LBP operator is one of the best performing texture descriptors and it has been used in various applications, for example Further author information: (Send correspondence to Ayman Abaza) A.A.: E-mail: [email protected], Telephone: +1 304 333 6452. A.A. is also affiliated with Cairo University. M.H.: E-mail: [email protected], Telephone: +1 304 333 6432 1 Figure 1. Ear recognition system face5 and periocular recognition.6 It has proven to be highly discriminative and invariant to monotonic gray level changes.5 The idea of using LBP for ear description is motivated by the fact that ears can be seen as a composition of micro-patterns which are well described by LBP operator. We present detailed experiments using USTB database I7 and UND database, collection E.8 These experiments evaluate: • Performance optimization by tuning various LBP parameters such as the uniform pattern, number of neighbor pixels, radius of neighbor pixels, division of the input image into blocks, and feature selection techniques to assign weights for the blocks. • The LBP performance compared to bench mark techniques such as the principal component analysis (PCA), and scale-invariant feature transform (SIFT). • The error accumulates from automation of various components of the ear system. • The system performance in case of occlusion. This article is organized as follows: Section 2 highlights some related work on 2D ear recognition, and reviews the previous assessment of performance in case of hair occlusion. Section 3 and Section 4 give brief overviews of the Haarbased ear detection and the LBP technique respectively. Experimental results are presented in Section 5.3. Finally, Section 6 presents conclusions and sketches our future plans. 2. RELATED WORK Ear biometrics refer to the automatic measurement of distinctive ear features to identify or confirm the identity of the owner. In this section, we present a review of previous work on 2D ear detection and feature extraction, followed by ear occlusion. 2.1 Ear Detection Burge and Burger4 presented one of the most cited ear biometrics method in the literature. They located the ear using deformable contours on a Gaussian pyramid representation of the side profile image gradient. Ear edges (contours) are then computed. Ansari and Gupta9 used a similar approach for localization of the ear from an arbitrary 2D side face image based on the ears outer helix curves. Abdel-Mottaleb and Zhou,10 and Yuizono et al.11 attempted ear detection using a model-based (template matching) technique. Later, Prakash et al.12 modified the template-based technique by adding skin color. The technique first separates skin regions from non-skin regions and then searches for the ear within skin regions using the template matching approach. Finally, the ear region is validated using a moment-based shape descriptor. Hajsaid et al.13 addressed the problem of fully automated ear segmentation, using morphological operators. Additionally, they used low computational cost, appearance based features and a learning based Bayesian classifier to determine whether the output segment is a proper or improper ear segment. 2 2.2 Feature Extraction Chang et al.14 used principal component analysis (PCA), and introduced the concept of Eigen-Ear. They reported a performance of 72.7% for ear in an identification experiment, compared to 90.9% for the multi-modal system. This technique has been widely used in the literature as a base reference, and that is why we used for comparison. Dewi and Yahagi15 used ear scale-invariant feature transform (SIFT).16 They classified the owner of an ear by calculating the number of key points matches and the average distance of the closest square distance. Kisku et al.17 used SIFT as feature descriptor for structural representation of ear images. They formed an ear skin color model using a Gaussian mixture model (GMM) and clustered the ear color pattern using vector quantization. After segmentation of ear images in specified color slice regions, they extracted SIFT key-points. They fused the extracted SIFT key-points from all color slice regions. Feng and Mu18 combined wavelet transform and local binary patterns (LBP). They used non-uniform LBP8,1 operator, and evaluated the performance of various similarity measures and two matchers (K Nearest Neighbor, and two-class Support Vector Machine). They used 70 subjects from the USTB database III,7 and 10 images per subject at various poses. They reported 96.86% cross validation recognition rate using chi-square distance. Wang et al.19 also used wavelet transforms and uniform local binary patterns for ear recognition. They decomposed ear images by a Haar wavelet transform, and then applied Uniform LBP simultaneously with block-based and multi-resolution methods to describe the texture features. They used 79 subjects from the USTB database III,7 and 10 images per subject at various poses. They reported the best recognition rate by combining uniform LBP of decomposed ear images with multi-resolution and block-based. The recognition rate was as high as 100% for 5o angle, and deteriorated to 42.41% for 45o angle. Feng and Mu18 and Wang et al.19 mentioned that LBP of the whole image did not yield good performance and that is why they transformed the image to the wavelet domain first. Based on experimental studies, shown in the experimental section, it was proven that dividing the image into blocks can yield the same performance as the best wavelet transforms. 2.3 Hair Occlusion Yuan et al.20 proposed an Improved Non-Negative Matrix Factorization with sparseness constraints (INMFSC) by imposing an additional constraint on the objective function of NMFSC. They showed by experiments that their enhanced technique yielded better performance even with partially occluded images. Later Yuan et al.21 separated the normalized ear image into 28 sub-windows. Then, they used Neighborhood Preserving Embedding for feature extraction on each sub-window, and selected the most discriminative sub-windows according to the recognition rate. Finally, they applied weighted majority voting for fusion at the decision level. They tested recognition with partially occluded images from 33% top, 33% middle, 33% bottom, 50% left, and 50% right of the ear respectively. Kocaman et al.22 applied principal component analysis (PCA), fisher linear discriminant analysis (FLDA), discriminative common vector analysis (DCVA), and locality preserving projections (LPP). The error and hit rates of the four algorithms were calculated by random sub-sampling and k-fold cross validation for various occlusion scenarios. ArbabZavar et al.23 used Scale Invariant Feature Transform (SIFT) to detect the features within the ear images. They presented a comparison with PCA to show the advantage derived by the use of the model in successful occlusion handling. Later Bustard and Nixon24 evaluated the SIFT technique using various occlusion ratios and presented rank-1 of 92%, and 74% for 20%, and 30% occlusion from ear top part respectively. 3. EAR DETECTION USING CASCADED ADABOOST In order to detect an ear in a given image, the image is scanned using rectangular features arranged in a cascaded AdaBoost system.25 These rectangular features encode ad-hoc domain knowledge, and work faster than pixel-based.26 Each rectangle feature (f) represents the main component of the weak classifier (h(x, f, p, φ)), where (φ) is a threshold, and (p) is polarity indicating the direction of the inequality: ( 1 if pf (x) < pφ h(x, f, p, φ) = (1) 0 otherwise 3 System block diagram Figure 2. Schematic diagram of the ear detection system: An input image (1) is scaled multiple times, and the set of all possible 24x16 pixel sub-images are extracted from each scale, including overlapping sub-images (2). Each of the sub-images is put through a cascaded Adaboost algorithm (3), after which a decision about the presence and location of any ears is made (4). The learner is called a weak classifier as it has very low performance, which is still better than guessing. The AdaBoost learning algorithm is used to enhance the classification performance. Each ensemble classifier consists of T weak classifiers as follows: ( PT Continue t=1 αt ht (x) > θ H(x) = (2) Reject otherwise where θ is the threshold of the strong classifier. These strong classifiers are arranged in cascade in order to form the detection system, as shown in fig. 2. The input image is divided into overlapped sub-images, and each region is evaluated using the cascaded classifier. Then, the image is scaled down by a scaling factor (s) and the above mentioned process is repeated. Finally, all the detected regions, at various levels of the pyramid of scale, are scaled back to the original image resolutions and the overlapped detected regions are combined. Detecting the face region, using a commercial software from Pittsburgh Pattern Recognition ∗ (PittPatt), helps cut the ear detection time as follows: (i) Reducing the search area; and (ii) Cutting the pyramid of scales by defining the minimum and maximum scale relative to the head size. 4. LOCAL BINARY PATTERNS EAR DESCRIPTION Ojala et al.27 quantify the intensity patterns in local pixel neighborhood patches such as spots, line ends, edges, corners, and other distinct texture patterns, and have been used in face recognition.27 They have shown the LBP operator to be highly discriminative and computationally efficient. Using LBP operators for ear recognition is based on the description of ears as a composition of micro-patterns. The basic LBP operator assigns a a decimal value to each pixel in the image by thresholding (P=8) neighbor pixels at distance (R=1), as shown in fig. 3. The histogram (H) of these decimal values represents the feature vector. Ojala et al.5 extended the LBP operator by using neighborhood of different sizes. Generally, P neighborhood pixels, at distance R, can be used; and bilinear interpolation is used for points out of grid to calculate an approximation of a pixel’s intensity based on the values at surrounding pixels. Figure 4 shows three examples for circular neighborhoods, that will be used through out the paper. 4.1 Uniform Local Binary Patterns Ojala et al.28 called a local binary pattern “uniform”, if it contains at most two bitwise transitions from 0 to 1 or vice versa when the bit pattern is considered circular. For example, the patterns “00000000” and “11111111” have 0 transitions, while ∗ http://www.pittpatt.com/ 4 Figure 3. Basic LBP operator: (1) For a given input image pixel and its 8 neighbors, (2) Each neighbor pixel greater than or equal to the center pixel is assigned 1 otherwise it is assigned 0, (3) These binary values are arranged to form a binary number (01110010), which is transferred to a decimal equivalent (114). (a) (b) (c) Figure 4. Examples for circular neighborhoods: (a) P=8 and R=1, (b) P=8 and R=2, and (c) P=16 and R=2. “00000111” and 10000001 have 2 transitions. LBP histogram, for uniform patterns, has a separate bin for every uniform pattern and only one bin for all non-uniform patterns. Experimentally, using 180 ear images from USTB database, we find: • 90.41% of the patterns in the 8,1 neighborhood (4-a) are uniform, • 88.84% of the patterns in the 8,2 neighborhood (4-b) are uniform, • 81.53% of the patterns in the 16,2 neighborhood (4-c) are uniform. This feature selection method reduced the number of features, for (p=16), from 216 using regular histogram to 243 using uniform pattern histogram; Hence, we decided to use uniform pattern for the rest of the paper. 4.2 Block Based Division For the block based division, the image is divided into N blocks. These blocks can be of arbitrary size and can overlap. The LBP operator is applied to each block separately, and their corresponding histograms H = [h1 h2 ...hN ] are calculated. Integration of these blocks can be: • At the feature level: By concatenating the histograms extracted from various blocks, then the overall histogram H is used for matching. • At the score level: By fusing the various blocks scores (s1 , s2 , ..sN ) generated by matching the histogram extracted from each block alone. Sub-blocks are expected to be more discriminative than using the whole image. However, this approach needs the detected images to be aligned, and depends on the fusing method. We presented experiments to evaluate the effect of this division on ear recognition performance. 5 Figure 5. Sample of PittPatt aligned ear images. 5. EXPERIMENTAL RESULTS This section presents several experiments on evaluating various components of the proposed ear recognition system, as well as a case study of performance in case of occlusion, as follows: • description of the databases used in the experiments. • various experiments to tune the various parameters of the LBP. • studies on detection and registration errors assessment. • comparison of LBP performance against other feature extraction techniques presented in the literature. • a case study about ear recognition performance in case of occlusion. 5.1 Ear Databases Three databases were used for various experiments: 1. The University of Notre Dame (UND) databases † are available to the public (free of charge). Collection E contains 464 visible-light face side profile (ear) images from 114 human subjects. We refer to this data set as UND, and it contains 106 subjects to maintain 2 images per subjects. We used this database for testing, and we call test set 1. 2. The University of Science and Technology Beijing (USTB) databases ‡ are available for academic research.7 IMAGE DATABASE I contains 180 images of 60 volunteers. The ear images in the USTB database I7 are vertically aligned the roll Ψ rotation. We refer to this data set as USTB, and it contains 60 subjects, 3 images per subject. We used this set to tune the parameters of LBP and to train Eigen-Ear technique, and we call it train set. 3. 200 subjects with occluded ear images from FERET database29, 30 were used in the ear occlusion assessment study, and we call it test set 2. † ‡ http://www3.nd.edu/ cvrl/CVRL/Data Sets.html http://www1.ustb.edu.cn/resb/en/index.htm 6 Table 1. Comparison of identification (Rank-1) rate of the LBP technique to tune LBP parameters Operator Whole Image Whole Image Whole Image Whole Image (2 × 2) (3 × 3) (5 × 5) (7 × 7) U LBP8,1 U LBP8,2 U LBP16,2 U LBP16,2 U LBP16,2 U LBP16,2 U LBP16,2 U LBP16,2 (Rank1%) 41.67 58.33 61.67 61.67 80.00 81.67 83.33 83.33 5.2 Tuning Local Binary Patterns To tune various parameters of Local Binary Patterns (LBP) method, we use the train set from USTB database. For all the subjects in the USTB data sets, we used: (i) One ear image per subject as a gallery; (ii) One ear image per subject as a probe. To measure the similarity between the probe histogram H p and gallery histogram H g generated by LBP operator, we used the chi-square distance: p g 2 (Hi,j − Hi,j ) SChi (H p , H g ) = Σj,i ωj ∗ (3) p g (Hi,j + Hi,j ) where i and j refer to the ith bin in histogram corresponding to the j th block, and ωj is the weight for block j. We set up several identification experiments to tune LBP operators; and hence to optimize the performance (details in table 1). The first experiments select the number of neighbors points (P) and the radius of these points from the center (R). U This experiment shows that the LBP16,2 operator achieves the best performance. In a pilot study, we evaluated fusion of these multi-resolution operators at the score level, and found that it did not improve the performance. Hence we decide to U operator for the remaining experiments. use the LBP16,2 We setup a second identification experiment to compare the LBP performance using the whole image against dividing U the image into blocks (2 × 2, 3 × 3, 5 × 5, and 7 × 7). For this experiment, we use the LBP16,2 operator. This experiment indicates that using block size of (5 × 5) yields the best performance. Hence we decide to divide the ear images into (5 × 5) blocks for the remaining experiments. 5.3 Ear Detection and Registration Errors Assessment The face, and hence the ear, can rotate along 3 axis to define rotation angles Φ the roll, Θ the pitch, and Ψ the azimuth (yaw). For a yaw angle beyond ±45, the face view is more toward profile and hence the ear structure starts to appear and can be used for recognition. To test the effect of automatic registration on the proposed ear recognition system, the head was automatically aligned using commercial system developed by Pittsburgh Pattern Recognition (PittPatt). This software detects not only the faces, but also the yaw and roll of the head. We use this roll angular information to automatically align the images and register the ear. We applied this automatic face detection to test set 1, UND data set (as shown in fig. 5). The experiment shows correctly detected face in 105 out of 106 subjects. We refer to the resulted data set as U N DP . Ear detection was implemented using cascaded Adaboost based of Haar features.25 However this method yields segmentation errors that affect the overall performance. To test the effect of automating this step, we used the U N DP . The experiment shows correctly segmented ears in 103 out of 105 subjects. We refer to the resulted data set as U N DA . We conducted an experiment to study the accumulated error effect of automating the alignment, using PittPatt, and the detection for the proposed ear recognition system using LBP technique. The identification rate (Rank-1) for the mentioned experiment: (i) 94.34% using UND set, (ii) 93.33% using U N DP , and 83.50% using U N DA . The general drop in performance returns to segmentation accuracy, in other words the system detects an ear but does not accurately segment the ear region. 7 Figure 6. Cumulative Match Characteristic (CMC) curve of the uniform LBP technique compared to the Eigen-Ear (PCA) and SIFT techniques Table 2. Comparison of identification rate (Rank1) of the LBP technique compared to the PCA and SIFT techniques Operator U LBP (5 × 5, LBP16,2 ) SIFT PCA UND (Rank1%) 94.34 67.92 60.38 U N DA (Rank1%) 83.50 60.19 63.11 5.4 LBP versus other feature extraction techniques We setup a comparison identification experiment against a shape-based technique SIFT16 and an intensity-based technique Eigen-Ear. Other intensity-based techniques as Linear discriminative analysis were not applicable to use as they require U more than one image per subject for training. For this experiment, we use the LBP16,2 operator for an ear images divided into (5 × 5) blocks. The performance measure used to evaluate the various schemes is the Cumulative Match Characteristic (CMC) curve, which is commonly used to depict the performance of an identification system. The horizontal axis of the CMC represents rank n, and the vertical axis represents the cumulative rank probability. In the CMC curve, the y-value is the probability of obtaining the correct identity in the top n positions cumulatively. Figure 6 shows the CMC curves of the block-based LBP technique compared to the Eigen-Ear and SIFT techniques. Table 2, shows Rank-1 of the performance comparison experiment. This low performance of the SIFT technique is due to failure in enrollment; in other words ear images with no or insufficient extracted SIFT points: (i) 28 out of the 106 subjects using UND set; and (ii) 33 out of the 103 subjects for the U N DA data set. The general drop in performance returns to segmentation accuracy, in other words the system detect an ear but did not accurate segment the ear region. We setup another experiment to compare the identification performance of the proposed LBP technique against techU niques which combined wavelet transform and local binary patterns (LBP)18 .19 For this experiment, we use the LBP16,2 operator for an ear images divided into (2 × 2) blocks. Table 3 shows the performance of applying the following wavelet filters: Daubechies (db1, db2, db10, db45), Coiflets (coif1, coif5), Symlets (sym2, sym8, sym24), Discrete Meyer (dmey), biorthogonal (bior1.1, bior1.3), reverse biorthogonal (rbio1.1, rbio3.1).18 Table 3 shows that dividing into (2 × 2) blocks achieve superior performance compared to applying wavelet filters. Hence it can be concluded that wavelet transform is not necessary and regular portioning of the image will yield better performance at lower computational cost. 5.5 Ear Recognition incase of occlusion Hair occlusion is more likely to be top down. Based on visual assessment to 200 occluded ear images from FERET database (test set 2),29, 30 we found 81% of the hair occlusion is top down, 17% from the hair side, and 2% from bottom up (Usually the occlusion for the bottom part return to huge ear rings) or random parts of the ear, as shown in fig. 7 (a-d). 8 U U Table 3. Identification rate (Rank1) using LBP16,2 ) after wavelet transform compared to block-based LBP16,2 ) Operator (2x2 ) (db1) (db2) (db10) (db45) (coif1) (coif5) (sym2) (sym8) (sym24) (dmey) (bior1.1) (bior1.3) (rbio1.1) (rbio3.1) UND (Rank1%) 90.84 65.71 70.48 60.00 31.43 63.81 50.48 70.48 61.90 47.62 38.10 65.71 74.29 65.71 79.05 U N DA (Rank1%) 82.52 48.54 53.40 46.60 36.89 48.54 42.72 53.40 50.49 37.86 34.95 48.54 49.51 48.54 57.28 (a) (b) (c) (d) (e) (f ) (g) (h) Figure 7. Example of real occluded ear images are given (a-d). Test images using mask covering part of the test images: (e) left; (f) right; (c) top; and (d) bottom. We setup an experiment with various occlusion masks as shown in fig. 7 (e-h), simulating the real hair occlusion shown U in fig. 7 (a-d). For this experiment, we use the LBP16,2 operator for an ear images divided into (5 × 5) blocks. Table VI shows the performance of the proposed method with various occlusion percentages up to 80%. Table 4. Identification (Rank1) rate for various occlusion percentages and location using UND database Occlusion Top (20%) Bottom (20%) Left (20%) Right (20%) Top (60%) Bottom (60%) Left (60%) Right (60%) Rank1% 92.45 94.34 97.17 93.40 80.19 92.45 94.34 78.30 Occlusion Top (40%) Bottom (40%) Left (40%) Right (40%) Top (80%) Bottom (80%) Left (80%) Right (80%) 9 Rank1% 88.68 97.17 94.34 89.62 60.38 84.91 90.57 51.89 6. CONCLUSION AND FUTURE WORK In this paper, we proposed a fully automated real-time ear recognition system. The proposed system can process up to five images (640 x 480) per sec. This system used Haar based cascaded AdaBoost for ear detection, and depends on the head rotation angle to align these detect ears. Then LBP histograms, corresponding to various non-overlapped blocks, are concatenated into one histogram to represent the detected ear texture. We presented detailed experiments to tune the LBP operators, and studied the error accumulates by various components. We compared the performance of the proposed technique to the bench mark and related ear recognition techniques. The experimental results show that the proposed ear recognition method achieved superior performance. Then we presented a set of experiments to determine the effect of hair occlusion, and found that the system can still achieve reasonable performance even in presence of 60% occlusion. Future work to enhance the performance of the proposed methods includes: (i) Advanced methods for dividing ear images into significant regions based on ear anatomy; and (ii) Integrating another preprocessing steps to enhance the ear segmentation step. REFERENCES [1] Jain, A. and Ross, A., [Handbook of Biometrics], ch. 1 Introduction to Biometrics, 1–22, Springer (2007). [2] Bertillon, A., [Signaletic Instructions Including: The Theory and Practice of Anthropometrical Identification ], R.W. McClaughry translation, The Werner Company (1896). [3] Iannarelli, A., [Ear Identification, Forensic Identification Series ], Paramount Publishing Company, Fremont, California (1989). [4] Burge, M. and Burger, W., “Ear biometrics in computer vision,” in [the 15th International Conference on Pattern Recognition (ICPR)], 826–830 (2000). [5] Ojala, T., Pietikainen, M., and Maenpaa, T., “Multiresolution gray-scale and rotation invariant texture classification with local binary patterns,” IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI) 24(7), 971–987 (2002). [6] Miller, P. E., Lyle, J. R., Pundlik, S. J., and Woodard, D. L., “Performance evaluation of local appearance based periocular recognition,” in [the Biometrics: Theory, Applications, and Systems (BTAS)], (2010). [7] USTB, “University of science and technology beijing USTB database.” available at: http : //www1.ustb.edu.cn/resb/en/index.htm. [8] UND, “University of notre dame UND databases.” available at: http : //www3.nd.edu/ cvrl/CV RL/Data Sets.html. [9] Ansari, S. and Gupta, P., “Localization of ear using outer helix curve of the ear,” in [the International Conference on Computing: Theory and Applications (ICCTA)], 688–692 (2007). [10] AbdelMottaleb, M. and Zhou, J., “Human ear recognition from face profile images,” in [the 2nd International Conference on Biometrics (ICB)], 786–792 (2006). [11] Yuizono, T., Wang, Y., Satoh, K., and Nakayama, S., “Study on individual recognition for ear images by using genetic local search,” in [the Congress on Evolutionary Computation (CEC)], 237–242 (2002). [12] Prakash, S., Jayaraman, U., and Gupta, P., “A skin-color and template based technique for automatic ear detection,” in [the 17th International Conference on Advances in Pattern Recognition (ICAPR)], (2009). [13] HajSaid, E., Abaza, A., and Ammar, H., “Ear segmentation in color facial images using mathematical morphology,” in [the 6th Biometric Consortium Conference (BCC)], (2008). [14] Chang, K., Bowyer, K., Sarkar, S., and Victor, B., “Comparison and combination of ear and face images in appearance-based biometrics,” IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI) 25, 1160– 1165 (2003). [15] Dewi, K. and Yahagi, T., “Ear photo recognition using scale invariant keypoints,” in [the 2nd International Association of Science and Technology for Development (IASTED): Conference on Computational Intelligence], 253–258 (2006). [16] Lowe, D. G., “Distinctive image features from scale-invariant keypoints,” International Journal of Computer Vision (IJCV) 60(2), 91–110 (2004). 10 [17] Kisku, D. R., Mehrotra, H., Gupta, P., and Sing, J. K., “SIFT-based ear recognition by fusion of detected keypoints from color similarity slice regions,” in [the International Conference on Advances in Computational Tools for Engineering Applications (ACTEA)], 380–385 (2009). [18] Feng, J. and Mu, Z., “Texture analysis for ear recognition using local feature descriptor and transform filter,” in [SPIE: Pattern Recognition and Computer Vision], (2009). [19] Wang, Y., chun Mu, Z., and Zeng, H., “Block-based and multi-resolution methods for ear recognition using wavelet transform and uniform local binary patterns,” in [the 19th International Conference on Pattern Recognition (ICPR)], (2008). [20] Yuan, L., chun Mu, Z., Zhang, Y., and Liu, K., “Ear recognition using improved non-negative matrix factorization,” in [the 18th International Conference on Pattern Recognition (ICPR)], 501–504 (2006). [21] Yuan, L., hua Wang, Z., and chun Mu, Z., “Ear recognition under partial occlusion based on neighborhood preserving embedding,” in [the SPIE: Biometric Technology for Human Identification VII], 7667 (2010). [22] Kocaman, B., Kirci, M., Gunes, E. O., Cakir, Y., and Ozbudak, O., “On ear biometrics,” in [the IEEE Region 8 Conference (EUROCON) ], (2009). [23] ArbabZavar, B., Nixon, M., and Hurley, D., “On model-based analysis of ear biometrics,” in [the Biometrics: Theory, Applications, and Systems (BTAS)], (2007). [24] Bustard, J. and Nixon, M., “Robust 2D ear registration and recognition based on SIFT point matching,” in [the Biometrics: Theory, Applications, and Systems (BTAS)], (2008). [25] Abaza, A., Hebert, C., and Harrison, M. A. F., “Fast learning ear detection for real-time surveillance,” in [the Biometrics: Theory, Applications, and Systems BTAS], (2010). [26] Viola, P. and Jones, M., “Robust real-time face detection,” International Journal of Computer Vision (IJCV) 57(2), 137–154 (2004). [27] Ojala, T., Pietikinen, M., and Harwood, D., “A comparative study of texture measures with classification based on featured distributions,” Pattern Recognition 29(1), 51–59 (1996). [28] Ojala, T., Hadid, A., and Pietikainen, M., “Face description with local binary patterns: Application to face recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI) 28(12), 2037–2041 (2006). [29] Phillips, P. J., Wechsler, H., Huang, J., and Rauss, P. J., “The feret database and evaluation procedure for face recognition algorithms,” Image and Vision Computing 16(5), 295–306 (1998). [30] Phillips, P., Moon, H., Rizvi, S. A., and Rauss, P. J., “The feret evaluation methodology for face recognition algorithms,” IEEE Transactions on Pattern Analysis and Machine Intelligence 22(10), 1090–1104 (2000). 11
© Copyright 2026 Paperzz