IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 37, NO. 5, OCTOBER 2007 1373 Correspondence Comments on “On Image Matrix Based Feature Extraction Algorithms” where s= Quanxue Gao, Lei Zhang, David Zhang, and Jian Yang K m 1 akj . K ×m (4) k=1 j=1 Abstract—A class of image-matrix-based feature extraction algorithms has been discussed earlier. The correspondence argues that 2-D principal component analysis and Fisher linear discriminant (FLD) are equivalent to block-based PCA and FLD. In this correspondence, we point out that this statement is not rigorous. Index Terms—Feature extraction, 2-D linear discriminant analysis (LDA), 2-D principal component analysis (PCA). Comparing (2) with (4), it is easy to obtain s = aj , j = 1, 2, . . . , m (5) and hence Gb = Gt I. T WO -D IMENSIONAL P RINCIPAL C OMPONENT A NALYSIS (PCA) V ERSUS B LOCK -B ASED PCA In the above paper [1], it is argued that 2-D PCA [2] is equivalent to block-based PCA [3]. This statement is rigorous if we assume that the input images have been shifted to zero mean. However, in the actual implementations of 2-D PCA and block-based PCA, they have different procedures of zero-mean shifting. This leads to different results of 2-D PCA and block-based PCA. Denote by akj ∈ R1×n and aj ∈ R1×n the jth row of Ak (k = 1, 2, . . . , K) and A, respectively, where Ak ∈ Rm×n is the training image, K is the number of training samples, and A is the mean image of training samples. In 2-D PCA [2], the image covariance matrix can be expressed as K 1 (Ak − A)T (Ak − A) Gt = K k=1 = K m T k 1 k aj − a j a j − aj K (1) k=1 j=1 rather than Gb = Gt , as claimed in [1]. We have the following comment: Comment 1. Two-dimensional PCA is not equivalent to block-based PCA by setting the blocks as the rows of an image, i.e., 2-D PCA is different from row-based PCA. II. T WO -D IMENSIONAL F ISHER L INEAR D ISCRIMINANT (FLD) V ERSUS B LOCK -B ASED FLD . . , K), includGiven K training samples, Ak ∈ Rm×n (k = 1, 2, . C ing C classes. The ith class Ci has Ki samples, and i=1 Ki = K. i Denote by bkl , bl , and bl the lth column of Ak , A, and Ai , respectively, where A is the global mean image and Ai is the mean image of the ith class Ci . In 2-D FLD [4]–[6], if the right projection matrix R is an identity matrix, then the image between-class scatter matrix Gb and within-class scatter matrix Gw may be expressed as Gb = 1 k aj , K Gw = K j = 1, 2, . . . , m. In general, aj1 = aj2 ∀j1 = j2 . In block-based PCA [3], if we view each row of an image as a block, then the image covariance matrix is m K T k 1 k aj − s aj − s K C n T 1 1 i i bkl − bl bkl − bl C Ki i=1 (2) k=1 Gb = C n T 1 i i bl − bl bl − bl C (7) i=1 l=1 where aj = (6) i (8) Ak ∈Ci l=1 K where bl = (1/Ki ) A ∈C bkl , and bl = (1/K) k=1 bkl . i k Similar to block-based PCA, if each column of an image is considered a block, i.e., bkl is viewed as a training sample in FLD, then the image between-class scatter matrix Sb and within-class scatter matrix Sw are (3) k=1 j=1 Sb = Manuscript received December 6, 2006. This paper was recommended by Associate Editor M. Pantic. Q. Gao is with the School of Telecommunication Engineering, Xidian University, Xi’an 710071, China, and also with the Department of Computing, Hong Kong Polytechnic University, Kowloon, Hong Kong. L. Zhang, D. Zhang, and J. Yang are with the Department of Computing, Hong Kong Polytechnic University, Kowloon, Hong Kong. Color versions of one or more of the figures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identifier 10.1109/TSMCB.2007.899415 C 1 (si − s)(si − s)T C (9) i=1 Sw = C n T 1 1 k bl − si bkl − si C Ki i=1 (10) Ak ∈Ci l=1 n bk and s = (1/(K × n)) where si = (1/(Ki × n)) A ∈C l=1 l i k K n k b are the mean vectors of the ith class and all samples, k=1 l=1 l respectively. 1083-4419/$25.00 © 2007 IEEE 1374 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 37, NO. 5, OCTOBER 2007 Fig. 1. (a) Recognition accuracy curves of row-based PCA and 2-D PCA. (b) Recognition accuracy curves of row-FLD and 2-D FLD. IV. C ONCLUSION Compare (7) with (8) and (9) with (10), it is readily seen that Gb = Sb Gw = Sw . (11) We have the following comments: Comment 2. Two-dimensional FLD is not equivalent to columnbased linear discriminant analysis (LDA) by assuming that the right projection matrix is identity. Comment 3. Two-dimensional FLD is not equivalent to line-based LDA by assuming that the left projection matrix is identity. Comment 4. Two-dimensional FLD methods are not equivalent to block-based FLD approaches. This correspondence indicates that the statement in [1], i.e., 2-D PCA and 2-D FLD are equivalent to block-based PCA and FLD, is not rigorous. Our comments can be helpful in understanding the concepts of 2-D PCA, 2-D FLD, block-based PCA, and FLD. ACKNOWLEDGMENT The authors would like to thank the anonymous reviewers and the editor for their constructive comments and suggestions. III. E XPERIMENTAL R ESULT R EFERENCES We use the Olivetti Research Ltd. database (http://www.cam-orl. co.uk) to illustrate the difference between the block-based PCA and FLD methods, and the 2-D PCA and 2-D FLD methods. In the experiments, the last five images per class are used for training, and the remaining images are used for testing. Fig. 1(a) plots the recognition accuracy curves of the row-based PCA and 2-D PCA by using the cosine distance. Fig. 1(b) plots the recognition accuracy curves of rowFLD and 2-D FLD by using Euclidean distance and assuming that the left projection matrix is identity. From Fig. 1(a), it can be seen that 2-D PCA and row-based PCA have different performances. From Fig. 1(b), we see that 2-D FLD by using identity left projection matrix obviously has better recognition accuracy than that of row-FLD. [1] L. Wang, X. Wang, and J. Feng, “On image matrix based feature extraction algorithms,” IEEE Trans. Syst., Man, Cybern. B, Cybern., vol. 36, no. 1, pp. 194–197, Feb. 2006. [2] J. Yang and D. Zhang, “Two-dimensional PCA: A new approach to appearance-based face representation and recognition,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 26, no. 1, pp. 131–137, Jan. 2004. [3] G. Rajkiran and K. Vijayan, “An improved face recognition technique based on modular PCA approach,” Pattern Recognit. Lett., vol. 25, no. 4, pp. 429–436, Apr. 2004. [4] J. Ye, R. Janardan, and Q. Li, “Two-dimensional linear discriminant analysis,” in Proc. NIPS, Vancouver, BC, Canada, 2004, pp. 1569–1576. [5] J. Ye, “Generalized low rank approximations of matrices,” in Proc. ICML, Banff, AB, Canada, 2004, pp. 887–894. [6] M. Li and B. Yuan, “2D-LDA: A novel statistical linear discriminant analysis for image matrix,” Pattern Recognit. Lett., vol. 26, no. 5, pp. 527– 532, 2005.
© Copyright 2026 Paperzz