Application of 2DPCA Based Techniques in DCT Domain for Face Recognition Messaoud Bengherabi1, Lamia Mezai1, Farid Harizi1, Abderrazak Guessoum2, and Mohamed Cheriet3 1 Centre de Développement des Technologies Avancées- Algeria Division Architecture des Systèmes et MultiMédia Cité 20 Aout, BP 11, Baba Hassen, Algiers-Algeriabengherabi@yahoo.com, l_mezai@yahoo.fr, harizihourizi@yahoo.fr 2 Université Saad Dahlab de Blida – Algeria Laboratoire Traitement de signal et d’imagerie Route De Soumaa BP 270 Blida guessouma@hotmail.com 3 École des Technologies Supérieur –Québec- CanadaLaboratoire d’Imagerie, de Vision et d’Intelligence Artificielle 1100, Rue Notre-Dame Ouest, Montréal (Québec) H3C 1K3 Canada mohamed.cheriet@gpa.etsmtl.ca Abstract. In this paper, we introduce 2DPCA, DiaPCA and DiaPCA+2DPCA in DCT domain for the aim of face recognition. The 2D DCT transform has been used as a preprocessing step, then 2DPCA, DiaPCA and DiaPCA+2DPCA are applied on the upper left corner block of the global 2D DCT transform matrix of the original images. The ORL face database is used to compare the proposed approach with the conventional ones without DCT under Four matrix similarity measures: Frobenuis, Yang, Assembled Matrix Distance (AMD) and Volume Measure (VM). The experiments show that in addition to the significant gain in both the training and testing times, the recognition rate using 2DPCA, DiaPCA and DiaPCA+2DPCA in DCT domain is generally better or at least competitive with the recognition rates obtained by applying these three 2D appearance based statistical techniques directly on the raw pixel images; especially under the VM similarity measure. Keywords: Two-Dimensional PCA (2DPCA), Diagonal PCA (DiaPCA), DiaPCA+2DPCA, face recognition, 2D Discrete Cosine Transform (2D DCT). 1 Introduction Different appearance based statistical methods for face recognition have been proposed in literature. But the most popular ones are Principal Component Analysis (PCA) [1] and Linear Discriminate Analysis (LDA) [2], which process images as 2D holistic patterns. However, a limitation of PCA and LDA is that both involve eigendecomposition, which is extremely time-consuming for high dimensional data. Recently, a new technique called two-dimensional principal component analysis 2DPCA was proposed by J. Yang et al. [3] for face recognition. Its idea is to estimate the covariance matrix based on the 2D original training image matrices, resulting in a E. Corchado et al. (Eds.): CISIS 2008, ASC 53, pp. 243–250, 2009. © Springer-Verlag Berlin Heidelberg 2009 springerlink.com 244 M. Bengherabi et al. covariance matrix whose size is equal to the width of images, which is quite small compared with the one used in PCA. However, the projection vectors of 2DPCA reflect only the variations between the rows of images, while discarding the variations of columns. A method called Diagonal Principal Component Analysis (DiaPCA) is proposed by D. Zhang et al. [4] to resolve this problem. DiaPCA seeks the projection vectors from diagonal face images [4] obtained from the original ones to ensure that the correlation between rows and those of columns is taken into account. An efficient 2D techniques that results from the combination of DiaPCA and 2DPCA (DiaPCA+2DPCA) is proposed also in [4]. Discrete cosine transform (DCT) has been used as a feature extraction step in various studies on face recognition. This results in a significant reduction of computational complexity and better recognition rates [5, 6]. DCT provides excellent energy compaction and a number of fast algorithms exist for calculating it. In this paper, we introduce 2DPCA, DiaPCA and DiaPCA+2DPCA in DCT domain for face recognition. The DCT transform has been used as a feature extraction step, then 2DPCA, DiaPCA and DiaPCA+2DPCA are applied only on the upper left corner block of the global DCT transform matrix of the original images. Our proposed approach is tested against conventional approaches without DCT under Four matrix similarity measures: Frobenuis, Yang, Assembled Matrix Distance (AMD) and Volume Measure (VM). The rest of this paper is organized as follows. In Section 2 we give a review of 2DPCA, DiaPCA and DiaPCA+2DPCA approaches and also we review different matrix similarity measures. In section 3, we present our contribution. In section 4 we report the experimental results and highlight a possible perspective of this work. Finally, in section 5 we conclude this paper. 2 Overview of 2DPCA, DiaPCA, DiaPCA+2DPCA and Matrix Similarity Measures 2.1 Overview of 2D PCA, DiaPCA and DiaPCA+2DPCA 2.1.1 Two-Dimensional PCA Given M training face images, denoted by m×n matrices Ak (k = 1, 2… M), twodimensional PCA (2DPCA) first uses all the training images to construct the image covariance matrix G given by [3] G= 1 M ∑ (A M k −A k =1 ) (A T k −A ) (1) Where A is the mean image of all training images. Then, the projection axes of 2DPCA, Xopt=[x1… xd] can be obtained by solving the algebraic eigenvalue problem Gxi=λixi, where xi is the eigenvector corresponding to the ith largest eigenvalue of G [3]. The low dimensional feature matrix C of a test image matrix A is extracted by C = AX opt (2) In Eq.(2) the dimension of 2DPCA projector Xopt is n×d, and the dimension of 2DPCA feature matrix C is m×d. Application of 2DPCA Based Techniques in DCT Domain 245 2.1.2 Diagonal Principal Component Analysis Suppose that there are M training face images, denoted by m×n matrices Ak(k = 1, 2, …, M). For each training face image Ak, we calculate the corresponding diagonal face image Bk as it is defined in [4]. Based on these diagonal faces, diagonal covariance matrix is defined as [4]: G DIAG = Where B = 1 M 1 M ∑ (B M k =1 k −B ) (B T k −B ) (3) M ∑B k k =1 is the mean diagonal face. According to Eq. (3), the projection vectors Xopt=[x1, …, xd] can be obtained by computing the d eigenvectors corresponding to the d biggest eigenvalues of GDIAG. The training faces Ak’s are projected onto Xopt, yielding m×d feature matrices. C k = Ak X opt (4) Given a test face image A, first use Eq. (4) to get the feature matrix C = AX opt , then a matrix similarity metric can be used for classification. 2.1.3 DiaPCA+2DPCA Suppose the n by d matrix X=[x1, …, xd] is the projection matrix of DiaPCA. Let Y=[y1, …, yd] the projection matrix of 2DPCA is computed as follows: When the height m is equal to the width n, Y is obtained by computing the q eigenvectors corresponding to the q biggest eigenvalues of the image covarinace matrix 1 M (A − A)T (A − A) . On the other hand, when the height m is not equal to the width M ∑ k =1 k k n, Y is obtained by computing the q eigenvectors corresponding to the q biggest ei- ∑ (A M genvalues of the alternative image covariance matrix 1 M k =1 k )( ) T − A Ak − A . Projecting training faces Aks onto X and Y together, yielding the q×d feature matrices D k = Y T Ak X (5) Given a test face image A, first use Eq. (5) to get the feature matrix D = Y T AX , then a matrix similarity metric can be used for classification. 2.2 Overview of Matrix Similarity Measures An important aspect of 2D appearance based face recognition approaches is the similarity measure between matrix features used at the decision level. In our work, we have used four matrix similarity measures. 2.2.1 Frobenius Distance Given two feature matrices A = (aij)m×d and B = (bij)m×d, the Frobenius distance [7] measure is given by: ⎛ m d F ( A, B ) = ⎜⎜ ∑ ⎝ i =1 ∑ (a 12 d j =1 ij 2⎞ − bij ) ⎟⎟ ⎠ (6) 246 M. Bengherabi et al. 2.2.2 Yang Distance Measure Given two feature matrices A = (aij)m×d and B = (bij)m×d, the Yang distance [7] is given by: 12 d ⎛ m 2⎞ dY ( A, B ) = ∑ ⎜ ∑ (aij − bij ) ⎟ j =1 ⎝ i =1 ⎠ (7) 2.2.3 Assembled Matrix Distance (AMD) A new distance called assembled matrix distance (AMD) metric to calculate the distance between two feature matrices is proposed recently by Zuo et al [7]. Given two feature matrices A = (aij)m×d and B = (bij)m×d, the assembled matrix distance dAMD(A,B) is defined as follows : (1 2 ) p ⎞ ⎛ d ⎛ m 2⎞ ⎟ d AMD ( A, B ) = ⎜ ∑ ⎜ ∑ (aij − bij ) ⎟ ⎟ ⎜ j =1 ⎝ i =1 ⎠ ⎠ ⎝ 12 ( p > 0) (8) It was experimentally verified in [7] that best recognition rate can be obtained when p≤0.125 while it decrease as p increases. In our work the parameter p is set equal to 0.125. 2.2.4 Volume Measure (VM) The VM similarity measure is based on the theory of high-dimensional geometry space. The volume of an m×n matrix of rank p is given by [8] ∑det Vol A = ( I ,J )∈N 2 A IJ (9) where AIJ denotes the submatrix of A with rows I and columns J, N is the index set of p×p nonsingular submatrix of A, and if p=0, then Vol A = 0 by definition. 3 The Proposed Approach In this section, we introduce 2DPCA, DiaPCA and DiaPCA+2DPCA in DCT domain for the aim of face recognition. The DCT is a popular technique in imaging and video compression, which was first applied in image compression in 1974 by Ahmed et al [9]. Applying the DCT to an input sequence decomposes it into a weighted sum of basis cosine sequences. our methodology is based on the use of the 2D DCT as a feature extraction or preprocessing step, then 2DPCA, DiaPCA and DiaPCA+2DPCA are applied to w×w upper left block of the global 2D DCT transform matrix of the original images. In this approach, we keep only a sub-block containing the first coefficients of the 2D DCT matrix as shown in Fig.1, from the fact that, the most significant information is contained in these coefficients. 2D DCT c11 c12 … c1w . . cw1 cw2 … cww Fig. 1. Feature extraction in our approach Application of 2DPCA Based Techniques in DCT Domain 247 With this approach and inversely to what is presented in literature of DCT-based face recognition approaches, the 2D structure is kept and the dimensionality reduction is carried out. Then, the 2DPCA, DiaPCA and DiaPCA+2DPCA are applied to w×w block of 2D DCT coefficients. The training and testing block diagrams describing the proposed approach is illustrated in Fig.2. Training a lgorithm based on 92DPCA 9Dia PCA 9Dia PCA+2DPCA Block w*w of 2D DCT coefficients 2D DCT Tra ined Model Training data 2D DCT ima ge Projection of the DCT bloc of the test ima ge using the eigenvectors of 92DPCA 9Dia PCA 9Dia PCA+2DPCA Block w*w of 2D DCT coefficients 2D DCT Test ima ge 2D DCT Block Features 2D DCT ima ge Compa rison using 9Frobenius 9Yang 9AMD 9VM 2D DCT Block Fea tures Decision Fig. 2. Block diagram of 2DPCA, DiaPCA and DiaPCA+2DPCA in DCT domain 4 Experimental Results and Discussion In this part, we evaluate the performance of 2DPCA, DiaPCA and DiaPCA+2DPCA in DCT domain and we compare it to the original 2DPCA, DiaPCA and DiaPCA+2DPCA methods. All the experiments are carried out on a PENTUIM 4 PC with 3.2GHz CPU and 1Gbyte memory. Matlab [10] is used to carry out these experiments. The database used in this research is the ORL [11] (Olivetti Research Laboratory) face database. This database contains 400 images for 40 individuals, for each person we have 10 different images of size 112×92 pixels. For some subjects, the images captured at different times. The facial expressions and facial appearance also vary. Ten images of one person from the ORL database are shown in Fig.3. In our experiment, we have used the first five image samples per class for training and the remaining images for test. So, the total number of training samples and test samples were both 200. Herein and without DCT the size of diagonal covariance matrix is 92×92, and each feature matrix with a size of 112×p where p varies from 1 to 92. However with DCT preprocessing the dimension of these matrices depends on the w×w DCT block where w varies from 8 to 64. We have calculated the recognition rate of 2DPCA, DiaPCA, DiaPCA+2DPCA with and without DCT. In this experiment, we have investigated the effect of the matrix metric on the performance of the 2D face recognition approaches presented in section 2. We see from table 1, that the VM provides the best results whereas the Frobenius gives the worst ones, this is justified by the fact that the Frobenius metric is just the sum of the 248 M. Bengherabi et al. (a) (b) Fig. 3. Ten images of one subject in the ORL face database, (a) Training, (b) Testing Euclidean distance between two feature vectors in a feature matrix. So, this measure is not compatible with the high-dimensional geometry theory [8]. Table 1. Best recognition rates of 2DPCA, DiaPCA and DiaPCA+2DPCA without DCT Methods 2DPCA DiaPCA DiaPCA+2DPCA Frobenius 91.50 (112×8) 91.50 (112×8) 92.50 (16×10) Yang 93.00 (112×7) 92.50 (112×10) 94.00 (13×11) AMD p=0,125 95.00 (112×4) 91.50 (112×8) 93.00 (12×6) Volume Distance 95.00 (112×3) 94.00 (112×9) 96.00 (21×8) Tables 2, and Table 3 summarize the best performances under different 2D DCT block sizes and different matrix similarity measures. Table 2. 2DPCA, DiaPCA and DiaPCA+2DPCA under different DCT block sizes using the Frobenius and Yang matrix distance Best Recognition rate (feature matrix dimension) DiaPCA+2DPCA 2DPCA DiaPCA Yang 91.50 (6×6) 93.50 (8×6) 93.50 (8×5) 92.00 (9×5) 93.00 (9×6) 95.00 (9×9) 92.00 (10×5) 94.50 (10×6) 95.50 (10×9) 92.00 (9×5) 94.00 (11×6) 95.50 (11×5) 91.50 (9×5) 94.50 (12×6) 95.50 (12×5) 92.00 (12×11) 94.50 (13×6) 95.00 (13×5) 92.00 (12×7) 94.50 (14×6) 94.50 (14×5) 2D DCT block size 8×8 9×9 10×10 11×11 12×12 13×13 14×14 91.50 (8×8) 92.00 (9×9) 91.50 (10×5) 92.00 (11×8) 92.00 (12×8) 91.50 (13×7) 92.00 (14×7) DiaPCA Frobenius 91.50 (8×6) 92.00 (9×5) 92.00 (10×5) 91.50 (11×5) 91.50 (12×10) 92.00 (13×11) 91.50 (14×7) 15×15 16×16 32×32 91.50 (15×5) 92.50 (16×10) 92.00 (32×6) 91.50 (15×5) 91.50 (16×11) 91.50 (32×6) 92.00 (13×15) 92.00 (4×10) 92.00 (11×7) 94.00 (15×9) 94.00 (16×7) 93.00 (32×6) 94.50 (15×5) 94.50 (16×5) 93.50 (32×5) 95.50 (12×5) 95.00 (12×5) 95.00 (12×5) 64×64 91.50 (64×6) 91.00 (32×6) 92.00 (14×12) 93.00 (64×7) 93.50 (64×5) 95.00 (12×5) 2DPCA DiaPCA+2DPCA 93.50 (8×5) 95.00 (9×9) 95.50 (10×9) 95.50 (11×5) 95.50 (12×5) 95.00 (11×5) 95.00 (12×5) From these four tables, we notice that in addition to the importance of matrix similarity measures, by the use of DCT we have always better performance in terms of recognition rate and this is valid for all matrix measures, we have only to choose the DCT block size and appropriate feature matrix dimension. An important remark is that a block size of 16×16 or less is sufficient to have the optimal performance. So, this results in a significant reduction in training and testing time. This significant gain Application of 2DPCA Based Techniques in DCT Domain 249 Table 3. 2DPCA, DiaPCA and DiaPCA+2DPCA under different DCT block sizes using the AMD distance and VM similarity measure on the ORL database 2D DCT block size 2DPCA 8×8 9×9 10×10 11×11 12×12 13×13 14×14 15×15 16×16 32×32 64×64 94.00 (8×4) 94.50 (9×4) 94.50 (10×4) 95.50 (11×5) 95.50 (12×5) 96.00 (13×4) 96.00 (14×4) 96.00 (15×4) 96.00 (16×4) 95.50 (32×4) 95.00 (64×4) DiaPCA AMD 95.00 (8×6) 94.50 (9×5) 95.50 (10×5) 96.00 (11×5) 96.50 (12×7) 95.50 (13×5) 95.00 (14×5) 95.00 (15×5) 95.50 (16×5) 95.00 (32×9) 94.50 (64×9) Best Recognition rate (feature matrix dimension) DiaPCA+2DPCA 2DPCA DiaPCA VM 95.00 (7×5) 96.00 (8×3) 93.50 (8×4) 94.50 (9×5) 95.00 (9×4) 95.00 (9×5) 96.00 (9×7) 95.00 (10×3) 95.00 (10×4) 94.50 (11×3) 95.50 (11×3) 96.50 (9×6) 95.50 (12×5) 96.00 (12×5) 96.50 (9×7) 95.50 (12×5) 96.00 (13×9) 96.00 (13×5) 95.50 (10×5) 95.00 (14×3) 95.50 (14×5) 96.00 (9×7) 96.00 (15×8) 96.00 (15×5) 95.50 (16×8) 96.00 (16×5) 96.50 (12×5) 96.00 (11×5) 95.00 (32×3) 95.50 (32×5) 96.00 (12×5) 95.00 (64×3) 95.00 (64×5) DiaPCA+2DPCA 93.50 (8×4) 95.00 (9×5) 95.00 (10×4) 95.50 (11×3) 96.00 (11×5) 96.50 (10×5) 96.50 (10×5) 96.50 (10×5) 96.50 (10×5) 96.50 (9×5) 96.50 (21×5) in computation is better illustrated in table 4 and table 5, which illustrate the total training and total testing time of 200 persons -in seconds - of the ORL database under 2DPCA, DiaPCA and DiaPCA+2DPCA without and with DCT, respectively. We should mention that the computation of DCT was not taken into consideration when computing the training and testing time of DCT based approaches. Table 4. Training and testing time without DCT using Frobenius matrix distance Methods Training time in sec Testing time in sec 2DPCA 5.837 (112×8) 1.294 (112×8) DiaPCA 5.886 (112×8) 2.779 (112×8) DiaPCA+2DPCA 10.99 (16×10) 0.78 (16×10) Table 5. Training and testing time with DCT using the Frobenius distance and the same matrixfeature dimensions as in Table2 2D DCT block size 8×8 9×9 10×10 11×11 12×12 13×13 14×14 15×15 16×16 2DPCA 0.047 0.048 0.047 0.048 0.063 0.062 0.079 0.094 0.125 Training time in sec DiaPCA DiaPCA+2DPCA 0.047 0.047 0.048 0.124 0.048 0.094 0.047 0.063 0.046 0.094 0.047 0.126 0.062 0.14 0.078 0.173 0.141 0.219 2DPCA 0.655 0.626 0.611 0.578 0.641 0.642 0.656 0.641 0.813 Testing time in sec DiaPCA DiaPCA+2DPCA 0.704 0.61 0.671 0.656 0.719 0.625 0.734 0.5 0.764 0.657 0.843 0.796 0.735 0.718 0.702 0.796 0.829 0.827 We can conclude from this experiment, that the proposed approach is very efficient in weakly constrained environments, which is the case of the ORL database. 5 Conclusion In this paper, 2DPCA, DiaPCA and DiaPCA+2PCA are introduced in DCT domain. The main advantage of the DCT transform is that it discards redundant information and it can be used as a feature extraction step. So, computational complexity is significantly reduced. The experimental results show that in addition to the significant gain in both the training and testing times, the recognition rate using 2DPCA, DiaPCA and DiaPCA+2DPCA in DCT domain is generally better or at least competitive with the 250 M. Bengherabi et al. recognition rates obtained by applying these three techniques directly on the raw pixel images; especially under the VM similarity measure. The proposed approaches will be very efficient for real time face identification applications such as telesurveillance and access control. References 1. Turk, M., Pentland, A.: “Eigenfaces for Recognition. Journal of Cognitive Neurosicence 3(1), 71–86 (1991) 2. Belhumeur, P.N., Hespanha, J.P., Kriegman, D.J.: Eigenfaces vs. fisherfaces: Recognition using class specific linear projection. IEEETrans. on Patt. Anal. and Mach. Intel. 19(7), 711–720 (1997) 3. Yang, J., Zhang, D., Frangi, A.F., Yang, J.Y.: Two-Dimensional PCA: A New Approach to Appearance- Based Face Representation and Recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence 26(1), 131–137 (2004) 4. Zhang, D., Zhou, Z.H., Chen, S.: “Diagonal Principal Component Analysis for Face Recognition. Pattern Recognition 39(1), 140–142 (2006) 5. Hafed, Z.M., Levine, M.D.: “Face recognition using the discrete cosine transform. International Journal of Computer Vision 43(3) (2001) 6. Chen, W., Er, M.J., Wu, S.: PCA and LDA in DCT domain. Pattern Recognition Letters 26(15), 2474–2482 (2005) 7. Zuo, W., Zhang, D., Wang, K.: An assembled matrix distance metric for 2DPCA-based image recognition. Pattern Recognition Letters 27(3), 210–216 (2006) 8. Meng, J., Zhang, W.: Volume measure in 2DPCA-based face recognition. Pattern Recognition Letters 28(10), 1203–1208 (2007) 9. Ahmed, N., Natarajan, T., Rao, K.: Discrete cosine transform. IEEE Trans. on Computers 23(1), 90–93 (1974) 10. Matlab, The Language of Technical Computing, Version 7 (2004), http://www.mathworks.com 11. ORL. The ORL face database at the AT&T (Olivetti) Research Laboratory (1992), http://www.uk.research.att.com/facedatabase.html
© Copyright 2025 Paperzz