1
Training Database Step 1 : In general approach of PCA, each image is divided into nxn blocks or pixels. Then all pixel values are taken into a single one dimensional matrix. Showed in figure:1 Figure : 1 Figure : 2 Our proposal here is that if we find half image of actual image, we divide each test image and then convert to single column matrix. Afterwards, general PCA will be applied. Step-2 : Center data The images are mean centered by subtracting the mean image from each image vector. Step-3 : Create data matrix These vectors are combined, side-by-side, to create a data matrix of size NxP (where P is the number of images). An Approach to Face Recognition System Using Principal Component Analysis (PCA) Fattah Muhammad Tahabi, Ratul Antik Das Department of Computer Science and Engineering (CSE), BUET Motivation Face recognition has attracted significant attention in the past decades because of its potential applications in biometrics, information security, and law enforcement. Many methods have been suggested to recognize faces. PCA has been widely investigated and has become one of the most popular face-recognition approaches. Methodology =[ ] T = , where = = = |… ] = Step-4 : Calculate covariance matrix The data matrix X is multiplied by its transpose to calculate the covariance matrix. Step-5 :Compute the eigenvalues and eigenvectors The eigenvalues and corresponding eigenvectors are computed for the covariance matrix. here V is the set of eigenvectors associated with the eigenvalues Step-6 :Order eigenvectors Order the eigenvectors according to their corresponding eigenvalues from high to low. Keep only the eigenvectors associated with non-zero eigenvalues. This matrix of eigenvectors is the eigenspace V , where each column of V is an eigenvector. Step-7 : Project training image Each of the centered training images ( ) is projected into the eigenspace. To project an image into the eigenspace, calculate the dot product of the image with each of the ordered eigenvectors. Therefore, the dot product of the image and the first eigenvector will be the first value in the new vector. The new vector of the projected image will contain as many values as eigenvectors. Step-8 : Identify test images Each test image is first mean centered by subtracting the mean image, and is then projected into the same eigenspace defined by V . and The projected test image is compared to every projected training image and the training image that is found to be closest to the test image is used to identify the training image. The images can be compared using any number of similarity measures; the most common is the L 2 norm or euclidean distance. = = = , where = × × =[ | | | ...... | ] = -6 -4 -2 0 2 4 6 8 -8 -6 -4 -2 0 2 4 6 8 10 12 V ariable X 1 V ariab le X 2 PC 1 PC 2 Input Image Face Detectio n Feature Extracti on Face Recogniti on Identificat ion/ Verificatio n Limitations of PCA References × Can not detect non-frontal image. Can not detect if sunglasses or any object that prevents getting face information. Not applicable to noisy image. [1] Belhumeur P., Hespanha J., and Kriegman D. (1997), Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection, IEEE Trans. PAMI, 19(7):711-720. [2]Jim Austin, Thomas Heseltine, Nick Pears and Zezhi Chen, “Face recognition: A comparison of appearance-based approaches”, ACA Group, Deptt. of Computer Science, University of York, 2003. Ω= T Objective × oWorking on principal component analysis approach. oTrying to enhance the capability of PCA for semi frontal face image

Training Database Step 1 : In general approach of PCA, each image is divided into nxn blocks or pixels. Then all pixel values are taken into a single one

Embed Size (px)

Citation preview

Page 1: Training Database Step 1 : In general approach of PCA, each image is divided into nxn blocks or pixels. Then all pixel values are taken into a single one

Training Database

Step 1 : In general approach of PCA, each image is divided into nxn blocks or pixels. Then all pixel values are taken into a single one dimensional matrix. Showed in figure:1

Figure : 1 Figure : 2

Our proposal here is that if we find half image of actual image, we divide each test image and then convert to single column matrix. Afterwards, general PCA will be applied.

Step-2 : Center data The images are mean centered by subtracting the mean image from each image vector.

Step-3 : Create data matrixThese vectors are combined, side-by-side, to create a data matrix of size NxP (where P is the number of images).

An Approach to Face Recognition SystemUsing Principal Component Analysis (PCA)

Fattah Muhammad Tahabi, Ratul Antik Das

Department of Computer Science and Engineering (CSE), BUET

MotivationMotivation

Face recognition has attracted significant attention in the past decades because of its potential applications in biometrics, information security, and law enforcement. Many methods have been suggested to recognize faces.PCA has been widely investigated and has become one of the most popular face-recognition approaches.

Methodology Methodology

𝒙𝒊 = [ 𝒙𝟏𝒊 … 𝒙𝑵𝒊 ]T

𝒙ഥ𝒊 = 𝒙𝒊 − 𝒎 , where 𝒎= 𝟏𝑷 𝒙𝒊𝑷

𝒊=𝟏

𝑿ഥ= ሾ 𝒙ഥ𝟏 ห 𝒙ഥ𝟐 | … 𝒙ഥ𝑷 ]

Ω𝑽= 𝝀𝑽

Step-4 : Calculate covariance matrixThe data matrix X is multiplied by its transpose to calculate the covariance matrix.

Step-5 :Compute the eigenvalues and eigenvectorsThe eigenvalues and corresponding eigenvectors are computed for the covariance matrix.

here V is the set of eigenvectors associated with the eigenvalues

Step-6 :Order eigenvectors Order the eigenvectors according to their corresponding eigenvalues from high to low. Keep only the eigenvectors associated with non-zero eigenvalues. This matrix of eigenvectors is the eigenspace V , where each column of V is an eigenvector.

Step-7 : Project training imageEach of the centered training images ( ) is projected into the eigenspace. To project an image into the eigenspace, calculate the dot product of the image with each of the ordered eigenvectors.

Therefore, the dot product of the image and the first eigenvector will be the first value in the new vector. The new vector of the projected image will contain as many values as eigenvectors.

Step-8 : Identify test imagesEach test image is first mean centered by subtracting the mean image, and is then projected into the same eigenspace defined by V .

and

The projected test image is compared to every projected training image and the training image that is found to be closest to the test image is used to identify the training image. The images can be compared using any number of similarity measures; the most common is the L2 norm or euclidean distance.

𝒚ഥ𝒊 = 𝒚𝒊 − 𝒎 𝒎= 𝟏𝑷 𝒙𝒊𝑷𝒊=𝟏

, where

𝒚 𝒊 = 𝑽𝑻𝒚ഥ𝒊

𝑵𝟐 × 𝟏 𝑵× 𝑵

𝑽= [ 𝐯𝟏 |𝐯𝟐 |𝐯𝟑 | ......|𝐯𝐩 ]

𝒙 𝒊 = 𝑽𝑻𝒙ഥ𝒊

-6

-4

-2

0

2

4

6

8

-8 -6 -4 -2 0 2 4 6 8 10 12

Variable X1

Va

ria

ble

X 2

PC 1

PC 2

InputImage

FaceDetection

FeatureExtraction

Face Recognition

Identification/ Verification

Limitations of PCA Limitations of PCA

References References

𝑵× 𝑵𝟐

Can not detect non-frontal image. Can not detect if sunglasses or any object that prevents getting face information. Not applicable to noisy image.

[1] Belhumeur P., Hespanha J., and Kriegman D. (1997), Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection, IEEE Trans. PAMI, 19(7):711-720.[2]Jim Austin, Thomas Heseltine, Nick Pears and Zezhi Chen, “Face recognition: A comparison of appearance-based approaches”, ACA Group, Deptt. of Computer Science, University of York, 2003.

Ω= 𝑿𝑿തതതത T

Objective Objective

𝝀

𝑵𝟐𝟐 × 𝟏

oWorking on principal component analysis approach.oTrying to enhance the capability of PCA for semi frontal face image

𝒗𝒊 ∈ 𝑽

𝒙ഥ𝒊