Você está na página 1de 9

Asian Face Image Database PF01

Intelligent Multimedia Lab.


Department of Computer Science and Engineering Pohang University of Science and Technology San 31, Hyoja-Dong, Nam-Gu, Pohang, 790-784, Korea

Abstract We constructed a face database PF01(Postech Faces 01). PF01 contains the true-color face images of 103 people, 53 men and 50 women, representing 17 various images (1 normal face, 4 illumination variations, 8 pose variations, 4 expression variations) per person. All of the people in the database are Asians. There are three kinds of systematic variations, such as illumination, pose, and expression variations in the database. The database is expected to be used to evaluate the technology of face recognition for Asian people or for people with systematic variations. Keywords: Face database, Image database, Face recognition

1. Introduction

Face recognition is an active research area spanning several research elds, such as image processing, pattern recognition, computer vision, and neural network [1]. Face recognition has many applications, such as biometrics system, surveillance, and content-based video processing. Today, face recognition technology is becoming more important. For the development of face recognition technology, we need the face image database to evaluate the performance of face recognition technology. Many face databases have been reported. Face databases can be categorized according to their purposes. Some databases can be used for face recognition technology, other for face expression recognition or for pose estimation. In this work, we concentrate on face databases for face recognition. The representative databases are FERET database[2], XM2VTS database[3], Yale DB[4], AR Face database[5], MIT face database[6] and ORL database[7]. We explain the representative databases briey. FERET database[2] The FERET database was constructed to develop automatic face recognition capabilities that can be employed to assist security,
Technical Report

intelligence and law enforcement personnel, sponsored by the Department of Defenses Conterdrug Technology Development Program through the Defense Advanced Research Products Agency. It has 14051 eight-bit grayscale images of human faces. It includes face images of various poses, including proles of alternative expressions and of dierent illuminations. For some people, it includes face images with eye glasses worn, with dierent hair length, and both. Pose variations are systematic and various, but other variations are not. Nevertheless, since it includes many face images of many people, it is one of the best-known face databases. XM2VTS database[3] The XM2VTS database is a large multi-modal database to test multi-modal face verication algorithms. It was made by the Center for Vision, Speech and Signal Processing, at the University of Surrey, England. It contains four recordings of 295 people over four months. Each recording contains a speaking head shot and a rotating head shot. The shots were stored in high quality color images, 23KHz 16-bit sound les of speech, video sequences of rotating heads, and a 3D model of a face. Yale DB[4] The Yale face database was made by the Center for Computational Vision and Control, at Yale University. It contains gray face images of 15 people, where there are images of 11 variations for each person. Images for each person are normal images, images with glasses and without glasses, images with light variations (such as center-light, left-light and right-light), and images with expression variations (such as happy, sad, sleepy, surprised and winking). AR face database[5] AR face database was made by Computer Vision Center at UAB(Universitat Autonoma de Barcelona). It contains the color images of 126 people (70 men and 56 women), where there are frontal faces with dierent facial expressions, illumination conditions, and occlusions (sun glasses and scarf). The pictures were taken at the CVC under strictly controlled conditions. No restrictions on wear (clothes, glasses, etc.), make-up, hair style, etc. were imposed on participants. Each person participated in two sessions, separated by two weeks (14 days). The same pictures were taken in both sessions. MIT face database [6] The MIT face database was made by the MIT Media Laboratory. It contains the face images of 16 male people. It includes images of 3 pose variations (upright, right, left), 3 light variations (head-on, 45 deg, 90 deg) and 3 scale (camera zoom) variations (full, medium, small). It includes 6 levels of gaussian pyramidals (480x512, 240x256, 120x128, 60x64, 30x32, 15x16). ORL database [7] The ORL database was made by AT&T laboratories at Cambridge. It contains the face images of 40 people,where there are ten dierent images for each person. For some subjects, the images were taken at dierent times, the lighting variations, facial expressions (open / closed eyes, smiling / not smiling) and facial details (glasses / no glasses). All the images were taken against a dark homogeneous background with the subjects in an upright, frontal position (with tolerance for some side movement). 2

There are two drawbacks in the existing face databases including those mentioned above. The rst drawback is that most databases do not contain many Asian people. Some database include only a limited number of Asian people. The second drawback is that there is no systematic and various changes about variations. Some databases have systematic changes about some variations, but not for other variations. In even some database having systematic changes, the range of variations is limited. To overcome these drawbacks, we have built a new database called PF01 (Postech Faces 01). It contains face images of Asian people, mostly Korean people. There are face variations for light, pose and expression. Variations are systematic and the amount of variations is broad enough. The database PF01 is expected to be used to evaluate face recognition technology for Asian people and the systematic variation-invariant recognition. This work is organized as follows. Section II describes the process of capturing the face images. Section III explains the detailed prole of the database. Section IV discusses construction the process of the database and problems we encountered in the process. Finally, a conclusion is drawn.

2. Collection of Data 2. .1 Face Variations

The face image database should include many variations, since it is used to evaluate the performance of face recognition technologies appropriately. The sources of variability in facial appearance can be either intrinsic or extrinsic[8]. Intrinsic variables for face variations are facial expression, clothing, facial occlusion, hair, and so on. Extrinsic variables are alignment, background, camera adjustment, illumination, imaging device, pose, scale, and so on. In the existing face image databases, there are many variations, including illumination, pose, expression, hair, glasses, earrings, etc. Some databases include many kinds of variations, but other databases do not. Our database includes one intrinsic variation of expression and two extrinsic variations of illumination and pose. The variations in our database are systematic. The rst row of Figure 1 shows a normal face. The second row of Figure 1 shows 4 illumination variations which correspond to up, down, left and right illumination from the left to right. The third row of Figure 1 shows dierent facial expressions which correspond to happy, surprised, irritated and closedeye facial expression from the left to right. The last two rows of 1 shows eight dierent poses which correspond to the direction of 45 degree increment, 3

counterclockwise, starting from the upper 90 degree direction from the top-left to bottom-right. We do not consider variations of wearing glasses, but there are people with or without glasses in the database.

Fig. 1. Face Examples

2. .2 Gathering Volunteers

We gathered volunteers from students and stas on the POSTECH campus. The nature of people in the database can be considered in terms of intrinsic natures, such as ethnicity, sex, and age [8]. All of the volunteers are Asian people. One is a Chinese, one is a Vietnamese, one is a Bangladeshi, and the remaining people are Korean. People in the database consist of 53 men and 50 women. Age distributions for men and women are plotted in Figure 2 and Figure 3, respectively. Distributions for men and women with glasses or without glasses are shown in Table 1. 4

10

20

30

40

50

60

Fig. 2. Age distributions of men


14

12

10

10

20

30

40

50

60

Fig. 3. Age distributions of women Table 1 Distributions for men and women with glasses or without glasses Men With glasses Without glasses 23 30 Women With glasses Without glasses 8 42

2. .3 Capturing Environment

We describe the capturing environment. First, we explain the camera conguration. The camera is a digital camera, the Sony DSC-F505V. Exposure is controlled automatically. The position of the camera is 113-123cm up from the ground and 105cm away from the back of a chair for a volunteer. The height of the camera is changeable according to the position of a volunteers face. The chair on which a volunteer sits is 46cm high. 5

Second, we explain the illumination conguration. We occluded the windows in a small studio room by dark papers. We used the uorescent light lamps partially occluded by dark papers in the ceiling for an ambient light eect. As a directional light eect for obtaining faces of pose variations, we used 4,500W-max-Halogen lamps. One lamp was positioned on the ground in front of a chair. Two lamps were positioned 165cm left and right from a chair and 110cm high from the center of the lamp positioned on the ground. The last lamp was positioned on the ceiling, 254cm up from the center of the lamp positioned on the ground. For the detailed conguration, please refer to Figure 4. Third, we explain the conguration for pose variations. We asked a volunteer to move his head by gazing at a mark. We placed 8 marks on the 30cm-radius circle around the camera.

Fig. 4. Camera Conguration

3. Prole of the Database

The database includes 1751 images because there are 53 men and 50 women with 17 images per each person. Each image is recorded in JPEG format. The image is of size 1280x960 and of true color. We append the information for normalization, which consists of 5 (left, right, up, down, central) points for a left eye, 5 points for a right eye, 2 points (of nostrils) for a nose, and 4 (left, right, up, down) points for mouth. Figure 5 illustrates 16 points used for normalization. We convert true-color images into the gray color images, normalize them using the central points of eyes, and append the processed 6

images. The normalization process requires that each image be 256 x 256 pixels in size. The centre of the right eye should be located at pixel 97 on row 100 and the centre of the left eye should be located at pixel 160 of row 100. Figure 6 shows one typical example of the normalized images.

Fig. 5. 16 points used for normalization

4. Discussions

We encountered three problems in building the database. First, it was dicult obtain volunteers. Some people are reluctant to have their face image publicly used, so we mainly used volunteers who are well known to us. So the current face image database has a limited number of people. Second, it is dicult to control the capturing environment. Our database has three kinds of variations. For illumination variations, in the case of down illumination, there are variations according to sitting poses. For pose variations, there are small dierences among people because we obtain pose variations by making volunteers look at marks for pose variations. For expression variations, it is dicult to make some expressions, such as an irritated look. Because of these diculties, some images have inappropriate variations, and we discarded them. To control these variations more exactly, we need an active sensor that measures the 3D pose direction accurately. 7

Fig. 6. one typical example of normalized face images

Third, it is dicult to extract information for normalization. As we mentioned above, we used 16 points per one image for normalization. It requires much eort to mark their points accurately. Although we nished making the information, there is another problem. Relative positions of these feature points such as eyes, a nose, and a mouth are dierent among people. So this approach may not be appropriate for normalization. We need further research to discover a more appropriate normalization technique.

5. Conclusion

We have built the face database PF01. The database contains face images of 103 people having 17 images per each person. All people in the database are Asian extraction. Images in the database are normal images and images of systematic variations for illumination, pose and expression. Yet, there is no well-known database for Asian people or with such variations like PF01. We 8

expect our database to be widely used for the evaluation of the technology of face recognition for Asian people. We have described the process of building the database PF01 in detail, and discussed the problem we encountered in the process. These will help researchers to construct a face database.

Acknowledgements

The authors would like to thank the Ministry of Education of Korea for its nancial support toward the Electrical and Computer Engineering Division at POSTECH through its BK21 program. This research was also supported by the Brain Science and Engineering Research Program of the Ministry of Science and Technology of Korea, and by the Basic Research Institute Support Program of Korea Research Foundation.

References
[1] R. Chellappa, C. L. Wilson, S. Sirohey, 1995, Human and Machine Recognition of Faces: A Survey, Proceedings of the IEEE, 83(5), pp. 705-740. [2] P. J. Phillips, H. Moon, P. J. Rauss, and S. Rizvi, 2000, The FERET evaluation methodology for face recognition algorithms, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 22, No. 10. [3] J. Matas, M. Hamouz, K. Jonsson, J. Kittler, Y. Li, C. Kotropoulos, A. Tefas, I. Pitas, T. Tan, H. Yan, F. Smeraldi, J. Bigun, N. Capdevielle, W. Gerstner, S. Ben-Yacoub, Y. Abdeljaoued , E. Mayoraz, 2000, Comparison of face verication results on the XM2VTS database, Proceedings of the 15th International Conference on Pattern Recognition, Barcelona (Spain), vol 4, September, 858-863. [4] P. Belhumeur, J. Hespanha, D. Kriegman, 1997. Eigenfaces vs.Fisherfaces: class specic linear projection, IEEE Transactions on PAMI, 19(7), 711-720. [5] A.M. Martinez, A. C. Kak, 2001, PCA versus LDA, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol 23. No. 2. [6] M. Turk and A. Pentland, 1991, Eigenfaces for Recognition, J. Cognitive Neuroscience, 3(1). [7] F. Samaria, A. Harter, 1994, Parameterisation of a Stochastic Model for Human Face Identication, Proceedings of 2nd IEEE Workshop on Applications of Computer Vision, Sarasota FL. [8] S. Gong, S. J. McKenna, A. Psarrou, 2000, Dynamic Vision : From Images to Face Recognition, ImperCooege Press.

Você também pode gostar