Escolar Documentos
Profissional Documentos
Cultura Documentos
Abstract
This paper presents a model-based approach to locate head and face boundaries in a head-shoulder image with plain
background. Three models are constructed for the images, where the head boundary is divided into left/right sub-
boundaries and the face boundary is divided into left/right and top/bottom sub-boundaries. The left/right head
boundaries are located from two thresholded images and the "nal result is the combination of them. After the head
boundary is located, the four face sub-boundaries are located from the grey edge image. The algorithm is carried out
iteratively by detecting low-level edges and then organizing/verifying them using high-level knowledge of the general
shape of a head. The experimental results using a database of 300 images show that this approach is promising.
1999 Pattern Recognition Society. Published by Elsevier Science Ltd. All rights reserved.
Keywords: Face detection; Face boundary models; Shape analysis; Feature extraction
0031-3203/99/$20.00 1999 Pattern Recognition Society. Published by Elsevier Science Ltd. All rights reserved.
PII: S 0 0 3 1 - 3 2 0 3 ( 9 8 ) 0 0 1 6 2 - 9
1318 J. Hu et al. / Pattern Recognition 32 (1999) 1317}1333
detected by the algorithms in [6]. The major limitation 2. Head/face boundary models
of this technique is the dependence on the accuracy of
the eye and mouth detection algorithm and the quality of The head boundary refers to the outermost boundary
edge detection. In [16], Lee et al. applied a knowledge- of the head, while the face boundary refers to the face
based boundary tracking method based on the binary contour excluding the hair, shoulders, and the neck.
edge map, which is obtained using four-directional Sobel According to the hairstyles and wearing of clothes, three
edge operators along with locally adaptive threshold head/face models can be constructed for ID type images.
(LAT). However, the head boundaries may be broken in Fig. 1 shows the models, where (a) corresponds to the
the binary edge images, and therefore reduce the reliabil- typical short hairstyle, (b) corresponds to the typical long
ity of the algorithm. We will discuss the edge detection hairstyle, and (c) corresponds to the situation when the
algorithms in detail in the next section. clothes and the background have a very low contrast
Another major approach to locating head/face bound- (e.g., white clothes against white background).
aries is to use snakes [3,8,14,21]. Snakes are the active In the following, the leftmost point at the topmost row
contour models introduced by Kass et al. [22] which of an image is chosen as the origin of the coordinate
aims at minimizing the energy function of the active system. The frame of the head in the image is located by
contour model. The energy function is the weighted sum horizontal and vertical histogram analysis of the edge
of three components: the internal energy of the contour image. The coordinates of the frame are denoted as y ,
KGL
due to bending or discontinuities, the image force, and x , and x (Fig. 1a).
KGL K?V
the external constraint forces. Huang and Chen [8] used For simplicity, the head boundary is divided into left
the greedy algorithm for the minimal energy process, head boundary and right head boundary. The x coordi-
and Lam and Yan [14] used a fast greedy algorithm to nates of the left and right head boundary points at
improve the computational cost. However, the success scanline i are denoted as h (i) (i"y , 2 , y ) and h (i)
J KGL K?V P
of this approach is heavily dependent on the position of (i"y , 2 , y ) respectively (Fig. 1). To describe the
KGL K?V
the initial head boundary. Furthermore, the snake may face boundary, we use four sub-boundaries: f (i), f (i), f (i),
J R P
be attracted to false edges, such as those caused by and f (i) for left, right, top, and bottom (chin) boundaries
@
shadows, beard and moustache. respectively.
In this paper, we combine traditional edge detection
and thresholding techniques and head/face models to
locate the head and face boundaries. By organizing local 3. Locating head boundary
edge evidence to high level interpretation, this approach
can locate the head and face boundaries e!ectively. In The #owchart of the algorithm for locating head
addition, some more useful information about the hair- boundary is shown in Fig. 2. The image processed here is
styles and the pose of the head can be obtained easily. In front-view ID type image with greylevel in the range
Section 2, we de"ne the head/face boundary models. The [0, 255]. To smooth the image, a 3;3 median "lter can
methods for locating head and face boundaries are de- be applied [23]. The grey edge image is obtained by
scribed in Sections 3 and 4, respectively. Section 5 pres- the templates described in Section 3.1, and the frame of
ents the experimental results, and Section 6 summarizes the face can be located using horizontal and vertical
the paper. histogram analysis. At this stage, we can also normalize
The low threshold ¹ can be determined similarly by From an initial boundary point, we trace along the
J
taking more pixels (45%) as the edge pixels: boundary by analyzing the connectivity of the boundary
pixels.
Denote a left boundary pixel at scanline i as pJ (x , y ).
hist(i)*0.45hg;wd. (3) G G G
The adjacent pixel at scanline i#1 is then
G2F pJ (x , y ), where y "y #1. Pixel pJ (x , y ) is
G> G> G> G> G G G G
If ¹ is very low (¹ )25), the minimum value of ¹ is set connected to pJ (x , y ) if the following conditions
J J J G> G> G>
to 25 to ensure the background noise is excluded. The are true:
thresholded edge images are then smoothed using mor-
phological operations. The smoothing is carried out by pJ (x #k, y )"255 (k"1, 2 , x !x !1)
G G G G> G
an opening followed by a closing, where a 3;3 square
if x *x ; (a)
structure element is used [25]. G> G
Fig. 5 shows some examples, where Fig. 5(1) and (5) is pJ (x #k, y )"255 (k"1, 2 , x !x !1)
the original images, 5(2) and (6) is the grey edge images, G G> G> G G>
Fig. 5(3) and (7) is the high thresholded images, and if x (x . (b)
Fig. 5(4) and (8) are the low thresholded images. For G> G
some images with good contrast and smooth back- The connectivity for right boundary pixels can be
ground, the di!erences between the head boundaries of determined similarly. Fig. 7 shows all the situations,
the high and low thresholded images are small (Fig. 5(3) where (a) and (b) are for left and right boundaries, res-
and (4)). In Fig. 5(7) the left head boundary is broken pectively. According to the connectivity of boundary
while in Fig. 5(8) the edges due to the shadows are pixels, a boundary can be divided into segments where
detected in the right head boundary. the boundary points in each segment are connected.
For the convenience of later processing, the boundary
3.2. Boundary tracing points in one segment are given the same label. In the
tracing procedure, the tracing for a boundary segment is
From the high/low thresholded images, we can trace stopped when the scanning point reaches the border of
the left and right boundaries. One boundary point is the image frame (y and y can then be determined),
K?V K?V
located at each scanline. If there is no background noise or the next boundary point cannot be found. A complete
(e.g., the high thresholded image in Fig. 6(2)), we can boundary segment is the boundary segment where the
simply take the "rst white pixel as the boundary pixel at starting point begins at y and ends at y or y .
KGL K?V K?V
each scanline. Since there may have background noise If the "rst traced segment is a complete boundary seg-
(e.g., Fig. 6(4)), we cannot trace the boundary in this way. ment, the tracing procedure is stopped. Otherwise, all
J. Hu et al. / Pattern Recognition 32 (1999) 1317}1333 1321
Fig. 4. Grey-level edge detection. First column: original images; second column: grey edge image by Sobel operations; third column:
grey edge image by the proposed operations; last column: edge image by Canny edge detection.
the boundary segments are traced. To eliminate the boundary consists of several boundary segments, we "rst
background noise, a boundary segment less than process the gaps between the two boundary segments. If
(hg!y )/20 is discarded. there is a large gap between two boundary segments, we
KGL
Figs. 6(3) and (5) shows the traced boundaries for the can segment a small area around the gap in the grey edge
high and low thresholded images in Fig. 6(2) and (4), image according to Eq. (3). For example, the area
respectively. We can see that the background noise was marked with a square in Fig. 5(7) is segmented again, and
excluded in the traced boundaries in Fig. 6(4). the result is shown in Fig. 5(9). In this case, the boundary
is relocated. If the gap between two segments is small, we
3.3. Boundary analysis draw a line segment to connect the two end points and
locate a boundary pixel nearby the line segment from the
The traced left/right boundary consists of a complete grey edge image (details are described in Step 7 in the
boundary segment or several boundary segments. If the following).
1322 J. Hu et al. / Pattern Recognition 32 (1999) 1317}1333
The located boundary is then smoothed. The bound- To measure the smoothness of a boundary, the "rst
ary data is "rst smoothed by a 3;3 median "lter, and it derivative of the data is used. The "rst derivative d (i) for
is further smoothed by morphological smoothing [25] the left boundary is determined by
where the length of the structure element is set to 7. The
d (i)"h (i)!h (i!1) (i"y #1, 2 , y ) (4)
smoothing scheme can smooth the boundary data e!ec- J J KGL K?V
tively, and it is also used in later boundary smoothing where h (i) is the coordinate of the boundary pixel (Sec-
J
(Fig. 2). After data smoothing, the smoothness analysis of tion 2). Based on the "rst derivative d (i), we can "nd the
a boundary is carried out. critical points of the boundary. Four kinds of critical
J. Hu et al. / Pattern Recognition 32 (1999) 1317}1333 1323
points are de"ned as follows: lem, we de"ned the above four kinds of critical points.
We can see that A1 and A2 points have unique de"ni-
(a) A1 point: a turning point where the boundary goes tions, while B1 and B2 are determined experimentally. In
left and then goes right (Fig. 8a); our experiment, B1 point is located if one of the following
(b) A2 point: a turning point where the boundary goes conditions is satis"ed:
right and then goes left (Fig. 8b);
(c) B1 point: a point where the boundary goes right at (a) d (i)'¹ (large leap);
a leap or at discontinuous (Fig. 8c); (b) d (i)'¹ and label(i)"label(i!1) (small leap but
(d) B2 point: a point where the boundary goes left at discontinuous).
a leap or at discontinuous (Fig. 8d).
B2 point can be determined in a similar way. ¹ and
The critical point detection algorithm is crucial in the ¹ can be determined experimentally according to the
algorithm. There are many methods such as the algo- size of the image frame. We set ¹ "8 and ¹ "3 in our
rithms in [28,29] for this purpose. To simplify the prob- experiments. After the critical boundary points are
1324 J. Hu et al. / Pattern Recognition 32 (1999) 1317}1333
3.4. Boundary combination we will only describe the algorithm for locating left face
boundary. The bottom face boundary, however, is more
The traced boundaries from the high thresholded im- di$cult to locate. In our approach, we use a model-based
age are more robust to shadows and background noise method to approximate the bottom (chin) boundary.
than that from the low thresholded image. However,
more boundary pixels may be missing in the high thre- 4.1. Left face boundary
sholded image. On the contrary, boundaries in the low
thresholded image tend to be smoother than that in the The procedure for locating the left face boundary is
high thresholded image, but edges formed by the shown in Fig. 10. The working images are the original
shadows may be detected in the head boundary. There- image I(x, y) and the grey edge image I (x, y). In order to
fore, we need to combine the two results to obtain a C
"nd the border pixel between the hair and face, the
smoother and reliable boundary. average intensity of the hair nearby the left boundary is
Denote the x coordinates of the left boundaries of the computed. The initial average hair intensity g is esti-
high and low thresholded images as h (i) and h (i) ?
J J mated by
(i"y , 2 , y ), respectively. Obviously, h (i)*
KGL K?V J
h (i). The combination result h (i) of h (i) and h (i) is W >
J J J J g " KGL g(h (i)#j, i)/100, (6)
determined as follows: ? J
GWKGL> H
(a) If the left boundary for the high thresholded image is
complete and smooth, then it is taken as the "nal left
head boundary: h (i)"h ;
J J
(b) If the left boundary for the high thresholded image is
not complete or smooth, but if the left boundary for
the low thresholded image is complete and smooth,
then we have h (i)"h ;
J J
(c) In all other cases, the "nal left boundary is taken as
the left boundary for the low thresholded image and
the boundary is analyzed and repaired according to
the procedure described in Section 3.3.
Fig. 11. Face boundary location: (1) initial left face boundary; (2) the "nal left face boundary after smoothing between the starting and
end points; (3) "nal results of the left/right/top face boundaries; (4) another example of the face boundaries; (5)}(8) illustration of the
bottom face boundary location.
where g(x, y) is the intensity of pixel I(x, y) and h (i) is the The next step is to "nd a face border pixel at each
J
left head boundary. In case of the hair area being very scanline. The searching ranges are [h (i)#1, w/3] for
J
small, g is re"ned by excluding very high and very low y (i)(y !y )/2 and [h (i)#1, w/2] for
? KGL K?V KGL J
intensity pixels ("I(x, y)!g "'80) in this area. (y !y )/2(i(y , where w is the image width
? K?V KGL K?V
J. Hu et al. / Pattern Recognition 32 (1999) 1317}1333 1327
and y and y are the starting and end points of the the skin of the face), the face boundary is the same as the
KGL K?V
left head boundary (Fig. 1). Let a"g( j#3, i)# head boundary.
g( j#4, i)#g( j#5, i), b"g( j!3, i)#g( j!4, i)#
g( j!5, i) and c""a!b" at scanline i. The border pixel 4.2. Bottom face boundary
I( j, i) is the point which has the maximum c value in the
searching range and satis"es the following conditions: The bottom face boundary is the border of the chin.
Since the beard and the shadow below the chin can
(a) c'40 and "b!g "(60, if g ( j, i)*200; produce &&wrong'' edges, and since the contrast at the chin
? C
may be very low (very weak edges), we cannot rely on the
(b) c'30 and "b!g "(50, if 60(g ( j, i)(200;
? C result of edge detection only (refer to Section 4 in [7] ). In
our approach, we use a segment of ellipse to approximate
where g is the average intensity of the hair and g (x, y) is
? C the chin shape. The initial position is determined accord-
the intensity of the grey edge pixel I (x, y).
C ing to the head and face boundaries. Edges are then
Fig. 11(1) shows the detected left face border pixels of
detected near the segment of ellipse, and the position of
a face image. If a border pixel is found at ( j, i), then we
the ellipse in turn is adjusted according to the detected
have f (i)"j. If no border pixel is found, we set f (i)"0.
J J edges.
The face border data is then smoothed using the same
The ellipse models are shown in Fig. 12, where (b) and
smoothing procedure described in Section 3.3. The initial
(c) are the rotated cases of (a). To draw the ellipse arc, we
starting point s of the face border is the "rst point from
need to know the lengths of the two axes of the ellipse
y which satis"es
KGL and the two points p (x , y ) and p (x , y ). If the min-
(a) f (s)'0, f (s#1)'0, f (s#2)'0, and imum y coordinate of the top face boundary is y , then
J J J
the length of the long axis of the ellipse is approximated
(b) " f (s)!f (s#1) "(4, " f (s#2)!f (s#1)"(4. by a"(y #y )/2!y . Denoting the minimum x coor-
J J J J
dinate of the left face segment above y as x and the
Similarly, we can search back from y to "nd the J
K?V maximum x coordinate of the right face segment above
initial end point t. For the face border between s and t, it y as x , the length of the short axis can be estimated as
is divided into several segments by detecting any two P
b"x !x . Therefore, the main problem is to deter-
adjacent points f (i) and f (i#1) which satisfy P J
J J mine the coordinates of p and p .
" f (i)!f (i#1)"(4. The points in a segment are labeled
J J Since the head boundary and the left and right face
with the same symbol, which is convenient for later boundaries have been located, we can determine the type
smoothing. Denoting the starting and end points of the of the face according to Fig. 1. First, shape analysis is
previous segment, the current segment and the next seg- carried out for the head boundary. The shape of the
ment as (s , t ), (s , t ), and (s , t ) respectively, the cur- boundary is characterized by its convex/concave cha-
rent segment can be repaired (it is substituted by the nges. We "nd out all the feature points as shown in
straight line connecting t and s ) if the following condi- Figs. 8a and b, and analyze the position of each feature
tions are true: point as well as their relative positions. For the "rst and
the third types of head boundaries in Fig. 1, the feature
(a) f (t )!f (s )'3 and f (s )!f (t )'3; or
J J J J points p and p are determined at one of the head
f (s )!f (t )'3 and f (t )!f (s )'3; boundaries. The left and right face boundaries are then
J J J J extended to join the feature points p and p . Fig. 11(5)
(b) t !s 't !s and t !s 't !s . shows the located feature points p and p and the initial
ellipse arc. For the second type in Fig. 1 (long hair), the
After repairing, we "nd the starting point s and the end feature points p and p are determined from the face
point t of the left face boundary again. The starting point boundaries only. Fig. 11(7) shows the located feature
s is taken as the "rst point of the "rst segment which are points p and p and the initial ellipse arc.
more than 4 pixels in length. From the starting point s, From the initial ellipse arc, we search for the edge
we "nd the end point t when a zero value ( f (i)"0) is met pixels of the chin. If the y coordinate of the ellipse at
J
or very short scattered segments are met. Fig. 11(2) shows column i is j, we search for the strongest edge pixel in the
the smoothed left face border between the "nal starting range [ j!10, j#10] at column i. Fig.s 11(6) and (8)
point and end point. show the detected edge pixels corresponding to Fig. 11(5)
Similar procedure can be applied to locate the right and 11(7), respectively, where the initial ellipse arcs are
face boundary f (i) and the top face boundary f (i). also drawn in the images. We then calculate the average
P R
Fig. 11(3) shows the result for the face image in Fig. 11(1) di!erence d (in y direction) of the initial ellipse arc and
while Fig. 11(4) shows the result for another face image. the detected edges. The "nal bottom (chin) boundary is
If no smooth face boundary segment can be found (no the initial ellipse adjusted by d in y direction (Fig. 13(1)
hair is present, or the color of the hair is very similar to and (10)).
1328 J. Hu et al. / Pattern Recognition 32 (1999) 1317}1333
5. Experiments
Fig. 13. The head and face boundaries located by the proposed algorithms for front view images in the database.
eyes and mouth features are located directly (see Section 6. Discussion and conclusion
1). The bottom face boundary (chin) detection is the most
di$cult problem in face boundary detection since the Locating head and face boundaries is an important
edge information is generally not su$cient (see also Sec- task in automatic face recognition. Many approaches are
tion 4 in [7]). In our current study, we deal with ID type based on &&global'' methods, for example, snakes and
front view images. More research works need to be car- Hough transform. It is believed that it is generally very
ried out for other situations (as shown in Fig. 14). di$cult to organise low-level (edge) information into a
1330 J. Hu et al. / Pattern Recognition 32 (1999) 1317}1333
sensible contour for an object. However, for face recogni- by histogram analysis. Therefore, the face image can be
tion, we have a priori knowledge of the general shape assumed at the center of the frame approximately. Thus,
of a head, which has been classi"ed into three models in the searching area for head and face boundary can be
our experiments. For images with plain background, the speci"ed. The division of head/face boundaries into sev-
frame of the head-shoulder image can easily be located eral sub-boundaries has the advantage of easy handling.
J. Hu et al. / Pattern Recognition 32 (1999) 1317}1333 1331
Table 1
Evaluation of the head and face boundary location by the proposed algorithms
Image Boundary Head Left face Right face Top face Bottom face
G 34 39 39 37 31
M 5 1 1 3 2
1,2 B 1 0 0 0 0
F 0 0 0 0 7
G 30 33 35 34 8
M 8 6 5 6 5
3,4 B 2 1 0 0 1
F 0 0 0 0 26
G 35 38 37 38 11
M 5 2 2 1 6
5,6 B 0 0 1 1 1
F 0 0 0 0 22
G 33 37 36 36 10
M 7 2 2 3 7
7,8 B 1 1 2 1 3
F 0 0 0 0 20
G 34 38 38 35 18
M 5 2 2 4 3
9,10 B 1 0 0 1 6
F 0 0 0 0 13
G 165 (82.5%) 185 (92.5%) 185 (92.5%) 180 (90%) 78 (39%)
Total M 30 (15%) 13 (6.5%) 12 (6%) 17 (8.5%) 23 (11.5%)
B 5 (2.5%) 2 (1%) 3 (1.5%) 3 (1.5%) 11 (5.5%)
F 0 0 0 0 88 (44%)
The boundary smoothing and boundary shape analy- sis. Thus, the searching area for head boundary can be
sis are crucial in the algorithm. With boundary smooth- speci"ed. Second, a grey edge image is obtained by ex-
ing, the redundant critical feature points in the boundary tended Sobel operations. From the grey edge image,
can be reduced signi"cantly. The algorithm is carried out two thresholded images are obtained. The high-thre-
iteratively by detecting low-level edges and then organis- sholded image is used to obtain strong edges while ex-
ing/verifying them using high-level knowledge of the gen- cluding the edges caused by shadows, and the low-thre-
eral shape of a head. The experimental results show that sholded image is used to provide more edges when the
this approach is promising. contrast of the image is low. The head boundary is
divided into left/right sub-boundaries where each of the
initial sub-boundary can be obtained from the edges
7. Summary directly. The boundary is then smoothed and analyzed
against the three models, where a broken boundary can
Locating head and face boundaries is an important be detected and repaired. The "nal result is the combina-
task in model based image coding and automatic face tion of the boundaries from the two thresholded images.
recognition. Many approaches are based on global The algorithm is carried out iteratively by detecting low-
methods such as snakes and the Hough transform. In level edges and then organizing/verifying them using
general, it is very di$cult to integrate low-level (edge) high-level knowledge of the general shape of a head. The
information into a contour of an object. However, for face boundary is located, the four face sub-boundaries
face recognition, we can make use of a priori knowledge are located from the insides of the head boundaries. The
of the general shape of a head, which has been classi"ed division of head/face boundaries into several sub-bound-
into three models in our experiments. For images with aries has the advantage of e$cient processing. The
plain background, we could detect the head and face boundary smoothing and boundary shape analyses are
edges and organize them into meaningful boundaries crucial steps in this algorithm. The experimental results
based on the three models. First, the frame of the head- on a database of 300 images show that this approach is
shoulder image can easily be located by histogram analy- promising.
1332 J. Hu et al. / Pattern Recognition 32 (1999) 1317}1333
Fig. 14. Examples of the results when the proposed algorithms were applied to other images in the database.
References [3] J.B. Waite, W.J. Welsh, Head boundary location using
snakes, Br. Telecom. Technol. J. 8 (3) (1990) 127}135.
[1] A. Samal, P.A. Iyengar, Automatic recognition and analy- [4] M. Turk, A. Pentland, Eigenfaces for recognition, J Cogni-
sis of human faces and facial expressions: a survey, Pattern tive Neurosci. 3 (1) (1991) 71}86.
Recognition 25 (1) (1992) 65}77. [5] A.L. Yuille, Deformable templates for face recognition,
[2] R. Chellappa, C.L. Wilson, S. Sirohey, Human and ma- J. Cognitive Neuroscience 3 (1) (1991) 59}70.
chine recognition of faces: a survey, Proc. IEEE 83 (5) [6] G. Chow, X. Li, Towards a system for automatic facial fea-
(1995) 705}740. ture detection, Pattern Recognition 26 (12) (1993) 1739}1755.
J. Hu et al. / Pattern Recognition 32 (1999) 1317}1333 1333
[7] X. Li, N. Roeder, Face contour extraction from front-view [18] G. Yang, T.S. Huang, Human face detection in a complex
images, Pattern Recognition 28 (8) (1995) 1167}1179. background, Pattern Recognition 27 (1994) 53}63.
[8] C.L. Huang, C.W. Chen, Human facial feature extraction [19] M. Kosugi, Human-face search and location in a scene by
for face interpretation and recognition, Pattern Recogni- multi-pyramid architecture for personal identi"cation,
tion 25 (12) (1992) 1435}1444. Systems Comput. Japan 26 (6) (1995) 27}38.
[9] R. Brunelli, T. Poggio, Face recognition: features versus [20] S.-H. Lin, S.-Y. Kung, L.-J. Lin, Face recognition/detec-
templates, IEEE Trans. Pattern Anal. Mach. Intell. 15 (10) tion by probabilistic decision-based neural network, IEEE
(1993) 1042}1052. Trans. Neural Networks 8 (1) (1997) 114}132.
[10] A. Pentland, B. Moghaddam, T. Starner, View-based and [21] C.W. Chen, C.L. Huang, Human face recognition from
modular eigenspaces for face recognition, in: Computer a single front view, Int. J. Pattern Recognition and Arti"-
Vision and Pattern Recognition Conf., IEEE Computer cial Intell. 6 (4) (1992) 571}593.
Society, 1994, pp. 84}91. [22] M. Kass, A. Witkin, D. Terzopoulos, Snakes: active
[11] H.P. Graf, T. Chen, E. Petajan, Eric Cosatto, Locating faces contour models, Int. J. Computer Vision, 1988, pp.
and facial parts, Int. Workshop on Automatic Face- and 321}331.
Gesture-Recognition, Zurich, Switzerland, 1995, pp. 41}46. [23] T.S. Huang, G.Y. Yang, G.Y. Tang, A fast two-dimen-
[12] N. Intrator, D. Reisfeld, Y. Yeshurun, Extraction of facial sional median "ltering algorithm, IEEE Trans. Accoust.
features for recognition using neural networks, Int. Work- Speech Signal Process. 27 (1) (1979) 13}18.
shop on Automatic Face- and Gesture-Recognition, [24] A.K. Jain, Fundamentals of Digital Image Processing,
Zurich, Switzerland, 1995, pp. 260}265. Prentice Hall, Englewood Cli!s, NJ, 1989.
[13] S.-H. Jeng, H.-Y.M. Liao, Y.-T. Liu, M.-Y. Chern, An [25] R.C. Gonzalez, R.E. Woods, Digital Image Processing,
e$cient approach for facial feature detection using geo- Addison-Wesley, Reading, MA, 1992.
metrical face model, Proc. ICPR'96, Vienna, Austria, [26] M. Sonka, V. Hlavac, R. Boyle, Image Process., Analysis
Track C, 1996, pp. 426}430. Mach. Vision. Chapman & Hall Computing, Cambridge,
[14] K.-M. Lam, H. Yan, Fast algorithm for locating head 1993.
boundaries, J. Electronic Imaging 3 (4) (1994) 351}359. [27] M. Heath, S. Sarkar, T. Sanocki, K. Bowyer, Comparison of
[15] A. Jacquin, A. Eleftheriadis, Automatic location tracking edge detectors: a methodology and initial study, Proc. Com-
of faces and facial features in video sequences, Int. Work- puter Vision and Pattern Recognition 1996, pp. 143}148.
shop on Automatic Face- and Gesture-Recognition, [28] M.A. Fischler, R.C. Bolles, Perceptual organization and
Zurich, Switzerland, 1995, pp. 142}147. curve partitioning, IEEE Trans. Pattern Anal. Mach. In-
[16] S.Y. Lee, Y.K. Ham, R.-H. Park, Recognition of human tell. 8 (1986) 100}105.
front faces using knowledge-based feature extraction and [29] J. Hu, H. Yan, Polygonal approximation of digital curves
neuro-fuzzy algorithm, Pattern Recognition 29 (11) (1996) based on the principals of perceptual organization, Pattern
1863}1876. Recognition 30 (5) (1997) 701}718.
[17] V. Govindaraju, D.B. Sher, R.K. Srihari, S.N. Srihari, [30] P.J. Phillips, H. Moon, P. Rauss, S.A. Rizvi, The FERET
Locating human faces in newspaper photographs, Proc. evaluation methodology for face-recognition algorithms,
IEEE-CS Conf. of Computer Vision and Pattern Recogni- Proc. Computer Vision and Pattern Recognition, (1997)
tion, San Diego, CA. 1989, pp. 549}554. 137}143.
About the Author*JIANMING HU received his MSE degree in Communication and Electronic Systems from Xidian University in
1990, and his Ph.D. degree in Electrical Engineering from the University of Sydney in 1997. He is now a senior research assistant in the
Department of Electrical Engineering, University of Sydney. His research interests are in the areas of information theory, coding theory,
image processing and pattern recognition.
About the Author*HONG YAN received his B.E. degree from Nanking Institute of Posts and Telecommunications in 1982, M.S.E.
degree from the University of Michigan in 1984, and Ph.D. degree from Yale University in 1989, all in electrical engineering. From 1986
to 1989 he was a research scientist at General Network Corporation, New Haven, CT, USA, where he worked on developing a CAD
system for optimizing telecommunication systems. Since 1989 he has been with the University of Sydney where he is currently
a Professor in Electrical Engineering. His research interests include medical imaging, signal and image processing, neural networks and
pattern recognition. He is an author or co-author of one book, and more than 200 technical papers in these areas. Dr. Yan is a fellow of
the Institution of Engineers, Australia (IEAust), a senior member of the IEEE, and a member of the SPIE, the International Neural
Network Society, the Pattern Recognition Society, and the International Society for Magnetic Resonance in Medicine.
About the Author*MUSTAFA SAKALLI received his BSc degree from the Faculty of Electrical Engineering, Istanbul Technical
University, in 1980 and MSc degree from the Institute of Biomedical Engineering, Bogazici University, in 1988. He is currently
a research assistant and is working toward a Ph.D. degree in the Image Processing Research Group, Department of Electrical
Engineering, University of Sydney.