Escolar Documentos
Profissional Documentos
Cultura Documentos
tan
tan
(2.2)
The data presented (in Table 2.1 and 2.2) shows that the measuring capa-
bilities of the available sensors are in the same order of magnitude as variations
in the anthropometric measures and in some cases are even ner. This compar-
ison suggests that the proposed idea of using these measures for identication
warrants further investigation.
13
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
2.3 Concept Evaluation of Head-to-Shoulder
Shape for Person Identication
Given the capabilities of the available sensors (relative to variations in anthropo-
metric dimensions) a proof of concept study was undertaken in order to further
investigate the possibility of person identication using physical dimensions of
the human head-to-shoulder region. Firstly, a dataset was gathered using a Swis-
sranger 4000 attached to the head of the RobotAssist platform. The robot was
placed in a xed position in an empty room with a wall as a backdrop (Fig.2.1).
A total of 25 individuals (the number of students available within CAS-UTS)
were separately recorded standing in front of the robot, at a distance of 2 metres,
facing towards it.
FEATURE
EXTRACTION ZONE
STAND
HERE
STAND
HERE
Figure 2.1: Data collection setup for proof of concept study: Robot is setup facing
a wall and participants stand facing the robot one at a time as 3D pointcloud data
is gathered.
From the pointcloud data gathered the physical dimensions of the head and
shoulders of each person were measured. An algorithm was developed to extract
the lateral span of the head and the shoulders of each person. The box plots in
Fig. 2.2 show the results obtained for the head (Fig. 2.2a) and shoulder spans
(Fig. 2.2b) where the red line marks the median measure for each individual and
the box represents the 25th - 75th percentile range. It is clear from the gure
that there is a separation between the median result of each persons head span
14
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
and even more so in their shoulder span, suggesting that this technique could
potentially be used for discriminating between people. The plots also show that
spread of results for each person was narrow in comparison to the separation from
other people, in many cases completely separating the middle 50% from other
people in the test and in some cases separating the entire range from other people
in the test. This comparison suggests that not only was the variation between
individuals signicant but that it could be measured using the SwissRanger 4000.
15
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
0.15
0.2
0.25
0.3
0.35
0.4
0.45
0.5
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25
(a) Box plot of head spans.
0.45
0.5
0.55
0.6
0.65
0.7
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25
(b) Box plot of shoulder spans.
Figure 2.2: Box plots of head and shoulder spans showing signigant separation
between individuals relative to the spread of data within each individual. This
separation suggests that these features may be descriptive enough to facilitate
identication.
16
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
Chapter 3
A Method of Person
Identication Using
Head-to-Shoulder Signatures
Based on encouraging results of preliminary investigations into using head and
shoulder shape for person identication, a system was developed to exploit this
idea. In order to identify people the system needed to detect people from a scene,
examine them individually, and make some decision as to their identity. The
identication system developed in this capstone is divided into three main stages:
person detection, feature extraction and person identication as shown in Fig. 3.1.
The input to the system is a 3D pointcloud from any type of 3D depth camera.
The pointcloud is passed to the person detection stage which analyses the scene
by segmenting it into smaller pointclouds for each of the regions of interest (ROI)
potentially representing people, so they can be examined individually. The feature
extraction stage then takes these pointcloud segments and extracts the HSS from
each of them to be considered by the person identication stage. Finally the
person identication stage determines the identity of each person by comparing
the observed HSS with a stored model of known people. These steps are explained
in more detail below.
17
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
!
!
P
E
R
S
O
N
D
E
T
E
C
T
I
O
N
F
E
A
T
U
R
E
E
X
T
R
A
C
T
I
O
N
P
E
R
S
O
N
I
D
E
N
T
I
F
I
C
A
T
I
O
N
SCENE FLATTENING
IDENTIFICATION TRAINING
Density Image!
Known
Person
HSS!
Unknown
Person
HSS!
Identity Label!
BLOB DETECTION
SIZE CHECKING
HEAD-TO-SHOULDER SLICING
CALCULATE ROTATIONALY INVARIANT SPANS
VERIFY DETECTED PEOPLE HUMAN
HSS
SORT KNOWN PEOPLE
KNOWN
PEOPLE
HSS
Blobs!
Labeled
HSS!
Labeled
HSS!
SEGMENT POINTCLOUD
ROI!
3D Pointcloud!
3D Pointcloud!
HSS!
HSS!
Pointcloud Segments!
HSS!
Human Only HSS!
3D DEPTH CAMERA
!
!"!"!"
N!
Head-to-Shoulder Signature!
Slice Pointclouds!
SUPERVISED
TRAINING
COMUNICATION
WITH USERS
Identity
Label!
WORLD
MODEL
!"!"!"
N!
Rotationally
Invariant Spans!
Figure 3.1: System Schematic for HSS Person Identication showing the three
main parts of the system: person detection, feature extraction and person
identication.
18
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
3.1 Scene Analysis for Person Detection
To enable detection of people within a scene, a method of scene analysis based
on (Hordern and Kirchner, 2010) was used. First the pointcloud is projected
onto a 2D horizontal plane by taking a bivariate histogram in the two horizontal
axes. This process results in a density image where the intensity of each pixel
represents the concentration of points measured in that area of the scene. This
image resembles an aerial view of the scene where clusters of points typically
represent vertical surfaces. An example of this density image is shown in Fig.
3.2a.
This approach to scene analysis exploits the assumption that people appear in
a scene as vertical surface with a relatively high point-density in the histogram,
when compared with the many horizontal surfaces (oor, tables, chairs) in a
scene. In order to detect these vertical surfaces a threshold is applied to the
density image to remove low density areas such as horizontal surfaces, resulting
in a binary image where only signicant point clusters were represented, as seen
in Fig. 3.2b.
Blob detection is performed on the binary image to allow each apparent verti-
cal surface to be segmented from the scene. Apart from humans, common features
identied via this method of scene analysis include, walls, doors, and tall items
of furniture. Many of these false positives can be eliminated by setting sensible
size constraints based on the expected minimum and maximum size of a person.
Some false positives may still pass this stage of the detection process, however
this is acceptable and preferred to the alternative of creating false negatives (re-
jecting real people) as further discrimination is performed after feature extraction
which can generally eliminate remaining false positives. The result of the blob
detection and size checking process is shown in Fig. 3.2c where red blobs are
those rejected by the size constraints and orange blobs are the remaining blobs
identied as ROI. Fig. 3.2c shows 5 blobs detected by the system. The large red
blob in the centre of the Fig. 3.2c corresponds with the wall and adjacent couch
(seen in Fig. 3.2d) and has been rejected as a possible person because it is too
19
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
large. The small red blob on the left of Fig 3.2c represents the small portion of
the wall visible in the top left of Fig. 3.2d and has been rejected because it is
too small. Of the three orange blobs two are correctly identied as people. The
orange blob on the far right of Fig. 3.2c corresponds to the white wall on the right
of Fig. 3.2d. This segment has been identied as a ROI but will be disregarded
in the person identication stage based on its HSS.
(a) Bivariate histogram (b) Thresholding
A!
D!
B!
C!
E!
(c) Blob detection
(d) Scene from perspective of the sensor
Figure 3.2: Scene analysis process takes a pointcloud of the scene and detects
regions of interest used to segment the pointcloud.
After detecting and ltering ROI in the density image, the location of the
pixels included in each ROI are used to recover the points belonging to that ROI
20
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
from the original pointcloud. These pointcloud segments are then passed on to
the feature extraction stage.
3.2 Extracting the Head-to-Shoulder Signature
The feature extraction stage of the identication system is the major innovation
developed in this capstone project. The objective of this stage was to capture
descriptive characteristics of the individual from a pointcloud representation of
their head, neck and shoulders. Based on the outcome of the concept evaluation
(in Section 2.3) where lateral measurements of the head and shoulders were iden-
tied as potentially descriptive quantities, a method was devised extending this
idea to involve a series of lateral size measurements taken at regular intervals over
the region of the head and shoulders. In this way lateral sizes over the region
were explicitly measured and vertical characteristics, such as the distance from
the top of the head to the shoulders, were captured implicitly by changes in size
over the range of intervals.
In order to obtain these lateral size measurements the head-shoulders region
of the pointcloud is divided into a number of lateral slices of equal thickness h
covering the region from the top of the head to the top of the shoulders H (Fig.
3.3). The distance H is chosen to include the shoulders of most adults while
minimising coverage of the arms. H in the experiments presented in this report
was chosen to be 40cm based on a manual survey of the pointcloud data collected
in the concept evaluation study. The number and thickness of slices was chosen to
maximise the resolution of the feature vector obtained while ensuring that each
slice would contain sucient data points to extract the required information.
Selection of the slice thickness depends upon the sensor used and the range of
distances over which people are expected to be measured (D
max
). Eq. 3.1 was
used to determine the minimum slice thickness (h
min
) given the vertical angular
resolution of the sensor used () and the maximum distance of people from the
sensor (D
max
). The formula ensures at least one full row of depth pixels are
captured within each slice.
21
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
0.3
0.2
0.1
0
0.1
0.2
2
2.1
2.2
0.6
0.5
0.4
0.3
0.2
0.1
0
0.1
0.2
0.3
0.4
h!
H!
S
l
i
c
e
s
!
Head-to-Shoulder Signature!
1.7 1.8 1.9
0.2
0.1
0
0.1
0.2
0.3
0.4
Slice 1
1.7 1.8 1.9
0.2
0.1
0
0.1
0.2
0.3
0.4
Slice 2
1.7 1.8 1.9
0.2
0.1
0
0.1
0.2
0.3
0.4
Slice 3
1.7 1.8 1.9
0.2
0.1
0
0.1
0.2
0.3
0.4
Slice 4
1.7 1.8 1.9
0.2
0.1
0
0.1
0.2
0.3
0.4
Slice 5
1.7 1.8 1.9
0.2
0.1
0
0.1
0.2
0.3
0.4
Slice 6
1.7 1.8 1.9
0.2
0.1
0
0.1
0.2
0.3
0.4
Slice 7
1.7 1.8 1.9
0.2
0.1
0
0.1
0.2
0.3
0.4
Slice 8
1.7 1.8 1.9
0.2
0.1
0
0.1
0.2
0.3
0.4
Slice 9
1.7 1.8 1.9
0.2
0.1
0
0.1
0.2
0.3
0.4
Slice 10
Span!
Figure 3.3: The Head-to-Shoulder Signature is made up of rotationally robust
span features taken from a series of lateral slices of the pointcloud.
22
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
h
min
= 2D
max
tan() (3.1)
In the case of the concept evaluation study the data gathered was of people
standing orthogonally to the sensor, as shown in Fig. 3.4a. In this case lateral
measurements representing the head length and breadth and shoulder breadth
were easily obtained from the maximum and minimum x and z values of the
relevant segments. However when persons are not oriented orthogonally with
the sensor (Fig. 3.4b) the same maximum and minimum values do not give a
meaningful result. What is needed is a consistent lateral measure than is robust
to changes in viewing angle. This has been achieved in this method by means
of a search through all pairs of points in the pointcloud segment to nd the
greatest distance between any two points. This measurement is illustrated on
real point cloud data in Fig. 3.3 and diagrammatically in Fig. 3.4c. Due to
the concave nature of the human form (in the horizontal plane), data captured
from almost any single perspective (with the exception of those close to parallel
with the major axis) can be used in this way to measure a relatively (relative
to variation between individuals) consistent result relating to the length of the
object along the major axis. This dimension is referred to in this report as the
span of the pointcloud, and rather than accurately measure prescribed dimensions
such as head length or shoulder breadth. The aim of the span feature is to give
a repeatable measurement of lateral length of a concave 3D body. This span
feature is extracted from each slice of the pointcloud and the resulting vector of
numbers is what we refer to in this paper as the Head-to-Shoulder Signature.
The procedure presented in Alg. 1 describes the extraction of the HSS where
inputs to the algorithm are as follows:
P - three dimensional pointcloud of person, collection of (X, Y, Z) points
h - slice height, selected according to 3.1
H - segment height, 40cm in all our experiments
For each of the horizontal slices identied in line 8 the maximum distance
between any two points in the slice was calculated in line 13. The span was
23
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
z!
x!
Head length!
Shoulder breadth!
Head breadth!
(a) Measuring specic
human dimensions is
simplied when the
person is orthogonal to
the sensor
z!
x!
c
a
t
i
o
n
A
c
c
u
r
a
c
y
%
Eect of Number of Classes on Classication Accuracy
Figure 4.7: Graph showing the detrimental eect of increasing the size of the
problem (number of classes) on classication accuracy. Interestingly the graph
indicates that this eect will plateau around 60%.
4.3 Online performance of HSS Based Person
Identication System
4.3.1 Objective
For the person identication system to be useful in the context of HRI it would
need to be able to work in real time to identify people in an unconstrained envi-
ronment, presenting a variety of dierent behaviours. The aim of this experiment
was to test the ability of the system to perform under such circumstances recog-
nising a greater range of human behaviours than simply standing or walking
towards the robot. This test also served to evaluate the usefulness of associating
multiple classication results to improve the identication accuracy of the system
38
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
as this was identied as a potentially useful technique in the previous test.
4.3.2 Experimental Procedure
In this experiment the RobotAssist platform was setup (as per Fig.4.8) in a room
facing the doorway. The Microsoft Kinect Sensor was attached to the head of the
robot and streaming depth images to the robot at 30 frames per second for online
processing. Participants were instructed to enter and exit the room presenting a
variety of poses to the robot including:
normal walking,
carrying objects,
facing away from the robot, and
sitting on a rolling chair
Conrmation of identication was verbally communicated by the robot. In
order to improve performance in an unstructured environment the online imple-
mentation of the algorithm used temporal association of classication data with
individuals. A particle lter (Kirchner et al., 2010) was used to continuously track
people enabling temporal association of data with an individual. The model of
known persons HSS was trained prior to the experiment using the GUI developed
for the previous experiment. The trained model of HSS was used with SVM to
classify each observation and upon classication a vector of probabilities associ-
ated with each class was returned. These class probabilities were accumulated
over a number of frames and the average of accumulated probabilities was used to
determine the outcome of classication. Applying an empirically derived thresh-
old (0.8 in our experiment) to the accumulated probabilities enabled rejection of
low condence classication results enabling the robot to withhold identication
results until a high condence. This approach is preferable in the context of HRI
as opposed to announcing numerous results per track.
39
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
Hello
Dan !!
Figure 4.8: Experimental setup for online evaluation where people were recog-
nised walking, carrying objects, from behind and seated
4.3.3 Results and Discussion
The experiment resulted in successful identication of all four participants over all
7 encounters with the robot. The test was video recorded and the video was used
in the team description paper (Alempijevic, 2011) for RobotAssists successful
qualication to compete in RoboCup@Home 2011. A number of images from
the video are shown in Fig. 4.9. Specically Fig. 4.9a shows a person who was
identied while facing away from the robot and holding a large painting. Fig.
4.9b shows a person being identied while facing the robot. Fig. 4.9c shows a
person identied as they leave the room holding a stack of books. Finally Fig.
4.9d shows a person identied while sitting in a rolling chair.
Although only a small scale test these results show the applicability of this
system to a real HRI situation. The test also demonstrates the value of temporal
40
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
(a) From behind carrying a painting (b) Front facing
(c) From behind carrying books (d) Sitting in a rolling chair
Figure 4.9: Images of successful identication results in a range of poses conrm-
ing the suitability of the system to the challenges of HRI
association of classication results facilitated by the use of a tracking system.
4.4 Performance of HSS Based Person
Identication at RoboCup@Home 2011
4.4.1 Objective
The purpose of competing in the RoboCup@Home 2011 competition was to test
the robustness of the complete robotic system in a number of simulated HRI
scenarios. With regard to the person identication system it served to evaluate
the performance of the system to distinguish between known (trained online) and
41
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
unknown persons.
4.4.2 Follow Me Task
The Follow Me task at RoboCup@Home 2011 required the robot to meet and
learn a randomly allocated person to be its leader (Fig. 4.10a), and then follow
this person through a series of checkpoints designed to test the capabilities of
the robot. This task served as a suitable evaluation of the person identication
systems performance in a realistic HRI scenario. The stages of the task relevant
to the evaluation of the person identication system are shown below in Fig. 4.10.
After learning its leader the robot must be able to follow them around the
arena and maintain tracking as another person passes in between robot and leader
as illustrated in Fig. 4.10b. The RobotAssist platform uses a laser scanner and
particle lter (Kirchner et al., 2010) to track the leader because this is a fast and
robust way to track a moving target. In the event that the target is lost the
robot uses the person identication system to identify the leader. As described
in the method the HSS of all visible targets is compared with the trained model
of the leader to determine which is a better match. At the end of the test the
robot is instructed by the leader (via a hand signal) to stop and wait for them to
return. The leader leaves the eld of view of the robot and returns with a second
unknown person. Upon their return the two people face the robot and wait. The
robot must identify the correct person as its leader and continue to follow them
(Fig. 4.10c). This is again a test of the systems ability to distinguish between a
known person, the leader, and an unknown person.
The system was successful and scored 1000 out of a possible 1100 points.
The remaining 100 points were bonus points for outstanding performance which
needed to be negotiated prior to the test. Nonetheless this was the highest score
in the competition for this task tied with 2 of the other teams where the average
score from all 19 teams was 242. The results from this test conrmed the ability
of the system to successfully distinguish two individuals based on their HSS.
42
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
LEARN
LEADER!
(a) The robot learns the HSS of its leader
FOLLOW
LEADER!
(b) The robot must follow its leader as another person passes in front
REMEMBER
LEADER!
(c) After losing its target the robot must identify which person is its leader
Figure 4.10: Stages of the RoboCup@Home 2011 Follow Me task where the
RobotAssist team scored full marks.
43
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
4.4.3 Grand Final
The RoboCup@Home 2011 Grand Final task allowed teams to exhibit any func-
tionality they wished, and the strength of the performance was used to decide 2/3
of the teams nal competition score. The 2011 RobotAssist team used this op-
portunity to demonstrate the performance of the HSS based person identication
system and at the same time evaluate the robustness of the system to repeated
identication tasks in dicult scenarios. In particular the systems ability to
recognise seated individuals from behind was tested. Not only is this one of the
more challenging identication scenarios tested but it represents the realistic sit-
uation of a robot approaching a table of seated people in a home environment.
In this scenario the people would most likely be facing the table and hence facing
away from the robot.
The test was structured similarly to the Follow Me task. The robot learned
a leader at the beginning of the test and proceeded to follow them through a
number of trials. The most relevant part of this test to the evaluation of the
person identication was the nal trial. In order to test the ability of the robot
to recognise its leader in challenging circumstances the robot was blinded with a
cloth while the leader and another person were positioned in front of the robot
with their backs to it as, illustrated in Fig. 4.11a. While covered the robot
would lose its ability to track the position of the leader forcing it to rely on the
HSS based identication system alone to recognise its leader. The blinding cloth
was then removed and the robot would recognise its leader and resume following.
This process was repeated with dierent people and dierent sitting and standing
congurations until the test time had elapsed.
The system performed awlessly in this test, correctly identifying its leader
in 3 out of 3 cases. This test showed the capability of the system for robust
performance in a simulated home environment. Although only tested with with a
choice between 2 people each time, robustly performing this task online involved
a number of important capabilities. As discussed in section 4.3, to work in a HRI
context the robot needed to not simply correctly classify the observed HSS but
44
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
?!
(a) Robot is blinded disrupting its position tracking system
REMEMBER
LEADER!
(b) Robot uses HSS based person identication to recognise its owner
Figure 4.11: RoboCup@Home 2011 Grand Final: Robot recognises its leader
from behind
45
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
accumulate classication results until a high condence decision could be made.
The performance of the system was judged by a panel of external judges as well
as the RoboCup@Home technical committee. The nal competition score was
a combination of the external judges score, the technical committee score and
the performance in the competition up to that point, with each representing 1/3
of the nal score. Based on this scoring scheme RobotAssist placed 4th (out
of 19 teams) in the RoboCup@Home competition signicantly improving on the
previous years placed of 12th.
46
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
Chapter 5
Conclusions and Future Work
5.1 Conclusions
This capstone project has presented a novel method for person recognition which
exploits physical shape of the head, neck and shoulders to facilitate non-intrusive
person identication suitable for the context of HRI. The method presented
employs a rotationally robust feature vector (HSS) to encapsulate the head-to-
shoulder shape of individuals. Combined with a method of scene analysis for
person detection and the use of a SVM classier this capstone presented a com-
plete system for person detection and, learning and identication. A series of
empirical evaluations were performed to asses the performance of the system.
The person identication system was shown to recognise people (from a pool
of 9) in a natural walking motion with 76.8% accuracy. The system was shown to
scale surprisingly well to a problem size of 438 people with an accuracy (based on
individual observations) of 62%. The system was also demonstrated to identify
individuals: walking towards the robot, away from the robot, carrying objects
and sitting down. Finally the person identication system was implemented and
tested on the RobotAssist platform and used in competition at RoboCup@Home
2011 in Istanbul where it was a major contributor to the success of the RobotAs-
sist team who placed 4th out of 19 teams in the competition.
47
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
5.2 Future Work
Although the system has been shown to perform well in many situations there
are several of areas to be considered in future work.
5.2.1 Identifying Unknown People
Although a method for detecting unknown persons was devised, implemented and
utilised at RoboCup@Home this component of the system remains to be throughly
evaluated. Worthwhile future work in this area could include experimental eval-
uation of the single-class SVM method used. The eect of adjusting the gamma
and cost parameters in the SVM model could be investigated in greater detail
with an aim to developing an ecient method for determining the optimal pa-
rameters, given a set of training data. This method could also be compared to
other methods of anomaly detection.
5.2.2 Benets of Tracking
Through the process of testing the identication system it became evident that
the ability to associate HSS feature vectors and classication results with per-
sistent tracks would signicantly improve the performance of the system. Some
simple tracking techniques were implemented on the RobotAssist platform at the
competition but not thoroughly tested. Scope for future work could be to inte-
grate a particle lter based tracking component into the system and evaluate the
benets of using these persistent tracks to accumulate multiple HSS observations,
multiple classication results, or both.
5.2.3 Robustness to Changes in Personal Attire
An interesting question that has not been addressed in this capstone project is
the robustness of the system to changes in the personal appearance or attire of
an individual from day to day. It is likely that items such as hats, scarves, high
collars and variations in hair style will have an adverse eect on the performance
48
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
of the system in its current state. An interesting topic for further research and
development of this identication system would be what eect such items have
have on the HSS and whether these eects can be mitigated, for instance via
sparse representation of the HSS with some front end ltering system to eliminate
unexpected measurements associated with hats and other items.
49
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
Bibliography
Alen Alempijevic. Robotassist - robocup@home 2011 team description paper. In
RoboCup 2011, Sydney, 2011.
J. Barreto, P. Menezes, and J. Dias. Human-robot interaction based on haar-
like features and eigenfaces. In Robotics and Automation, 2004. Proceedings.
ICRA 04. 2004 IEEE International Conference on, volume 2, pages 1888
1893 Vol.2, 26-may 1, 2004. doi: 10.1109/ROBOT.2004.1308099.
Chih-Chung Chang and Chih-Jen Lin. LIBSVM: a library
for support vector machines, 2001. Software available at
http://www.csie.ntu.edu.tw/ cjlin/libsvm.
S.A. Cole. Suspect identities: a history of ngerprinting and criminal
identication. Harvard University Press, 2002.
C. Darwin. The origin of species. Number v. 1 in The Origin of Species. D. Ap-
pleton, 1902. URL http://books.google.com.au/books?id=QrcRAAAAYAAJ.
Gordon C. et al. 1988 anthropometric survey of u.s. army personnel: Methods and
summary statistics. Technical report, United States Army Natick, Research,
Development and Engineering Center, 1989.
Michael A. Goodrich and Alan C. Schultz. Human-robot interaction:
a survey. Found. Trends Hum.-Comput. Interact., 1:203275, Jan-
uary 2007. ISSN 1551-3955. doi: 10.1561/1100000005. URL
http://dl.acm.org/citation.cfm?id=1348099.1348100.
50
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
Daniel Hordern and Nathan Kirchner. Robust and Ecient People Detection
with 3-D Range Data using Shape Matching. In Proc. of the 2010 Australasian
Conference on Robotics and Automation (ACRA-10), pages 18, 2010.
A. Hrdlicka. Anthropometry. The Wistar institute of anatomy and biology, 1920.
MO Irfanoglu, B. Gokberk, and L. Akarun. 3D shape-based face recognition using
automatically registered facial surfaces. In Pattern Recognition, 2004. ICPR
2004. Proceedings of the 17th International Conference on, volume 4, pages
183186. IEEE, 2004. ISBN 0769521282.
P. Kakumanu, S. Makrogiannis, and N. Bourbakis. A survey of skin-color
modeling and detection methods. Pattern Recognition, 40(3):1106
1122, 2007. ISSN 0031-3203. doi: DOI: 10.1016/j.patcog.2006.06.010. URL
http://www.sciencedirect.com/science/article/pii/S0031320306002767.
Nathan Kirchner, Alen Alempijevic, Sonja Caraian, Robert Fitch, Daniel
Hordern, Gibson Hu, Gavin Paul, David Richards, Surya P. N. Singh, and
Stephen Webb. RobotAssist - a Platform for Human Robot Interaction Re-
search. In Proc. of the 2010 Australasian Conference on Robotics and Automa-
tion (ACRA-10), pages 110, 2010.
Nathan Kirchner, Alen Alempijevic, and Gamini Dissanayake. Nonverbal robot-
group interaction using an imitated gaze cue. In HRI 11: Proc. 6th ACM/IEEE
Int. Conf. on Human robot interaction, pages 18, Switzerland, 2011.
Kuang-Chih Lee, Jerey Ho, and David J. Kriegman. Acquiring linear subspaces
for face recognition under variable lighting. IEEE Transactions on Pattern
Analysis and Machine Intelligence, 27:684698, 2005. ISSN 0162-8828. doi:
http://doi.ieeecomputersociety.org/10.1109/TPAMI.2005.92.
B. Moghaddam and A. Pentland. Probabilistic visual learning for object repre-
sentation. Pattern Analysis and Machine Intelligence, IEEE Transactions on,
19(7):696 710, jul 1997. ISSN 0162-8828. doi: 10.1109/34.598227.
51
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
Stephen Pheasant. Bodyspace: anthropometry, ergonomics, and the design of
work. Taylor & Francis, 1996. ISBN 9780748400676.
Pramila Rani, Changchun Liu, Nilanjan Sarkar, and Eric Vanman. An em-
pirical study of machine learning techniques for aect recognition in hu-
manrobot interaction. Pattern Analysis and Applications, 9:5869, 2006.
ISSN 1433-7541. URL http://dx.doi.org/10.1007/s10044-006-0025-y.
10.1007/s10044-006-0025-y.
J.H. Relethford. The Human Species: An Introduction to Biological Anthropology.
McGraw-Hill, 2009.
C. Sanderson. Biometric Person Recognition Face, Speech and Fusion. VDM
Verlag, 2008.
Stevenson A. Soanes C. Oxford Dictionary of English. Oxford University Press,
2005.
A. Young V. Bruce. Understanding face recognition. British
Journal of Psychology, 77(3):305327, 1986. ISSN 2044-
8295. doi: 10.1111/j.2044-8295.1986.tb02199.x. URL
http://dx.doi.org/10.1111/j.2044-8295.1986.tb02199.x.
Paul Viola and Michael Jones. Rapid object detection using a boosted cas-
cade of simple features. Computer Vision and Pattern Recognition, IEEE
Computer Society Conference on, 1:511, 2001. ISSN 1063-6919. doi:
http://doi.ieeecomputersociety.org/10.1109/CVPR.2001.990517.
D. Voth. You can tell me by the way i walk. Intelligent Systems, IEEE, 18(1):4
5, Jan 2003. ISSN 1541-1672. doi: 10.1109/MIS.2003.1179185.
L. Wang, T. Tan, H. Ning, and W. Hu. Silhouette analysis-based gait recognition
for human identication. Pattern Analysis and Machine Intelligence, IEEE
Transactions on, 25(12):15051518, 2003. ISSN 0162-8828.
52
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
Mike Woods. Caring for older australians: Overview, report no. 53, nal inquiry
report. Technical report, Productivity Commission, Australian Government,
2011.
J. Wright, A.Y. Yang, A. Ganesh, S.S. Sastry, and Yi Ma. Robust face recogni-
tion via sparse representation. Pattern Analysis and Machine Intelligence,
IEEE Transactions on, 31(2):210 227, Feb. 2009. ISSN 0162-8828. doi:
10.1109/TPAMI.2008.79.
53
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
Appendix A
3D Sensing Technology
Crucial to the development of the person identication system presented here is
the use of 3D range cameras. These sensors are a type of camera which measures
the distance of the scene from the sensor at each pixel in the cameras eld of
view. This type of data is commonly referred to as a depth image. Using a depth
image in combination with some understanding of the optics of the range camera
a 3D pointcloud can be obtained. A pointcloud is a list of 3D cartesian points
which in this case represents the locations in space of physical surfaces as seen
by the range camera. This type of information is very useful in robotics because
it enables robots to interpret their environment spatially. There are two main
types of range camera technology, time-of-ight (TOF) and coded-light (CL).
A.1 Time of Flight
Time of ight cameras work by illuminating the scene in front of the camera with
an infrared lamp and sensing the reected infrared light with the camera. The
processor on board the sensor uses the phase dierence of the reected infrared
light to calculate the distance of the observed surface from the sensor. The Swis-
sRanger 4000 is an example of a TOF camera and was used in the development
of the identication system presented here.
54
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
A.2 Structured Light
Structured light 3D sensors work be projecting a specially structured pattern of
infrared light into the scene from a slightly dierent perspective to the camera.
From the point of view of the camera this projected pattern appears distorted
depending on the spatial placement of visible surfaces in the scene. The onboard
processor uses the discrepancies between the projected pattern and the pattern
as seen by the camera to calculate the depth at each part of the image. The
Kinect sensor, which is sold by Microsoft as a game controller for their popular
XBox gaming console, is based on the PrimeSensor 1.08 reference design and is
an example of a depth camera which uses structured light technology.
55
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
Appendix B
RobotAssist as a Development
Platform
B.1 The RobotAssist Project
The RobotAssist project aims to
Provide a platform for undergraduate and postgraduate students to conduct
research into robotics in particular HRI
Contribute international HRI research community
Integrate teaching and research, providing a path for students at all levels
(high school - university) to become involved in robotics research
Raise the prole of the university in the international robotics community
through competition in the RoboCup@Home competition
Raise the public prole of the university by generating media attention in
exciting projects at UTS.
Develop an adaptable hardware-software platform capable of a wide range
of activities in the arena of HRI and domestic service robotics.
56
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
B.2 The RobotAssist Platform
B.2.1 Hardware
The platform consists of a collection of o-the-shelf sensors, computers, and actu-
ators coupled with custom made frame and moving parts. The robot is built on a
two-wheeled, motorised segway base for mobility on which is mounted a desktop
computer running Ubuntu which acts as the main processor for the robot. The
sensors on the robot include:
3D Range Cameras (Kinect and Swissranger)
2D Laser Rangenders (Hokuyo)
XSens 6dof inertial measuring unit
Hi-res rgb camera
a pair of high quality condenser microphones (and external audio device)
B.2.2 Software
The platform is primarily controlled by a component-based software framework
called Orca. The Orca framework allows for each separate capability of the robot
to be developed as an individual component which can be started or stopped
independently of other components. These components communicate with one
another via a system called IceBox using a set of predened interfaces. Com-
ponents can publish data, subscribe to data, expose task interfaces, and invoke
task interfaces. This type of system is very versatile because components can
be stopped, started or even interchanged without shutting down the entire sys-
tem, facilitating testing and experimentation. The fact that components can be
developed independently of one another also allows multiple members of the de-
velopment team to be working on dierent parts of the project at he same with
out the need for constant consultation.
57
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
Appendix C
Classication Using Support
Vector Machines
The identication system developed for this capstone works by extracting the HSS
of the human user and classifying it as one of its known user base. The problem
of classication is (although not the main focus of this project) an interesting
problem in its own right. The system described in this report uses support vector
machines (SVM) to address the important task of classication.
Support vector machines are a concept in machine learning used for solving re-
gression problems (to obtain continuous results) as well as classication problems
( to obtain discrete results in the form of labels). In this project SVM is being
used for the latter (classication) to obtain discrete labels representing known
users to the system. In theory SVM classication produces a binary result. Two
classes are trained with some data points and the SVM denes the boundary
between the two classes. Further data points can then be classied using the
trained SVM and each point will be found to be on one side of the boundary or
another. Despite the binary nature of SVM several strategies exist for solving
multi-class problems using SVM, most of which involve reducing the multi-class
problem to a set of two-class problems.
For the purpose of this project an openly available library called LIBSVM
(Chang and Lin, 2001) was used to implement the classication stage of the
58
Head-to-Shoulder Signature for Person Identication in Human-Robot Interaction
system. This library has been written to interface with Matlab, C, Java, Python
and several other programming languages and it is widely used in the research
community. In this project the Matlab programming interface was used for oine
testing and development of the feature extraction algorithm and system as a
whole. The C interface was also used to create real time implementation
59