Head Pose Estimation

What’s the liklihood that, using cameras, 3D sensors and lidar that we can estimate the head poses of people around the robot? Why should it matter? Knowing who is being attentive to the robot, as it passes them or as they pass by would be a useful interaction building block. Changes in head orientation could be a factor in choosing people for interaction. If, on making an interaction attempt, we see no deviation either of trajectory or head orientation, we can assume that we have failed to engage that person, and move on to our next candidate.

We want to examine head pose estimation using affordable, on-board robot sensing, such as RGB-D, laser sensors and camera data. Models using camera data exist for estimating the orientation of a persons head, and we want to investigate the effectiveness of these models also in high traffic scenarios. The ability to estimate head pose will be used in evaluation (we can use head movements by human subjects in the direction of the robot as evidence that attempts to attract their attention are successful).

With the robot situated in busy, public use areas, we will evaluate the accuracy of head pose estimation, both of stationary humans and those passing-by. We already annotate head orientation manually for existing experimental work, and so can use this approach as a gold standard measure.

Resources

*real-time head orientation

*tensor flow based head orientation

References

Mark S Cary. The role of gaze in the initiation of conversation. Social Psychology, pages 269–271, 1978.

Avatar
Nick Webb
Associate Professor of Computer Science / Director of Data Analytics

My research interests include Natural Language Processing, Social Robotics and Data Analytics.

Related