Head Pose Estimation
What’s the liklihood that, using cameras, 3D sensors and lidar that we can estimate the head poses of people around the robot? Why should it matter? Knowing who is being attentive to the robot, as it passes them or as they pass by would be a useful interaction building block. Changes in head orientation could be a factor in choosing people for interaction. If, on making an interaction attempt, we see no deviation either of trajectory or head orientation, we can assume that we have failed to engage that person, and move on to our next candidate.
We want to examine head pose estimation using affordable, on-board robot sensing, such as RGB-D, laser sensors and camera data. Models using camera data exist for estimating the orientation of a persons head, and we want to investigate the effectiveness of these models also in high traffic scenarios. The ability to estimate head pose will be used in evaluation (we can use head movements by human subjects in the direction of the robot as evidence that attempts to attract their attention are successful).
With the robot situated in busy, public use areas, we will evaluate the accuracy of head pose estimation, both of stationary humans and those passing-by. We already annotate head orientation manually for existing experimental work, and so can use this approach as a gold standard measure.
Resources
*tensor flow based head orientation
References
Mark S Cary. The role of gaze in the initiation of conversation. Social Psychology, pages 269–271, 1978.