Learning to Predict Intent from Gaze During Robotic Hand-Eye Coordination

TitleLearning to Predict Intent from Gaze During Robotic Hand-Eye Coordination
Publication TypeConference Paper
Year of Publication2017
AuthorsRazin, Y.
EditorFeigh, K.
Conference NameAAAI
Date Published02/2017
Conference LocationSan Francisco
KeywordsApplication of Supervised Learning, Gaze Tracking, Human-Aware AI, Human-Robot Interaction
URLhttps://aaai.org/ocs/index.php/AAAI/AAAI17/paper/view/14270/0
Full Text

Effective human-aware robots should anticipate their user’s intentions. During hand-eye coordination tasks, gaze often precedes hand motion and can serve as a powerful predictor for intent. However, cooperative tasks where a semi-autonomous robot serves as an extension of the human hand have rarely been studied in the context of hand-eye coordination. We hypothesize that accounting for anticipatory eye movements in addition to the movements of the robot will improve intent estimation. This research compares the application of various machine learning methods to intent prediction from gaze tracking data during robotic hand-eye coordination tasks. We found that with proper feature selection, accuracies exceeding 94% and AUC greater than 91% are achievable with several classification algorithms but that anticipatory gaze data did not improve intent prediction.

Map of Cognitive Engineering Center

Cognitive Engineering Center (CEC)
Georgia Institute of Technology
270 Ferst Drive
Atlanta GA 30332-0150