Affordable Access

Publisher Website

Inferring user tasks in pedestrian navigation from eye movement data in real-world environments

  • Liao, Hua
  • Dong, Weihua
  • Huang, Haosheng
  • Gartner, Georg
  • Liu, Huiping
Publication Date
Jan 01, 2019
DOI: 10.1080/13658816.2018.1482554
Ghent University Institutional Archive
External links


Eye movement data convey a wealth of information that can be used to probe human behaviour and cognitive processes. To date, eye tracking studies have mainly focused on laboratory-based evaluations of cartographic interfaces; in contrast, little attention has been paid to eye movement data mining for real-world applications. In this study, we propose using machine-learning methods to infer user tasks from eye movement data in real-world pedestrian navigation scenarios. We conducted a real-world pedestrian navigation experiment in which we recorded eye movement data from 38 participants. We trained and cross-validated a random forest classifier for classifying five common navigation tasks using five types of eye movement features. The results show that the classifier can achieve an overall accuracy of 67%. We found that statistical eye movement features and saccade encoding features are more useful than the other investigated types of features for distinguishing user tasks. We also identified that the choice of classifier, the time window size and the eye movement features considered are all important factors that influence task inference performance. Results of the research open doors to some potential real-world innovative applications, such as navigation systems that can provide task-related information depending on the task a user is performing.

Report this publication


Seen <100 times