Characterising eye movement events with multi-scale spatio-temporal awareness

Yang ZHENG, Zhiyong YU, Hong FU, Kaitai GUO, Jimin LIANG

Research output: Contribution to journalArticlespeer-review

Abstract

The intricate and dynamic nature of eye movements serves as a window into the realms of cognition, emotion, and physiological responses. Event detection, in turn, is instrumental in the precise recognition and categorization of these diverse eye movements. Deep learning methods have recently been applied to event detection, yielding promising results. However, the intrinsic multi-scale attributes of events have often been overlooked in existing approaches. To address this, we introduce “GazeUNet”, a novel network based on U-Net and Bi-GRU, which classifies gaze samples into three categories: fixation, saccade, and post-saccadic oscillations. Firstly, multi-scale spatial features are captured using a U-Net model, and then a hierarchical bidirectional gated recurrent unit (Bi-GRU) is employed to extract temporal correlations, followed by classification through fully connected layers. Our results, derived from the analysis of three publicly available datasets, consistently showcase the superiority of the proposed model compared with other state-of-the-art methods across all categories. Copyright © 2024 IEEE.

Original languageEnglish
Pages (from-to)2090-2094
JournalIEEE Signal Processing Letters
Early online dateAug 2024
DOIs
Publication statusPublished - 2024

Citation

Zheng, Y., Yu, Z., Fu, H., Guo, K., & Liang, J. (2024). Characterising eye movement events with multi-scale spatio-temporal awareness. IEEE Signal Processing Letters, 31, 2090-2094. https://doi.org/10.1109/LSP.2024.3441490

Keywords

  • Bidirectional gated recurrent unit
  • Deep learning
  • Eye movement event detection
  • Multi-scale convolution

Fingerprint

Dive into the research topics of 'Characterising eye movement events with multi-scale spatio-temporal awareness'. Together they form a unique fingerprint.