Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MambaPupil: Bidirectional Selective Recurrent model for Event-based Eye tracking (2404.12083v2)

Published 18 Apr 2024 in cs.CV

Abstract: Event-based eye tracking has shown great promise with the high temporal resolution and low redundancy provided by the event camera. However, the diversity and abruptness of eye movement patterns, including blinking, fixating, saccades, and smooth pursuit, pose significant challenges for eye localization. To achieve a stable event-based eye-tracking system, this paper proposes a bidirectional long-term sequence modeling and time-varying state selection mechanism to fully utilize contextual temporal information in response to the variability of eye movements. Specifically, the MambaPupil network is proposed, which consists of the multi-layer convolutional encoder to extract features from the event representations, a bidirectional Gated Recurrent Unit (GRU), and a Linear Time-Varying State Space Module (LTV-SSM), to selectively capture contextual correlation from the forward and backward temporal relationship. Furthermore, the Bina-rep is utilized as a compact event representation, and the tailor-made data augmentation, called as Event-Cutout, is proposed to enhance the model's robustness by applying spatial random masking to the event image. The evaluation on the ThreeET-plus benchmark shows the superior performance of the MambaPupil, which secured the 1st place in CVPR'2024 AIS Event-based Eye Tracking challenge.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (29)
  1. Bina-rep event frames: A simple and effective representation for event-based cameras. In IEEE International Conference on Image Processing, pages 3998–4002, 2022.
  2. A low-power neuromorphic approach for efficient eye-tracking. arXiv preprint arXiv:2312.00425, 2023a.
  3. Tinytracker: Ultra-fast and ultra-low-power edge vision in-sensor for gaze estimation. IEEE Sensors, pages 1–4, 2023b.
  4. 3et: Efficient event-based eye tracking using a change-based convlstm network. In IEEE Biomedical Circuits and Systems Conference, pages 1–5, 2023.
  5. Event-based eye tracking - ais2024 cvpr workshop, 2024. https://kaggle.com/competitions/event-based-eye-tracking-ais2024.
  6. Empirical evaluation of gated recurrent neural networks on sequence modeling. In Conference on Neural Information Processing Systems Workshop on Deep Learning, 2014.
  7. Event-based vision: A survey. IEEE Transactions on Pattern Analysis & Machine Intelligence, 44(01):154–180, 2022.
  8. Mamba: Linear-time sequence modeling with selective state spaces. arXiv preprint arXiv:2312.00752, 2023.
  9. Efficiently modeling long sequences with structured state spaces. In The International Conference on Learning Representations, 2022.
  10. Remoteeye: An open-source high-speed remote eye tracker: Implementation insights of a pupil- and glint-detection algorithm for high-speed remote eye tracking. Behavior Research Methods, 52:1387–1401, 2020.
  11. Gaze360: Physically unconstrained gaze estimation in the wild. In International Conference on Computer Vision, page 6912–6921, 2019.
  12. Masked event modeling: Self-supervised pretraining for event cameras. In IEEE/CVF Winter Conference on Applications of Computer Vision, 2024.
  13. Deep learning-based pupil center detection for fast and accurate eye tracking system. In European Conference on Computer Vision, pages 36–52, 2020.
  14. Eye tracking in augmented spaces: A deep learning approach. In 2018 IEEE Games, Entertainment, Media Conference, pages 1–6, 2018.
  15. SGDR: stochastic gradient descent with warm restarts. In International Conference on Learning Representations, pages 24–26, 2017.
  16. Gaze and eye-tracking solutions for psychological research. 13:261–265, 2012.
  17. Robust eye tracking based on multiple corneal reflections for clinical applications. Journal of biomedical optics, page 23, 2018.
  18. Recurrent CNN for 3d gaze estimation using appearance and shape cues. In British Machine Vision Conference, page 251, 2018.
  19. Pytorch: An imperative style, high-performance deep learning library. 32:8024–8035, 2019.
  20. Fixation detection for head-mounted eye tracking based on visual similarity of gaze targets. In Proceedings of the 2018 ACM Symposium on Eye Tracking Research & Applications, pages 23:1–23:9, 2018.
  21. Event-based kilohertz eye tracking using coded differential lighting. In IEEE/CVF Winter Conference on Applications of Computer Vision, pages 3937–3945, 2022.
  22. Real time eye gaze tracking for human machine interaction in the cockpit. AI and Optical Data Sciences III, pages 24–33, 2022.
  23. Efficient object localization using convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2015.
  24. Event-Based Eye Tracking. AIS 2024 Challenge Survey. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2024.
  25. Exploring figure-ground assignment mechanism in perceptual organization. Advances in Neural Information Processing Systems, 35:17030–17042, 2022a.
  26. One-shot object affordance detection in the wild. International Journal of Computer Vision, 130(10):2472–2500, 2022b.
  27. On exploring multiplicity of primitives and attributes for texture recognition in the wild. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
  28. Background activation suppression for weakly supervised object localization and semantic segmentation. International Journal of Computer Vision, 132(3):750–775, 2024.
  29. Evaluation of appearance-based methods and implications for gaze-based applications. pages 1–13, 2019.
Citations (3)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com