Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Panonut360: A Head and Eye Tracking Dataset for Panoramic Video (2403.17708v1)

Published 26 Mar 2024 in cs.CV, cs.HC, and cs.MM

Abstract: With the rapid development and widespread application of VR/AR technology, maximizing the quality of immersive panoramic video services that match users' personal preferences and habits has become a long-standing challenge. Understanding the saliency region where users focus, based on data collected with HMDs, can promote multimedia encoding, transmission, and quality assessment. At the same time, large-scale datasets are essential for researchers and developers to explore short/long-term user behavior patterns and train AI models related to panoramic videos. However, existing panoramic video datasets often include low-frequency user head or eye movement data through short-term videos only, lacking sufficient data for analyzing users' Field of View (FoV) and generating video saliency regions. Driven by these practical factors, in this paper, we present a head and eye tracking dataset involving 50 users (25 males and 25 females) watching 15 panoramic videos. The dataset provides details on the viewport and gaze attention locations of users. Besides, we present some statistics samples extracted from the dataset. For example, the deviation between head and eye movements challenges the widely held assumption that gaze attention decreases from the center of the FoV following a Gaussian distribution. Our analysis reveals a consistent downward offset in gaze fixations relative to the FoV in experimental settings involving multiple users and videos. That's why we name the dataset Panonut, a saliency weighting shaped like a donut. Finally, we also provide a script that generates saliency distributions based on given head or eye coordinates and pre-generated saliency distribution map sets of each video from the collected eye tracking data. The dataset is available on website: https://dianvrlab.github.io/Panonut360/.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (24)
  1. Salsabeel FM Alfalah “Perceptions Toward Adopting Virtual Reality as a Teaching Aid in Information Technology” In Education and Information Technologies 23 Springer, 2018, pp. 2633–2653
  2. “Virtual Reality for Stroke Rehabilitation” In Cochrane database of systematic reviews John Wiley & Sons, Ltd, 2017
  3. “Low-Cost Virtual Reality Environment for Engineering and Construction” In Visualization in Engineering 4.1 Springer, 2016, pp. 1–18
  4. “Evaluating Display Fidelity and Interaction Fidelity in a Virtual Reality Game” In IEEE Transactions on Visualization and Computer Graphics 18.4 IEEE, 2012, pp. 626–633
  5. “Utility-Driven Joint Caching and Bitrate Allocation for Real-Time Immersive Videos” In IEEE Journal of Selected Topics in Signal Processing 17.5, 2023, pp. 1106–1118
  6. Yashas Rai, Jesús Gutiérrez and Patrick Le Callet “A Dataset of Head and Eye Movements for 360 Degree Images” In Proceedings of the 8th ACM on Multimedia Systems Conference, 2017, pp. 205–210
  7. “A Dataset for Exploring User Behaviors in VR Spherical Video Streaming” In Proceedings of the 8th ACM on Multimedia Systems Conference, 2017, pp. 193–198
  8. “Gaze Prediction in Dynamic 360 Immersive Videos” In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 5333–5342
  9. “A Saliency Dataset for 360-Degree Videos” In Proceedings of the 10th ACM Multimedia Systems Conference, 2019, pp. 279–284
  10. “SalientVR: Saliency-Driven Mobile 360-Degree Video Streaming with Gaze Information” In Proceedings of the 28th Annual International Conference on Mobile Computing And Networking, 2022, pp. 542–555
  11. “Where Are You Looking? A Large-Scale Dataset of Head and Gaze Behavior for 360-Degree Videos and a Pilot Study” In Proceedings of the 30th ACM International Conference on Multimedia, 2022, pp. 1025–1034
  12. Xavier Corbillon, Francesca De Simone and Gwendal Simon “360-Degree Video Head Movement Dataset” In Proceedings of the 8th ACM on Multimedia Systems Conference, 2017, pp. 199–204
  13. “ATSal: An Attention Based Architecture for Saliency Prediction in 360∘superscript360360^{\circ}360 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT Videos” In International Conference on Pattern Recognition, 2021, pp. 305–320 Springer
  14. “Predicting Head Movement in Panoramic Video: A Deep Reinforcement Learning Approach ” In IEEE Transactions on Pattern Analysis and Machine Intelligence 41.11 IEEE, 2018, pp. 2693–2708
  15. Gov HK “Horizontal Field of View” Accessed: 2023-12-20, https://www.epd.gov.hk/eia/register/report/eiareport/eia_2522017/EIA/html/Appendix/Appendix11.1.pdf, 2023
  16. Yashas Rai, Patrick Le Callet and Philippe Guillotel “Which Saliency Weighting for Omni Directional Image Quality Assessment?” In 2017 Ninth International Conference on Quality of Multimedia Experience (QoMEX), 2017, pp. 1–6
  17. “Towards Retina-quality VR Video Streaming: 15ms Could Save You 80% of Your Bandwidth” In ACM SIGCOMM Computer Communication Review 52.1 ACM New York, NY, USA, 2022, pp. 10–19
  18. Mario Graf, Christian Timmerer and Christopher Mueller “Towards Bandwidth Efficient Adaptive Streaming of Omnidirectional Video over HTTP: Design, Implementation, and Evaluation” In Proceedings of the 8th ACM on Multimedia Systems Conference, 2017, pp. 261–271
  19. “Flare: Practical Viewport-Adaptive 360-Degree Video Streaming for Mobile Devices” In Proceedings of the 24th Annual International Conference on Mobile Computing and Networking, 2018, pp. 99–114
  20. “EPASS360: QoE-Aware 360-Degree Video Streaming Over Mobile Devices” In IEEE Transactions on Mobile Computing 20.7, 2021, pp. 2338–2353
  21. “Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting” In Proceedings of the AAAI Conference on Artificial Intelligence 35.12, 2021, pp. 11106–11115
  22. Dan Grois, Evgeny Kaminsky and Ofer Hadar “ROI Adaptive Scalable Video Coding for Limited Bandwidth Wireless Networks” In 2010 IFIP Wireless Days, 2010, pp. 1–5 IEEE
  23. “Collaborative Training between Region Proposal Localization and Classification for Domain Adaptive Object Detection” In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XVIII 16, 2020, pp. 86–102 Springer
  24. Anh Nguyen, Zhisheng Yan and Klara Nahrstedt “Your Attention is Unique: Detecting 360-Degree Video Saliency in Head-Mounted Display for Head Movement Prediction” In Proceedings of the 26th ACM International Conference on Multimedia, 2018, pp. 1190–1198
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Yutong Xu (3 papers)
  2. Junhao Du (2 papers)
  3. Jiahe Wang (8 papers)
  4. Yuwei Ning (2 papers)
  5. Sihan Zhou Yang Cao (1 paper)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets