Panonut360: A Head and Eye Tracking Dataset for Panoramic Video (2403.17708v1)
Abstract: With the rapid development and widespread application of VR/AR technology, maximizing the quality of immersive panoramic video services that match users' personal preferences and habits has become a long-standing challenge. Understanding the saliency region where users focus, based on data collected with HMDs, can promote multimedia encoding, transmission, and quality assessment. At the same time, large-scale datasets are essential for researchers and developers to explore short/long-term user behavior patterns and train AI models related to panoramic videos. However, existing panoramic video datasets often include low-frequency user head or eye movement data through short-term videos only, lacking sufficient data for analyzing users' Field of View (FoV) and generating video saliency regions. Driven by these practical factors, in this paper, we present a head and eye tracking dataset involving 50 users (25 males and 25 females) watching 15 panoramic videos. The dataset provides details on the viewport and gaze attention locations of users. Besides, we present some statistics samples extracted from the dataset. For example, the deviation between head and eye movements challenges the widely held assumption that gaze attention decreases from the center of the FoV following a Gaussian distribution. Our analysis reveals a consistent downward offset in gaze fixations relative to the FoV in experimental settings involving multiple users and videos. That's why we name the dataset Panonut, a saliency weighting shaped like a donut. Finally, we also provide a script that generates saliency distributions based on given head or eye coordinates and pre-generated saliency distribution map sets of each video from the collected eye tracking data. The dataset is available on website: https://dianvrlab.github.io/Panonut360/.
- Salsabeel FM Alfalah “Perceptions Toward Adopting Virtual Reality as a Teaching Aid in Information Technology” In Education and Information Technologies 23 Springer, 2018, pp. 2633–2653
- “Virtual Reality for Stroke Rehabilitation” In Cochrane database of systematic reviews John Wiley & Sons, Ltd, 2017
- “Low-Cost Virtual Reality Environment for Engineering and Construction” In Visualization in Engineering 4.1 Springer, 2016, pp. 1–18
- “Evaluating Display Fidelity and Interaction Fidelity in a Virtual Reality Game” In IEEE Transactions on Visualization and Computer Graphics 18.4 IEEE, 2012, pp. 626–633
- “Utility-Driven Joint Caching and Bitrate Allocation for Real-Time Immersive Videos” In IEEE Journal of Selected Topics in Signal Processing 17.5, 2023, pp. 1106–1118
- Yashas Rai, Jesús Gutiérrez and Patrick Le Callet “A Dataset of Head and Eye Movements for 360 Degree Images” In Proceedings of the 8th ACM on Multimedia Systems Conference, 2017, pp. 205–210
- “A Dataset for Exploring User Behaviors in VR Spherical Video Streaming” In Proceedings of the 8th ACM on Multimedia Systems Conference, 2017, pp. 193–198
- “Gaze Prediction in Dynamic 360 Immersive Videos” In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 5333–5342
- “A Saliency Dataset for 360-Degree Videos” In Proceedings of the 10th ACM Multimedia Systems Conference, 2019, pp. 279–284
- “SalientVR: Saliency-Driven Mobile 360-Degree Video Streaming with Gaze Information” In Proceedings of the 28th Annual International Conference on Mobile Computing And Networking, 2022, pp. 542–555
- “Where Are You Looking? A Large-Scale Dataset of Head and Gaze Behavior for 360-Degree Videos and a Pilot Study” In Proceedings of the 30th ACM International Conference on Multimedia, 2022, pp. 1025–1034
- Xavier Corbillon, Francesca De Simone and Gwendal Simon “360-Degree Video Head Movement Dataset” In Proceedings of the 8th ACM on Multimedia Systems Conference, 2017, pp. 199–204
- “ATSal: An Attention Based Architecture for Saliency Prediction in 360∘superscript360360^{\circ}360 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT Videos” In International Conference on Pattern Recognition, 2021, pp. 305–320 Springer
- “Predicting Head Movement in Panoramic Video: A Deep Reinforcement Learning Approach ” In IEEE Transactions on Pattern Analysis and Machine Intelligence 41.11 IEEE, 2018, pp. 2693–2708
- Gov HK “Horizontal Field of View” Accessed: 2023-12-20, https://www.epd.gov.hk/eia/register/report/eiareport/eia_2522017/EIA/html/Appendix/Appendix11.1.pdf, 2023
- Yashas Rai, Patrick Le Callet and Philippe Guillotel “Which Saliency Weighting for Omni Directional Image Quality Assessment?” In 2017 Ninth International Conference on Quality of Multimedia Experience (QoMEX), 2017, pp. 1–6
- “Towards Retina-quality VR Video Streaming: 15ms Could Save You 80% of Your Bandwidth” In ACM SIGCOMM Computer Communication Review 52.1 ACM New York, NY, USA, 2022, pp. 10–19
- Mario Graf, Christian Timmerer and Christopher Mueller “Towards Bandwidth Efficient Adaptive Streaming of Omnidirectional Video over HTTP: Design, Implementation, and Evaluation” In Proceedings of the 8th ACM on Multimedia Systems Conference, 2017, pp. 261–271
- “Flare: Practical Viewport-Adaptive 360-Degree Video Streaming for Mobile Devices” In Proceedings of the 24th Annual International Conference on Mobile Computing and Networking, 2018, pp. 99–114
- “EPASS360: QoE-Aware 360-Degree Video Streaming Over Mobile Devices” In IEEE Transactions on Mobile Computing 20.7, 2021, pp. 2338–2353
- “Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting” In Proceedings of the AAAI Conference on Artificial Intelligence 35.12, 2021, pp. 11106–11115
- Dan Grois, Evgeny Kaminsky and Ofer Hadar “ROI Adaptive Scalable Video Coding for Limited Bandwidth Wireless Networks” In 2010 IFIP Wireless Days, 2010, pp. 1–5 IEEE
- “Collaborative Training between Region Proposal Localization and Classification for Domain Adaptive Object Detection” In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XVIII 16, 2020, pp. 86–102 Springer
- Anh Nguyen, Zhisheng Yan and Klara Nahrstedt “Your Attention is Unique: Detecting 360-Degree Video Saliency in Head-Mounted Display for Head Movement Prediction” In Proceedings of the 26th ACM International Conference on Multimedia, 2018, pp. 1190–1198
- Yutong Xu (3 papers)
- Junhao Du (2 papers)
- Jiahe Wang (8 papers)
- Yuwei Ning (2 papers)
- Sihan Zhou Yang Cao (1 paper)