News
- February, 2024: Paper accepted to The Visual Computer (see Downloads).
- February, 2024: Web launched.
Abstract
Advances in virtual and augmented reality have increased the demand for immersive and engaging 3D experiences. To create such experiences, it is crucial to understand visual attention in 3D environments, which is typically modeled by means of saliency maps. While attention in 2D images and traditional media has been widely studied, there is still much to explore in 3D settings. In this work, we propose a deep learning-based model for predicting saliency when viewing 3D objects, which is a first step toward understanding and predicting attention in 3D environments. Previous approaches rely solely on low-level geometric cues or unnatural conditions, however, our model is trained on a dataset of real viewing data that we have manually captured, which indeed reflects actual human viewing behavior. Our approach outperforms existing state-of-the-art methods and closely approximates the ground-truth data. Our results demonstrate the effectiveness of our approach in predicting attention in 3D objects, which can pave the way for creating more immersive and engaging 3D experiences.
Downloads
Code
We will make code available soon.
Bibtex
Related Work
- 2020: Panoramic convolutions for 360º single-image saliency prediction
- 2022: Scangan360: A generative model of realistic scanpaths for 360 images
- 2023: D-SAV360: A Dataset of Gaze Scanpaths on 360° Ambisonic Videos