Attention guides our gaze to fixate the proper location of the scene and holds it in that location for the de-served amount of time given current processing demands, before shifting to the next one. As such, gaze deploy-ment crucially is a temporal process. Existing computational models have made significant strides in predicting spatial aspects of observer's visual scanpaths (where to look), while often putting on the background the tempo-ral facet of attention dynamics (when). In this paper we present TPP-Gaze, a novel and principled approach to model scanpath dynamics based on Neural Temporal Point Process (TPP), that Jointly learns the temporal dynamics of fixations position and duration, integrating deep learning methodologies with point process theory. We conduct ex-tensive experiments across five publicly available datasets. Our results show the overall superior performance of the proposed model compared to state-of-the-art approaches. Source code and trained models are publicly available at: https://github.com/phuselab/tppgaze.
TPP-Gaze: Modelling Gaze Dynamics in Space and Time with Neural Temporal Point Processes / A. D'Amelio, G. Cartella, V. Cuculo, M. Lucchi, M. Cornia, R. Cucchiara, G. Boccignone - In: WACV[s.l] : Institute of Electrical and Electronics Engineers (IEEE), 2025. - ISBN 9798331510831. - pp. 8786-8795 (( convegno Winter Conference on Applications of Computer Vision : 28 February through 4 March tenutosi a Tucson (AZ USA) nel 2025 [10.1109/wacv61041.2025.00851].
TPP-Gaze: Modelling Gaze Dynamics in Space and Time with Neural Temporal Point Processes
A. D'Amelio
Primo
;V. Cuculo;G. BoccignoneUltimo
2025
Abstract
Attention guides our gaze to fixate the proper location of the scene and holds it in that location for the de-served amount of time given current processing demands, before shifting to the next one. As such, gaze deploy-ment crucially is a temporal process. Existing computational models have made significant strides in predicting spatial aspects of observer's visual scanpaths (where to look), while often putting on the background the tempo-ral facet of attention dynamics (when). In this paper we present TPP-Gaze, a novel and principled approach to model scanpath dynamics based on Neural Temporal Point Process (TPP), that Jointly learns the temporal dynamics of fixations position and duration, integrating deep learning methodologies with point process theory. We conduct ex-tensive experiments across five publicly available datasets. Our results show the overall superior performance of the proposed model compared to state-of-the-art approaches. Source code and trained models are publicly available at: https://github.com/phuselab/tppgaze.| File | Dimensione | Formato | |
|---|---|---|---|
|
DAmelio_TPP-Gaze_Modelling_Gaze_Dynamics_in_Space_and_Time_with_Neural_WACV_2025_paper.pdf
accesso riservato
Tipologia:
Publisher's version/PDF
Dimensione
4.21 MB
Formato
Adobe PDF
|
4.21 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.




