Attention guides our gaze to fixate the proper location of the scene and holds it in that location for the de-served amount of time given current processing demands, before shifting to the next one. As such, gaze deploy-ment crucially is a temporal process. Existing computational models have made significant strides in predicting spatial aspects of observer's visual scanpaths (where to look), while often putting on the background the tempo-ral facet of attention dynamics (when). In this paper we present TPP-Gaze, a novel and principled approach to model scanpath dynamics based on Neural Temporal Point Process (TPP), that Jointly learns the temporal dynamics of fixations position and duration, integrating deep learning methodologies with point process theory. We conduct ex-tensive experiments across five publicly available datasets. Our results show the overall superior performance of the proposed model compared to state-of-the-art approaches. Source code and trained models are publicly available at: https://github.com/phuselab/tppgaze.

TPP-Gaze: Modelling Gaze Dynamics in Space and Time with Neural Temporal Point Processes / A. D'Amelio, G. Cartella, V. Cuculo, M. Lucchi, M. Cornia, R. Cucchiara, G. Boccignone - In: WACV[s.l] : Institute of Electrical and Electronics Engineers (IEEE), 2025. - ISBN 9798331510831. - pp. 8786-8795 (( convegno Winter Conference on Applications of Computer Vision : 28 February through 4 March tenutosi a Tucson (AZ USA) nel 2025 [10.1109/wacv61041.2025.00851].

TPP-Gaze: Modelling Gaze Dynamics in Space and Time with Neural Temporal Point Processes

A. D'Amelio
Primo
;
V. Cuculo;G. Boccignone
Ultimo
2025

Abstract

Attention guides our gaze to fixate the proper location of the scene and holds it in that location for the de-served amount of time given current processing demands, before shifting to the next one. As such, gaze deploy-ment crucially is a temporal process. Existing computational models have made significant strides in predicting spatial aspects of observer's visual scanpaths (where to look), while often putting on the background the tempo-ral facet of attention dynamics (when). In this paper we present TPP-Gaze, a novel and principled approach to model scanpath dynamics based on Neural Temporal Point Process (TPP), that Jointly learns the temporal dynamics of fixations position and duration, integrating deep learning methodologies with point process theory. We conduct ex-tensive experiments across five publicly available datasets. Our results show the overall superior performance of the proposed model compared to state-of-the-art approaches. Source code and trained models are publicly available at: https://github.com/phuselab/tppgaze.
gaze modelling; neural temporal point processes; scanpath prediction; visual attention;
Settore INFO-01/A - Informatica
2025
Adobe
IO Industries
Kitware
Mindtech
Prime Video Science
RapidFire.AI
Book Part (author)
File in questo prodotto:
File Dimensione Formato  
DAmelio_TPP-Gaze_Modelling_Gaze_Dynamics_in_Space_and_Time_with_Neural_WACV_2025_paper.pdf

accesso riservato

Tipologia: Publisher's version/PDF
Dimensione 4.21 MB
Formato Adobe PDF
4.21 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2434/1164318
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? 0
  • OpenAlex 3
social impact