The remarkable progress in deep learning (DL) show-cases outstanding results in various computer vision tasks. However, adaptation to real-time variations in data distributions remains an important challenge. Test- Time Training (TTT) was proposed as an effective solution to this issue, which increases the generalization ability of trained models by adding an auxiliary task at train time and then using its loss at test time to adapt the model. Inspired by the recent achievements of contrastive representation learning in unsupervised tasks, we propose ReC-TTT, a test-time training technique that can adapt a DL model to new un-seen domains by generating discriminative views of the input data. ReC- Ttt uses cross-reconstruction as an auxiliary task between a frozen encoder and two trainable en-coders, taking advantage of a single shared decoder. This enables, at test time, to adapt the encoders to extract features that will be correctly reconstructed by the decoder that, in this phase, is frozen on the source domain. Experimental results show that ReC- Ttt achieves better re-sults than other state-of-the-art techniques in most domain shift classification challenges. The code is available at: https://github.com/warpcut/ReC-TTT

ReC- Ttt: Contrastive Feature Reconstruction for Test-Time Training / M. Colussi, S. Mascetti, J. Dolz, C. Desrosiers (IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION). - In: WACV 2025[s.l] : Institute of Electrical and Electronics Engineers (IEEE), 2025. - ISBN 979-8-3315-1083-1. - pp. 6699-6708 (( Winter Conference on Applications of Computer Vision : 26 February - 06 March Tucson (AZ, USA) 2025 [10.1109/wacv61041.2025.00652].

ReC- Ttt: Contrastive Feature Reconstruction for Test-Time Training

M. Colussi
Primo
;
S. Mascetti
Secondo
;
2025

Abstract

The remarkable progress in deep learning (DL) show-cases outstanding results in various computer vision tasks. However, adaptation to real-time variations in data distributions remains an important challenge. Test- Time Training (TTT) was proposed as an effective solution to this issue, which increases the generalization ability of trained models by adding an auxiliary task at train time and then using its loss at test time to adapt the model. Inspired by the recent achievements of contrastive representation learning in unsupervised tasks, we propose ReC-TTT, a test-time training technique that can adapt a DL model to new un-seen domains by generating discriminative views of the input data. ReC- Ttt uses cross-reconstruction as an auxiliary task between a frozen encoder and two trainable en-coders, taking advantage of a single shared decoder. This enables, at test time, to adapt the encoders to extract features that will be correctly reconstructed by the decoder that, in this phase, is frozen on the source domain. Experimental results show that ReC- Ttt achieves better re-sults than other state-of-the-art techniques in most domain shift classification challenges. The code is available at: https://github.com/warpcut/ReC-TTT
contrastive feature reconstruction; domain adaptation; image classification; test-time training;
Settore INFO-01/A - Informatica
   Tight control of treatment adherence and efficacy by tElemedicine for an improved personalized Management of Patients with hemOphilia - TEMPO project
   TEMPO
   MINISTERO DELL'UNIVERSITA' E DELLA RICERCA
   2022PKTW2B_001

   MUSA - Multilayered Urban Sustainability Actiona
   MUSA
   MINISTERO DELL'UNIVERSITA' E DELLA RICERCA
2025
Adobe
IO Industries
Kitware
Mindtech
Prime Video Science
RapidFire.AI
Book Part (author)
File in questo prodotto:
File Dimensione Formato  
REC_TTT_WACV_2025 (2).pdf

accesso aperto

Tipologia: Post-print, accepted manuscript ecc. (versione accettata dall'editore)
Licenza: Creative commons
Dimensione 1.65 MB
Formato Adobe PDF
1.65 MB Adobe PDF Visualizza/Apri
ReC-_Ttt_Contrastive_Feature_Reconstruction_for_Test-Time_Training(2).pdf

accesso riservato

Tipologia: Publisher's version/PDF
Licenza: Nessuna licenza
Dimensione 1.72 MB
Formato Adobe PDF
1.72 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2434/1231017
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
  • OpenAlex 1
social impact