Reliable detection of slow-moving unmanned aerial vehicles (UAVs) under low signal-to-noise ratio (SNR) conditions remains a core challenge for the next-generation intelligent sensing systems. We propose AirSentinel, a novel deep radar perception network to leverage frame-synchronized real-time kinematic (RTK)-derived geocoordinates for supervised learning, bridging the gap between image- and radar-based methods. To address the challenge of weak feature discrimination and energy leakage in radar data, we propose a local search and guard band (LSGB) mechanism that identifies Doppler peaks across coherent processing intervals (CPIs). We further propose an information weight calibration adapter (IWCA) and a global temporal modeling head to facilitate feature alignment across views and capture long-range temporal dependencies, respectively. To mitigate data scarcity, an object transformation and generation scheme is introduced, significantly boosting data diversity and annotation efficiency. Extensive experiments on the newly curated radarUAV-24K dataset demonstrate that AirSentinel outperforms state-of-the-art methods in both accuracy and real-time performance. This work establishes a robust paradigm for radar-based object detection, with implications for autonomous monitoring and urban airspace security.
Deep Radar Perception Network with RTK Supervision for Slow-Moving UAV Detection / Z. Jiang, F. Ke, Y. Zhai, J. Zhou, P. Coscia, A. Genovese, X.Y. Zhang. - In: IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT. - ISSN 0018-9456. - 75:(2026), pp. 2510318.1-2510318.18. [10.1109/tim.2026.3687291]
Deep Radar Perception Network with RTK Supervision for Slow-Moving UAV Detection
P. Coscia;A. GenovesePenultimo
;
2026
Abstract
Reliable detection of slow-moving unmanned aerial vehicles (UAVs) under low signal-to-noise ratio (SNR) conditions remains a core challenge for the next-generation intelligent sensing systems. We propose AirSentinel, a novel deep radar perception network to leverage frame-synchronized real-time kinematic (RTK)-derived geocoordinates for supervised learning, bridging the gap between image- and radar-based methods. To address the challenge of weak feature discrimination and energy leakage in radar data, we propose a local search and guard band (LSGB) mechanism that identifies Doppler peaks across coherent processing intervals (CPIs). We further propose an information weight calibration adapter (IWCA) and a global temporal modeling head to facilitate feature alignment across views and capture long-range temporal dependencies, respectively. To mitigate data scarcity, an object transformation and generation scheme is introduced, significantly boosting data diversity and annotation efficiency. Extensive experiments on the newly curated radarUAV-24K dataset demonstrate that AirSentinel outperforms state-of-the-art methods in both accuracy and real-time performance. This work establishes a robust paradigm for radar-based object detection, with implications for autonomous monitoring and urban airspace security.| File | Dimensione | Formato | |
|---|---|---|---|
|
tim26b-min.pdf
accesso aperto
Tipologia:
Post-print, accepted manuscript ecc. (versione accettata dall'editore)
Licenza:
Creative commons
Dimensione
3.94 MB
Formato
Adobe PDF
|
3.94 MB | Adobe PDF | Visualizza/Apri |
|
Deep_Radar_Perception_Network_With_RTK_Supervision_for_Slow-Moving_UAV_Detection(4).pdf
accesso riservato
Tipologia:
Publisher's version/PDF
Licenza:
Nessuna licenza
Dimensione
6.59 MB
Formato
Adobe PDF
|
6.59 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.




