15

VReID-XFD: Video-based Person Re-identification at Extreme Far Distance Challenge Results

Kailash A. Hambarde
Hugo Proença
Md Rashidunnabi
Pranita Samale
Qiwei Yang
Pingping Zhang
Zijing Gong
Yuhao Wang
Xi Zhang
Ruoshui Qu
Qiaoyun He
Yuhang Zhang
Thi Ngoc Ha Nguyen
Tien-Dung Mai
Cheng-Jun Kang
Yu-Fan Lin
Jin-Hui Jiang
Chih-Chung Hsu
Tamás Endrei
György Cserey
Ashwat Rajbhandari
Main:8 Pages
10 Figures
Bibliography:2 Pages
2 Tables
Abstract

Person re-identification (ReID) across aerial and ground views at extreme far distances introduces a distinct operating regime where severe resolution degradation, extreme viewpoint changes, unstable motion cues, and clothing variation jointly undermine the appearance-based assumptions of existing ReID systems. To study this regime, we introduce VReID-XFD, a video-based benchmark and community challenge for extreme far-distance (XFD) aerial-to-ground person re-identification. VReID-XFD is derived from the DetReIDX dataset and comprises 371 identities, 11,288 tracklets, and 11.75 million frames, captured across altitudes from 5.8 m to 120 m, viewing angles from oblique (30 degrees) to nadir (90 degrees), and horizontal distances up to 120 m. The benchmark supports aerial-to-aerial, aerial-to-ground, and ground-to-aerial evaluation under strict identity-disjoint splits, with rich physical metadata. The VReID-XFD-25 Challenge attracted 10 teams with hundreds of submissions. Systematic analysis reveals monotonic performance degradation with altitude and distance, a universal disadvantage of nadir views, and a trade-off between peak performance and robustness. Even the best-performing SAS-PReID method achieves only 43.93 percent mAP in the aerial-to-ground setting. The dataset, annotations, and official evaluation protocols are publicly available atthis https URL.

View on arXiv
Comments on this paper