Content not yet available
This lecture has no active video or poster.
Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.
Video-based visible-infrared person re-identification (VVI-ReID) aims to match pedestrian video sequences captured across different modalities and viewpoints, and plays a critical role in all-day intelligent surveillance. While recent supervised methods have shown promising results, they rely on large-scale cross-modal video annotations, which are expensive and difficult to obtain in practice. To address this limitation, we introduce the task of unsupervised domain adaptation for video-based visible-infrared person re-identification (UDA-VVI-ReID), where a model is transferred from a labeled source domain to an unlabeled target domain. This setting presents unique challenges, including modality discrepancies, temporal variations, and the difficulty of generating reliable pseudo-labels under occlusion or motion noise. To tackle these issues, we propose a Dynamic-Static Collaboration (DSC) framework that integrates two key modules. The Dynamic-Static Label Unification (DSLU) module refines pseudo-labels by enforcing consistency between appearance and motion features across modalities. The Dynamic-Static Joint Learning (DSJL) module further enhances representation learning through contrastive objectives and neighbor-based feature alignment guided by both dynamic and static cues. Experimental results on the HITSZ-VCM and BUPTCampus datasets demonstrate that this method achieves state-of-the-art performance among unsupervised approaches without relying on target domain labels.