Looking Clearer with Text: A Hierarchical Context Blending Network for Occluded Person Re-IdentificationShow others and affiliations
2025 (English)In: IEEE Transactions on Information Forensics and Security, ISSN 1556-6013, E-ISSN 1556-6021, Vol. 20, p. 4296-4307Article in journal (Refereed) Published
Abstract [en]
Existing occluded person re-identification (re-ID) methods mainly learn limited visual information for occluded pedestrians from images. However, textual information, which can describe various human appearance attributes, is rarely fully utilized in the task. To address this issue, we propose a Text-guided Hierarchical Context Blending Network ( THCB-Net) for occluded person re-ID. Specifically, at the data level, informative multi-modal inputs are first generated to make full use of the auxiliary role of textual information and make image data have a strong inductive bias for occluded environments. At the feature expression level, we design a novel Hierarchical Context Blending (HCB) module that can adaptively integrate shallow appearance features obtained by CNNs and multi-scale semantic features from visual transformer encoder. At the model optimization level, a Multi-modal Feature Interaction (MFI) module is proposed to learn the multi-modal information of pedestrians from texts and images, then guide the visual transformer encoder and HCB module to further learn discriminative identity information for occluded pedestrians through Image-Multimodal Contrastive (IMC) learning. Extensive experiments on standard occluded person re-ID benchmarks demonstrate that the proposed THCB-Net outperforms state-of-the-art methods. © 2025 IEEE.
Place, publisher, year, edition, pages
Piscataway, NJ: IEEE, 2025. Vol. 20, p. 4296-4307
Keywords [en]
Contrastive Learning, Feature Blending, Multi-modal Learning, Occluded Person Re-identification, Transformer
National Category
Computer graphics and computer vision Computer Sciences Natural Language Processing
Identifiers
URN: urn:nbn:se:hh:diva-55929DOI: 10.1109/TIFS.2025.3558586Scopus ID: 2-s2.0-105002302035OAI: oai:DiVA.org:hh-55929DiVA, id: diva2:1955447
2025-04-302025-04-302025-04-30Bibliographically approved