hh.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
MV-ReID: 3D Multi-view Transformation Network for Occluded Person Re-Identification
Chinese Academy of Sciences, Beijing, China; University of Chinese Academy of Sciences, Beijing, China.ORCID iD: 0000-0002-3425-1153
Halmstad University, School of Information Technology.ORCID iD: 0000-0002-2851-4260
Beijing University of Posts and Telecommunications, Beijing, China.
Old Dominion University, Norfolk, United States.ORCID iD: 0000-0002-4323-2632
Show others and affiliations
2024 (English)In: Knowledge-Based Systems, ISSN 0950-7051, E-ISSN 1872-7409, Vol. 283, article id 111200Article in journal (Refereed) Published
Abstract [en]

Re-identification (ReID) of occluded persons is a challenging task due to the loss of information in scenes with occlusions. Most existing methods for occluded ReID use 2D-based network structures to directly extract representations from 2D RGB (red, green, and blue) images, which can result in reduced performance in occluded scenes. However, since a person is a 3D non-grid object, learning semantic representations in a 2D space can limit the ability to accurately profile an occluded person. Therefore, it is crucial to explore alternative approaches that can effectively handle occlusions and leverage the full 3D nature of a person. To tackle these challenges, in this study, we employ a 3D view-based approach that fully utilizes the geometric information of 3D objects while leveraging advancements in 2D-based networks for feature extraction. Our study is the first to introduce a 3D view-based method in the areas of holistic and occluded ReID. To implement this approach, we propose a random rendering strategy that converts 2D RGB images into 3D multi-view images. We then use a 3D Multi-View Transformation Network for ReID (MV-ReID) to group and aggregate these images into a unified feature space. Compared to 2D RGB images, multi-view images can reconstruct occluded portions of a person in 3D space, enabling a more comprehensive understanding of occluded individuals. The experiments on benchmark datasets demonstrate that the proposed method achieves state-of-the-art results on occluded ReID tasks and exhibits competitive performance on holistic ReID tasks. These results also suggest that our approach has the potential to solve occlusion problems and contribute to the field of ReID. The source code and dataset are available at https://github.com/yuzaiyang123/MV-Reid. © 2023 Elsevier B.V.

Place, publisher, year, edition, pages
Amsterdam: Elsevier, 2024. Vol. 283, article id 111200
Keywords [en]
3D multi-view learning, Occluded person Re-identification
National Category
Computer Vision and Robotics (Autonomous Systems)
Identifiers
URN: urn:nbn:se:hh:diva-52204DOI: 10.1016/j.knosys.2023.111200Scopus ID: 2-s2.0-85177792581OAI: oai:DiVA.org:hh-52204DiVA, id: diva2:1818216
Note

Funding: The National Natural Science Foundation of China No. 62373343, Beijing Natural Science Foundation No. L233036.

Available from: 2023-12-08 Created: 2023-12-08 Last updated: 2023-12-08Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Tiwari, Prayag

Search in DiVA

By author/editor
Yu, ZaiyangTiwari, PrayagLi, LusiLi, WeijunNing, Xin
By organisation
School of Information Technology
In the same journal
Knowledge-Based Systems
Computer Vision and Robotics (Autonomous Systems)

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 37 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf