Item Infomation
Title: | Omnidirectional Feature Learning for Person Re-Identification |
Authors: | Wu, D. |
Participants: | Yang, H. Huang, D. Yuan, C. Qin, X. Zhao, Y. Zhao, X. Sun, J. |
Issue Date: | 2019 |
Publisher: | IEEE Xplore |
Series/Report no.: | IEEE Access, (2019), Vol 7, pp28402-28411 |
Abstract: | Person re-identi cation (PReID) has received increasing attention due to it being an important role in intelligent surveillance. Many state-of-the-art PReID methods are part-based deep models. Most of these models focus on learning the part feature representation of a person's body from the horizontal direction. However, the feature representation of the body from the vertical direction is usually ignored. In addition, the relationships between these part features and different feature channels are not considered. In this paper, we introduce a multi-branch deep model for PReID. Speci cally, the model consists of ve branches. Among the ve branches, two branches learn the part features with spatial information from horizontal and vertical orientations; one branch aims to learn the interdependencies between different feature channels generated by the last convolution layer of the backbone network; the remaining two branches are identi cation and triplet sub-networks in which the discriminative global feature and a corresponding measurement can be learned simultaneously. All ve branches can improve the quality of representation learning. We conduct extensive comparison experiments on three benchmarks, including Market-1501, CUHK03, and DukeMTMC-reID. The proposed deep framework outperforms other competitive state-of-the-art methods. The code is available at https://github.com/caojunying/person-reidenti cation. |
URI: | http://tailieuso.tlu.edu.vn/handle/DHTL/9859 |
Appears in Collections: | Tài liệu hỗ trợ nghiên cứu khoa học |
ABSTRACTS VIEWS
14
VIEWS & DOWNLOAD
4
Files in This Item:
Bạn đọc là cán bộ, giáo viên, sinh viên của Trường Đại học Thuỷ Lợi cần đăng nhập để Xem trực tuyến/Tải về
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.