Deeply-Learned Spatial Alignment for Person Re-Identification

A large class of Person Re-identification (ReID) approaches identify pedestrians with the TriHard loss. Though the TriHard loss is a robust ReID method, pose variance and viewpoint in pedestrians constrain the performance. To address this problem, we introduce a spatial transformer network (STN) to...

Full description

Bibliographic Details
Main Authors: Dongyue Chen, Peng Chen, Xiaosheng Yu, Mengjiao Cao, Tong Jia
Format: Article
Language:English
Published: IEEE 2019-01-01
Series:IEEE Access
Subjects:
STN
Online Access:https://ieeexplore.ieee.org/document/8856191/
id doaj-b5308dce5ff54b65a6e22c9f79488eb2
record_format Article
spelling doaj-b5308dce5ff54b65a6e22c9f79488eb22021-04-05T17:34:30ZengIEEEIEEE Access2169-35362019-01-01714368414369210.1109/ACCESS.2019.29453538856191Deeply-Learned Spatial Alignment for Person Re-IdentificationDongyue Chen0https://orcid.org/0000-0003-0673-6767Peng Chen1https://orcid.org/0000-0001-6545-4941Xiaosheng Yu2Mengjiao Cao3Tong Jia4https://orcid.org/0000-0003-1424-798XCollege of Information Science and Engineering, Northeastern University, Shenyang, ChinaCollege of Information Science and Engineering, Northeastern University, Shenyang, ChinaFaculty of Robot Science and Engineering, Northeastern University, Shenyang, ChinaCollege of Information Science and Engineering, Northeastern University, Shenyang, ChinaCollege of Information Science and Engineering, Northeastern University, Shenyang, ChinaA large class of Person Re-identification (ReID) approaches identify pedestrians with the TriHard loss. Though the TriHard loss is a robust ReID method, pose variance and viewpoint in pedestrians constrain the performance. To address this problem, we introduce a spatial transformer network (STN) to align pedestrians. Then, we illustrate the generality of the STN module in pose variance problem through the evaluations on feature representation network (FRN) like VGG, ResNet and DenseNet architectures respectively. Furthermore, based on the evaluation results, we propose a robust and high-performance ReID model which consists of the STN module, DenseNet backbone and TriHard loss. And finally, we prove that our ReID model is whole differentiable by formula derivation, therefore achieving an end-to-end high-performance ReID system. The experiments show that our ReID system outperforms the state-of-art methods on Market-1501, DukeMTMC-reID and CUHK03 datasets.https://ieeexplore.ieee.org/document/8856191/ReIDSTNalignmentTriHardpose varianceviewpoint
collection DOAJ
language English
format Article
sources DOAJ
author Dongyue Chen
Peng Chen
Xiaosheng Yu
Mengjiao Cao
Tong Jia
spellingShingle Dongyue Chen
Peng Chen
Xiaosheng Yu
Mengjiao Cao
Tong Jia
Deeply-Learned Spatial Alignment for Person Re-Identification
IEEE Access
ReID
STN
alignment
TriHard
pose variance
viewpoint
author_facet Dongyue Chen
Peng Chen
Xiaosheng Yu
Mengjiao Cao
Tong Jia
author_sort Dongyue Chen
title Deeply-Learned Spatial Alignment for Person Re-Identification
title_short Deeply-Learned Spatial Alignment for Person Re-Identification
title_full Deeply-Learned Spatial Alignment for Person Re-Identification
title_fullStr Deeply-Learned Spatial Alignment for Person Re-Identification
title_full_unstemmed Deeply-Learned Spatial Alignment for Person Re-Identification
title_sort deeply-learned spatial alignment for person re-identification
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2019-01-01
description A large class of Person Re-identification (ReID) approaches identify pedestrians with the TriHard loss. Though the TriHard loss is a robust ReID method, pose variance and viewpoint in pedestrians constrain the performance. To address this problem, we introduce a spatial transformer network (STN) to align pedestrians. Then, we illustrate the generality of the STN module in pose variance problem through the evaluations on feature representation network (FRN) like VGG, ResNet and DenseNet architectures respectively. Furthermore, based on the evaluation results, we propose a robust and high-performance ReID model which consists of the STN module, DenseNet backbone and TriHard loss. And finally, we prove that our ReID model is whole differentiable by formula derivation, therefore achieving an end-to-end high-performance ReID system. The experiments show that our ReID system outperforms the state-of-art methods on Market-1501, DukeMTMC-reID and CUHK03 datasets.
topic ReID
STN
alignment
TriHard
pose variance
viewpoint
url https://ieeexplore.ieee.org/document/8856191/
work_keys_str_mv AT dongyuechen deeplylearnedspatialalignmentforpersonreidentification
AT pengchen deeplylearnedspatialalignmentforpersonreidentification
AT xiaoshengyu deeplylearnedspatialalignmentforpersonreidentification
AT mengjiaocao deeplylearnedspatialalignmentforpersonreidentification
AT tongjia deeplylearnedspatialalignmentforpersonreidentification
_version_ 1721539337183559680