Learning discriminative and generalizable features with multi-branch for person re-identification
Finer-grained local features play a supplementary role in the description of pedestrian global features, and the combination of them has been an essential solution to improve discriminative performances in person re-identification (PReID) tasks. The existing part-based methods mostly extract representational semantic parts according to human visual habits or some prior knowledge and focus on spatial partition strategies but ignore the significant influence of channel information on PReID task. So, we proposed an end-to-end multi-branch network architecture (MCSN) jointing multi-level global fusion features, channel features and spatial features in this paper to better learn more diverse and discriminative pedestrian features. It is worth noting that the effect of multi-level fusion features on the performance of the model is taken into account when extracting global features. In addition, to enhance the stability of model training and the generalization ability of the model, the BNNeck and the joint loss function strategy are applied to all vector representation branches. Extensive comparative evaluations are conducted on three mainstream image-based evaluation protocols, including Market-1501, DukeMTMC-ReID and MSMT17, to validate the advantages of our proposed model, which outperforms previous state-of-the-art in ReID tasks.