论文标题
观察不变的步态识别,并细心地重复学习部分表示
View-Invariant Gait Recognition with Attentive Recurrent Learning of Partial Representations
论文作者
论文摘要
步态识别是指根据步行过程中从身体运动中获得的特征来识别个体。尽管步态识别方面取得了进步,但深度学习,数据获取和外观的变化,即相机角度,主题姿势,遮挡和衣服,是具有挑战性的因素,需要考虑实现准确的步态识别系统。在本文中,我们提出了一个网络,该网络首先学会从框架级卷积特征提取步态卷积能图(GCEM)。然后,它采用双向复发性神经网络从GCEM的分裂箱中学习,从而利用了学到的部分时空表示之间的关系。然后,我们使用注意力机制选择性地专注于重要的经常学习的部分表示,因为不同情况下的身份信息可能位于不同的GCEM垃圾箱中。我们提出的模型已在两个大规模的CASIA-B和OU-MVLP步态数据集上使用四种不同的测试协议进行了广泛的测试,并已与许多最先进和基线解决方案进行了比较。此外,已经进行了全面的实验,以研究在六个不同合成的闭塞的情况下我们模型的鲁棒性。实验结果表明,我们提出的方法的优越性优于最先进的方法,尤其是在遇到不同衣服和携带条件的情况下。结果还表明,与最先进的方法相比,我们的模型在不同的遮挡方面更有鲁棒。
Gait recognition refers to the identification of individuals based on features acquired from their body movement during walking. Despite the recent advances in gait recognition with deep learning, variations in data acquisition and appearance, namely camera angles, subject pose, occlusions, and clothing, are challenging factors that need to be considered for achieving accurate gait recognition systems. In this paper, we propose a network that first learns to extract gait convolutional energy maps (GCEM) from frame-level convolutional features. It then adopts a bidirectional recurrent neural network to learn from split bins of the GCEM, thus exploiting the relations between learned partial spatiotemporal representations. We then use an attention mechanism to selectively focus on important recurrently learned partial representations as identity information in different scenarios may lie in different GCEM bins. Our proposed model has been extensively tested on two large-scale CASIA-B and OU-MVLP gait datasets using four different test protocols and has been compared to a number of state-of-the-art and baseline solutions. Additionally, a comprehensive experiment has been performed to study the robustness of our model in the presence of six different synthesized occlusions. The experimental results show the superiority of our proposed method, outperforming the state-of-the-art, especially in scenarios where different clothing and carrying conditions are encountered. The results also revealed that our model is more robust against different occlusions as compared to the state-of-the-art methods.