论文标题

对预训练的深卷卷神经网络的自适应剥削,用于鲁棒的视觉跟踪

Adaptive Exploitation of Pre-trained Deep Convolutional Neural Networks for Robust Visual Tracking

论文作者

Marvasti-Zadeh, Seyed Mojtaba, Ghanei-Yakhdan, Hossein, Kasaei, Shohreh

论文摘要

由于通过多层非线性转换的自动特征提取过程,基于深度学习的视觉跟踪器最近在挑战视觉跟踪目的的挑战场景方面取得了巨大的成功。尽管许多跟踪器都利用了预先训练的卷积神经网络(CNN)的特征图,但仍未完全比较选择不同模型并利用其特征图的各种组合的效果。据我们所知,所有这些方法都使用固定数量的卷积特征图,而无需考虑在跟踪过程中可能发生的场景属性(例如遮挡,变形和快速运动)。作为预示,本文提出了基于可以利用具有不同拓扑结构的CNN模型的方法自适应判别相关过滤器(DCF)。首先,本文对四个常用的CNN模型进行了全面分析,以确定每个模型的最佳特征图。其次,借助分析结果作为属性词典,提出了对深度特征的自适应开发,以提高视觉跟踪器在视频特征方面的准确性和鲁棒性。第三,在各种跟踪数据集以及具有相似体系结构的CNN模型上验证了所提出方法的概括。最后,广泛的实验结果证明了与最先进的视觉跟踪方法相比,提出的自适应方法的有效性。

Due to the automatic feature extraction procedure via multi-layer nonlinear transformations, the deep learning-based visual trackers have recently achieved great success in challenging scenarios for visual tracking purposes. Although many of those trackers utilize the feature maps from pre-trained convolutional neural networks (CNNs), the effects of selecting different models and exploiting various combinations of their feature maps are still not compared completely. To the best of our knowledge, all those methods use a fixed number of convolutional feature maps without considering the scene attributes (e.g., occlusion, deformation, and fast motion) that might occur during tracking. As a pre-requisition, this paper proposes adaptive discriminative correlation filters (DCF) based on the methods that can exploit CNN models with different topologies. First, the paper provides a comprehensive analysis of four commonly used CNN models to determine the best feature maps of each model. Second, with the aid of analysis results as attribute dictionaries, adaptive exploitation of deep features is proposed to improve the accuracy and robustness of visual trackers regarding video characteristics. Third, the generalization of the proposed method is validated on various tracking datasets as well as CNN models with similar architectures. Finally, extensive experimental results demonstrate the effectiveness of the proposed adaptive method compared with state-of-the-art visual tracking methods.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源