论文标题
多目标多摄像机车辆跟踪的半自动数据注释系统
Semi-automatic Data Annotation System for Multi-Target Multi-Camera Vehicle Tracking
论文作者
论文摘要
多目标多摄像机跟踪(MTMCT)在智能视频分析,监视视频检索和其他应用程序方案中起着重要作用。如今,基于深度学习的MTMCT一直是主流,并且在跟踪准确性和效率方面取得了令人着迷的改进。但是,根据我们的调查,缺乏关注现实应用程序方案的数据集限制了当前基于学习的MTMCT模型的进一步改进。具体而言,基于学习的MTMCT模型通过通用数据集培训通常无法在现实应用程序方案中获得令人满意的结果。在此激励的情况下,本文提出了一个半自动数据注释系统,以促进现实世界中的MTMCT数据集建立。提出的系统首先采用了基于深度学习的单相机轨迹生成方法来自动从监视视频中提取轨迹。随后,该系统在以下手动跨相机轨迹匹配过程中提供了建议列表。建议列表是根据侧面信息(包括相机位置,时间戳关系和背景场景)生成的。在实验阶段,广泛的结果进一步证明了拟议系统的效率。
Multi-target multi-camera tracking (MTMCT) plays an important role in intelligent video analysis, surveillance video retrieval, and other application scenarios. Nowadays, the deep-learning-based MTMCT has been the mainstream and has achieved fascinating improvements regarding tracking accuracy and efficiency. However, according to our investigation, the lacking of datasets focusing on real-world application scenarios limits the further improvements for current learning-based MTMCT models. Specifically, the learning-based MTMCT models training by common datasets usually cannot achieve satisfactory results in real-world application scenarios. Motivated by this, this paper presents a semi-automatic data annotation system to facilitate the real-world MTMCT dataset establishment. The proposed system first employs a deep-learning-based single-camera trajectory generation method to automatically extract trajectories from surveillance videos. Subsequently, the system provides a recommendation list in the following manual cross-camera trajectory matching process. The recommendation list is generated based on side information, including camera location, timestamp relation, and background scene. In the experimental stage, extensive results further demonstrate the efficiency of the proposed system.