论文标题
Tive:用于识别视频实例分割错误的工具箱
TIVE: A Toolbox for Identifying Video Instance Segmentation Errors
论文作者
论文摘要
自首次提出以来,视频实例细分(VIS)任务吸引了广泛的研究人员对建筑建模的关注以提高性能。尽管在在线和离线范式中取得的巨大进步,但仍然没有足够的方法来识别模型错误并区分方法之间的差异,并且方法在识别各种时间长度的对象实例时正确地反映了模型的性能。更重要的是,随着任务所需的基本模型能力,空间分割和时间关联在评估和相互作用机制中仍被研究了。在本文中,我们介绍了Tive,这是一个用于识别视频实例分割错误的工具箱。通过直接操作输出预测文件,TIVE定义了孤立的错误类型和权重每种类型对MAP的损坏,以区分模型字符。通过分解时空维度中的定位质量,可以揭示模型对空间分割和时间关联的潜在缺点。 Tive还可以报告实际应用程序的实例时间长度。我们通过工具箱进行广泛的实验,以进一步说明空间分割和时间关联如何相互影响。我们希望对Tive进行分析可以为研究人员提供更多的见解,从而指导社区促进更有意义的视频实例细分探索。提出的工具箱可在https://github.com/wenhe-jia/tive上找到。
Since first proposed, Video Instance Segmentation(VIS) task has attracted vast researchers' focus on architecture modeling to boost performance. Though great advances achieved in online and offline paradigms, there are still insufficient means to identify model errors and distinguish discrepancies between methods, as well approaches that correctly reflect models' performance in recognizing object instances of various temporal lengths remain barely available. More importantly, as the fundamental model abilities demanded by the task, spatial segmentation and temporal association are still understudied in both evaluation and interaction mechanisms. In this paper, we introduce TIVE, a Toolbox for Identifying Video instance segmentation Errors. By directly operating output prediction files, TIVE defines isolated error types and weights each type's damage to mAP, for the purpose of distinguishing model characters. By decomposing localization quality in spatial-temporal dimensions, model's potential drawbacks on spatial segmentation and temporal association can be revealed. TIVE can also report mAP over instance temporal length for real applications. We conduct extensive experiments by the toolbox to further illustrate how spatial segmentation and temporal association affect each other. We expect the analysis of TIVE can give the researchers more insights, guiding the community to promote more meaningful explorations for video instance segmentation. The proposed toolbox is available at https://github.com/wenhe-jia/TIVE.