论文标题

单眼3D对象检测的均匀丢失

Homography Loss for Monocular 3D Object Detection

论文作者

Gu, Jiaqi, Wu, Bojian, Fan, Lubin, Huang, Jianqiang, Cao, Shen, Xiang, Zhiyu, Hua, Xian-Sheng

论文摘要

单程3D对象检测是自动驾驶中的重要任务。但是,大多数当前方法将场景中的每个3D对象视为一个独立的训练样本,同时忽略了它们固有的几何关系,因此不可避免地导致缺乏利用空间约束。在本文中,我们提出了一种新颖的方法,该方法将所有对象考虑在内,并探索它们的相互关系,以帮助更好地估计3D框。此外,由于当前2D检测更可靠,因此我们还研究了如何使用检测到的2D框作为指导,以全局约束相应预测的3D框的优化。为此,提出了一个可区分的损失函数,称为同构损失,以实现目标,该目标利用了2D和3D信息,旨在通过全局约束来平衡不​​同对象之间的位置关系,以便获得更准确的预测3D框。借助简洁的设计,我们的损失功能是通用的,并且可以插入任何成熟的单眼3D检测器中,同时显着提高了其基线的性能。实验表明,与其他最先进的方法相比,我们的方法在Kitti 3D数据集上具有很大的利润。

Monocular 3D object detection is an essential task in autonomous driving. However, most current methods consider each 3D object in the scene as an independent training sample, while ignoring their inherent geometric relations, thus inevitably resulting in a lack of leveraging spatial constraints. In this paper, we propose a novel method that takes all the objects into consideration and explores their mutual relationships to help better estimate the 3D boxes. Moreover, since 2D detection is more reliable currently, we also investigate how to use the detected 2D boxes as guidance to globally constrain the optimization of the corresponding predicted 3D boxes. To this end, a differentiable loss function, termed as Homography Loss, is proposed to achieve the goal, which exploits both 2D and 3D information, aiming at balancing the positional relationships between different objects by global constraints, so as to obtain more accurately predicted 3D boxes. Thanks to the concise design, our loss function is universal and can be plugged into any mature monocular 3D detector, while significantly boosting the performance over their baseline. Experiments demonstrate that our method yields the best performance (Nov. 2021) compared with the other state-of-the-arts by a large margin on KITTI 3D datasets.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源