论文标题
重新连接还是densenet?引入密集的快捷方式以重新设置
ResNet or DenseNet? Introducing Dense Shortcuts to ResNet
论文作者
论文摘要
重新连接还是densenet?如今,大多数基于深度学习的方法是通过开创性骨干网络实施的,其中两个最著名的方法是Resnet和Densenet。尽管他们的竞争性能和极大的受欢迎程度,但他们俩都存在固有的缺点。对于Resnet而言,稳定训练的身份快捷方式还限制了其表示能力,而Densenet具有更高的能力,具有多层特征串联。但是,密集的串联引起了新的问题,即需要高GPU记忆和更多的训练时间。部分原因是,它不是Resnet和Densenet之间的微不足道选择。本文提供了一个统一的观点来分析它们,这有助于更好地理解其核心差异。我们进一步提出了密集的加权归一化快捷方式,以解决它们之间的困境。我们提出的密集快捷方式继承了Resnet和Densenet中简单设计的设计理念。在几个基准数据集上,实验结果表明,所提出的DSNET比RESNET取得了明显更好的结果,并且具有与Densenet相当的性能,但需要更少的计算资源。
ResNet or DenseNet? Nowadays, most deep learning based approaches are implemented with seminal backbone networks, among them the two arguably most famous ones are ResNet and DenseNet. Despite their competitive performance and overwhelming popularity, inherent drawbacks exist for both of them. For ResNet, the identity shortcut that stabilizes training also limits its representation capacity, while DenseNet has a higher capacity with multi-layer feature concatenation. However, the dense concatenation causes a new problem of requiring high GPU memory and more training time. Partially due to this, it is not a trivial choice between ResNet and DenseNet. This paper provides a unified perspective of dense summation to analyze them, which facilitates a better understanding of their core difference. We further propose dense weighted normalized shortcuts as a solution to the dilemma between them. Our proposed dense shortcut inherits the design philosophy of simple design in ResNet and DenseNet. On several benchmark datasets, the experimental results show that the proposed DSNet achieves significantly better results than ResNet, and achieves comparable performance as DenseNet but requiring fewer computation resources.