Wider or Deeper: Revisiting the ResNet Model for Visual Recognition

作者:

Highlights:

• We further develop the unravelled view of ResNets, which helps us better understand their behaviours. We demonstrate this in the context of a training process, which is the key difference from the original version 1.

• We propose a group of relatively shallow convolutional networks based on our new understanding. Some of them perform comparably with the state-of-the-art approaches on the ImageNet classification dataset 2.

• We evaluate the impact of using different networks on the performance of semantic image segmentation, and show these networks, as pre-trained features, can boost existing algorithms a lot.

摘要

•We further develop the unravelled view of ResNets, which helps us better understand their behaviours. We demonstrate this in the context of a training process, which is the key difference from the original version 1.•We propose a group of relatively shallow convolutional networks based on our new understanding. Some of them perform comparably with the state-of-the-art approaches on the ImageNet classification dataset 2.•We evaluate the impact of using different networks on the performance of semantic image segmentation, and show these networks, as pre-trained features, can boost existing algorithms a lot.

论文关键词:Image classification,Semantic segmentation,Residual network

论文评审过程:Received 24 June 2018, Revised 4 November 2018, Accepted 4 January 2019, Available online 6 January 2019, Version of Record 28 January 2019.

论文官网地址:https://doi.org/10.1016/j.patcog.2019.01.006