R2Net: Residual refinement network for salient object detection

作者:

Highlights:

摘要

The multi-scale features and the fusion strategy of contextual features are the keys to salient object detection task. Previous multi-scale-based works often overlooked the completeness of features when acquiring multi-scale features. Moreover, the decoders were hard to accurately capture the salient object and refine the object's boundaries simultaneously when in a complex environment, which leads to unsatisfactory saliency maps. To address the above problems, we present a Residual Refinement Network (R2Net), which is composed of the Residual Pyramid Module (RPM), the Residual Fusion Module (RFM) and the Feature Optimize Module (FOM), for salient object detection. RPM integrates different feature information through different receptive fields, can not only obtain multi-scale information but also retain the local details information of features. RFM can better locate salient objects and refine the boundaries through the interweaving and fusion of multi-layer features. And FOM is designed to further refine the fused features. Furthermore, we propose a Structural Polishing (SP) loss, which better guides the network through pixel-level supervision, global supervision and boundary supervision to generate high-quality saliency maps with fine boundaries. Experimental results on 6 benchmark datasets demonstrate that the proposed method has superior performance compared with 18 state-of-the-art methods. The code and results of our method are available at https://github.com/zhangjin12138/R2Net

论文关键词:Deep learning,Salient object detection,Multi-scale feature,Feature fusion

论文评审过程:Received 19 December 2021, Accepted 19 February 2022, Available online 25 February 2022, Version of Record 8 March 2022.

论文官网地址:https://doi.org/10.1016/j.imavis.2022.104423