Co-segmentation inspired attention module for video-based computer vision tasks

作者:

Highlights:

• Exploring the applicability of “Co-segmentation” concept in video-based tasks.

• Generic “Co-Segmentation inspired Attention Module” (COSAM) to plug-in to any CNN.

• The COSAM module captures task-specific salient regions in an end-to-end manner.

• The COSAM module aids interpretability through its spatial attention mask.

• Experiments on three video-based tasks show the effectiveness of the COSAM module.

摘要

•Exploring the applicability of “Co-segmentation” concept in video-based tasks.•Generic “Co-Segmentation inspired Attention Module” (COSAM) to plug-in to any CNN.•The COSAM module captures task-specific salient regions in an end-to-end manner.•The COSAM module aids interpretability through its spatial attention mask.•Experiments on three video-based tasks show the effectiveness of the COSAM module.

论文关键词:Attention,Co-segmentation,Person re-ID,Video-captioning,Video classification

论文评审过程:Received 26 November 2021, Revised 2 August 2022, Accepted 3 August 2022, Available online 9 August 2022, Version of Record 23 August 2022.

论文官网地址:https://doi.org/10.1016/j.cviu.2022.103532