MAIN: Multi-Attention Instance Network for video segmentation

作者:

Highlights:

摘要

Instance-level video segmentation requires a solid integration of spatial and temporal information. However, current methods rely mostly on domain-specific information (online learning) to produce accurate instance-level segmentations. We propose a novel approach that relies exclusively on the integration of generic spatio-temporal attention cues. Our strategy, named Multi-Attention Instance Network (MAIN), overcomes challenging segmentation scenarios over arbitrary videos without modeling sequence- or instance-specific knowledge. We design MAIN to segment multiple instances in a single forward pass, and optimize it with a novel loss function that favors class agnostic predictions and assigns instance-specific penalties. We achieve state-of-the-art performance on the challenging Youtube-VOS dataset and benchmark, improving the unseen Jaccard and F-Metric by 6.8% and 12.7% respectively, while operating at real-time (30.3 FPS).

论文关键词:

论文评审过程:Received 13 February 2020, Revised 1 June 2021, Accepted 4 June 2021, Available online 24 June 2021, Version of Record 3 July 2021.

论文官网地址:https://doi.org/10.1016/j.cviu.2021.103240