Interactive convolutional learning for visual tracking

作者:

Highlights:

摘要

Recent years have witnessed significant improvements of ensemble trackers based on independent models. However, existing ensemble trackers only combine the responses of independent models and pay less attention to the learning process, which hinders their performance from further improvements. To this end, we propose an interactive learning framework to strengthen the independent models in the learning process. Specifically, in the interactive network, we force convolutional filter models to interact with each other by sharing their responses during the learning. The interaction between the convolutional filter models can mine hard samples and prevent easy samples from overwhelming them, which improve their discriminative capacity. In addition, to achieve a more accurate target location, we develop a fusion mechanism based on the confidences of the independent predictions. We evaluate the proposed method on five public datasets including OTB-2013, OTB-2015, VOT-2016, VOT-2017, Temple-Color-128, and LaSOT. The comprehensive experimental results show that the proposed algorithm performs favorably against state-of-the-art methods.

论文关键词:Visual tracking,Convolutional neural network,Interactive learning

论文评审过程:Received 4 September 2020, Revised 16 December 2020, Accepted 21 December 2020, Available online 7 January 2021, Version of Record 15 January 2021.

论文官网地址:https://doi.org/10.1016/j.knosys.2020.106724