Image retrieval with mixed initiative and multimodal feedback

作者:

Highlights:

摘要

How would you search for a unique, flamboyant shoe that a friend wore and you want to buy? What if you did not take a picture? Existing approaches propose interactive image search, but they either entrust the user with taking the initiative to provide informative feedback, or give all control to the system which determines informative questions to ask. Instead, we propose a mixed-initiative framework where both the user and system can be active participants, depending on whose input will be more beneficial for obtaining high-quality search results. We develop a reinforcement learning approach which dynamically decides which of four interaction opportunities to give to the user: drawing a sketch, marking images as relevant or not, providing free-form attribute feedback, or answering attribute-based questions. By allowing these four options, our system optimizes both the informativeness of feedback, and the ability of the user to explore the data, allowing faster image retrieval. We outperform five baselines on three datasets under extensive settings.

论文关键词:

论文评审过程:Received 4 July 2020, Revised 19 March 2021, Accepted 22 March 2021, Available online 26 March 2021, Version of Record 13 April 2021.

论文官网地址:https://doi.org/10.1016/j.cviu.2021.103204