|Xiaoxiao Guo||IBM Research|
|Hui Wu||IBM Research|
|Yu Cheng||Microsoft AI & Research|
|Gerald Tesauro||IBM TJ Watson Research Center|
|Rogerio Feris||IBM Research AI|
Existing methods for interactive image retrieval have demonstrated the merit of integrating user feedback, improving retrieval results.In this paper, the authors introduce a new approach to interactive image search that enables users to provide feedback via natural language, allowing for more natural and effective interaction.
Existing methods for interactive image retrieval have demonstrated the merit of integrating user feedback, improving retrieval results. However, most current systems rely on restricted forms of user feedback, such as binary relevance responses, or feedback based on a fixed set of relative attributes, which limits their impact. In this paper, we introduce a new approach to interactive image search that enables users to provide feedback via natural language, allowing for more natural and effective interaction. We formulate the task of dialog-based interactive image retrieval as a reinforcement learning problem, and reward the dialog system for improving the rank of the target image during each dialog turn. To mitigate the cumbersome and costly process of collecting human-machine conversations as the dialog system learns, we train our system with a user simulator, which is itself trained to describe the differences between target and candidate images. The efficacy of our approach is demonstrated in a footwear retrieval application. Experiments on both simulated and real-world data show that 1) our proposed learning framework achieves better accuracy than other supervised and reinforcement learning baselines and 2) user feedback based on natural language rather than pre-specified attributes leads to more effective retrieval results, and a more natural and expressive communication interface.