How would you search for a unique, fashionable shoe that a friend wore and you want to buy, but you didn't take a picture? Existing approaches propose interactive image search as a promising venue. However, they either entrust the user with taking the initiative to provide informative feedback, or give all control to the system which determines informative questions to ask. Instead, we propose a mixed-initiative framework where both the user and system can be active participants, depending on whose initiative will be more beneficial for obtaining high-quality search results. We develop a reinforcement learning approach which dynamically decides which of three interaction opportunities to give to the user: drawing a sketch, providing free-form attribute feedback, or answering attribute-based questions. By allowing these three options, our system optimizes both the informativeness and exploration capabilities allowing faster image retrieval. We outperform three baselines on three datasets and extensive experimental settings.
翻译:如何寻找一个独特的、时尚的鞋, 朋友可以穿戴而你想要买的鞋子, 但你没有拍照? 现有的方法建议交互式图像搜索是一个充满希望的场所。 但是, 它们要么委托用户主动提供信息反馈, 要么将所有控制权赋予决定信息提问的系统。 相反, 我们提议了一个混合倡议框架, 让用户和系统都能成为积极的参与者, 取决于谁的主动行动更有利于获得高质量的搜索结果。 我们开发了一个强化学习方法, 以动态方式决定给用户的三个互动机会中的哪一个: 绘制草图, 提供自由形式属性反馈, 或回答基于属性的问题。 通过允许这三个选项, 我们的系统优化了信息性和探索能力, 以便更快地检索图像。 我们比三个数据集和广泛的实验设置的三条基线要强。