Learning goal-oriented visual dialog agents: Imitating and surpassing analytic experts

Yen Wei Chang, Wen-Hsiao Peng

研究成果: Conference contribution同行評審

3 引文 斯高帕斯(Scopus)

摘要

This paper tackles the problem of learning a questioner in the goal-oriented visual dialog task. Several previous works adopt model-free reinforcement learning. Most pretrain the model from a finite set of human-generated data. We argue that using limited demonstrations to kick-start the questioner is insufficient due to the large policy search space. Inspired by a recently proposed information theoretic approach, we develop two analytic experts to serve as a source of high-quality demonstrations for imitation learning. We then take advantage of reinforcement learning to refine the model towards the goal-oriented objective. Experimental results on the GuessWhat?! dataset show that our method has the combined merits of imitation and reinforcement learning, achieving the state-of-the-art performance.

原文English
主出版物標題Proceedings - 2019 IEEE International Conference on Multimedia and Expo, ICME 2019
發行者IEEE Computer Society
頁面520-525
頁數6
ISBN(電子)9781538695524
DOIs
出版狀態Published - 1 7月 2019
事件2019 IEEE International Conference on Multimedia and Expo, ICME 2019 - Shanghai, China
持續時間: 8 7月 201912 7月 2019

出版系列

名字Proceedings - IEEE International Conference on Multimedia and Expo
2019-July
ISSN(列印)1945-7871
ISSN(電子)1945-788X

Conference

Conference2019 IEEE International Conference on Multimedia and Expo, ICME 2019
國家/地區China
城市Shanghai
期間8/07/1912/07/19

指紋

深入研究「Learning goal-oriented visual dialog agents: Imitating and surpassing analytic experts」主題。共同形成了獨特的指紋。

引用此