Efficient Dialog Policy Learning via Positive Memory Retention.

2018 IEEE Spoken Language Technology Workshop (SLT)(2018)

引用 8|浏览52757
暂无评分
摘要
This paper is concerned with the training of recurrent neural networks as goal-oriented dialog agents using reinforcement learning. Training such agents with policy gradients typically requires a large amount of samples. However, the collection of the required data in form of conversations between chatbots and human agents is time-consuming and expensive. To mitigate this problem, we describe an efficient policy gradient method using positive memory retention, which significantly increases the sample-efficiency. We show that our method is 10 times more sample-efficient than policy gradients in extensive experiments on a new synthetic number guessing game. Moreover, in a real-word visual object discovery game, the proposed method is twice as sample-efficient as policy gradients and shows state-of-the-art performance.
更多
查看译文
关键词
Trajectory,Proposals,Games,Training,Task analysis,Monte Carlo methods,Mathematical model
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要