Dialogue Control Algorithm for Ambient Intelligence based on Partially Observable Markov Decision Processes

msra

引用 24|浏览7
暂无评分
摘要
From the viewpoint of supporting users' natural dialogue communication with conversational agents, their dialogue management has to determine any agent's action, based on probabilistic methods derived from noisy data through sensors in the real world. We believe unique Partially Observable Markov Decision Processes (POMDPs) should be applied to such action control systems. The agents must flexibly choose their actions to reach a state suitable for the users while re- taining as many statistical characteristics of the data as possible. We offer two technical points to resolve this issue. One is the automatic acquisition of POMDPs ¡ú state transition probabilities through DBNs with a large amount of dialogue data, and the other is applying re- wards from the emission probabilities of agent actions into POMDPs' reinforcement learning. This paper proposes a method to simultane- ously achieve purpose-oriented and stochastic naturalness-oriented ac- tion controls. Our experimental results demonstrate the effectiveness of our framework, which shows that the agent can generate both actions without being locked into either of them.
更多
查看译文
关键词
hidden markov model hmm,dynamic bayesian net- work dbn,dialogue management,reinforcement learning rl,expectation-maximization em algorithm,agent,multi-modal interaction,partially observable markov decision process pomdp
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要