Get 20M+ Full-Text Papers For Less Than $1.50/day. Start a 14-Day Trial for You or Your Team.

Learn More →

Effective handling of dialogue state in the hidden information state POMDP-based dialogue manager

Effective handling of dialogue state in the hidden information state POMDP-based dialogue manager Effective Handling of Dialogue State in the Hidden Information State POMDP-Based Dialogue Manager MILICA GASIC and STEVE YOUNG, University of Cambridge Effective dialogue management is critically dependent on the information that is encoded in the dialogue state. In order to deploy reinforcement learning for policy optimization, dialogue must be modeled as a Markov Decision Process. This requires that the dialogue state must encode all relevent information obtained during the dialogue prior to that state. This can be achieved by combining the user goal, the dialogue history, and the last user action to form the dialogue state. In addition, to gain robustness to input errors, dialogue must be modeled as a Partially Observable Markov Decision Process (POMDP) and hence, a distribution over all possible states must be maintained at every dialogue turn. This poses a potential computational limitation since there can be a very large number of dialogue states. The Hidden Information State model provides a principled way of ensuring tractability in a POMDP-based dialogue model. The key feature of this model is the grouping of user goals into partitions that are dynamically built during the dialogue. In this article, we extend this model further to incorporate the notion http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png ACM Transactions on Speech and Language Processing (TSLP) Association for Computing Machinery

Effective handling of dialogue state in the hidden information state POMDP-based dialogue manager

Loading next page...
 
/lp/association-for-computing-machinery/effective-handling-of-dialogue-state-in-the-hidden-information-state-3g3xqVvYbv
Publisher
Association for Computing Machinery
Copyright
Copyright © 2011 by ACM Inc.
ISSN
1550-4875
DOI
10.1145/1966407.1966409
Publisher site
See Article on Publisher Site

Abstract

Effective Handling of Dialogue State in the Hidden Information State POMDP-Based Dialogue Manager MILICA GASIC and STEVE YOUNG, University of Cambridge Effective dialogue management is critically dependent on the information that is encoded in the dialogue state. In order to deploy reinforcement learning for policy optimization, dialogue must be modeled as a Markov Decision Process. This requires that the dialogue state must encode all relevent information obtained during the dialogue prior to that state. This can be achieved by combining the user goal, the dialogue history, and the last user action to form the dialogue state. In addition, to gain robustness to input errors, dialogue must be modeled as a Partially Observable Markov Decision Process (POMDP) and hence, a distribution over all possible states must be maintained at every dialogue turn. This poses a potential computational limitation since there can be a very large number of dialogue states. The Hidden Information State model provides a principled way of ensuring tractability in a POMDP-based dialogue model. The key feature of this model is the grouping of user goals into partitions that are dynamically built during the dialogue. In this article, we extend this model further to incorporate the notion

Journal

ACM Transactions on Speech and Language Processing (TSLP)Association for Computing Machinery

Published: May 1, 2011

There are no references for this article.