Multi-domain Conversation Quality Evaluation via User Satisfaction Estimation. (arXiv:1911.08567v1 [cs.LG])
An automated metric to evaluate dialogue quality is vital for optimizing data driven dialogue management. The common approach of relying on explicit user feedback during a conversation is intrusive and sparse. Current models to estimate user satisfaction use limited feature sets and employ annotation schemes with limited generalizability to conversations spanning multiple domains. To address…