1 / 1

Using Model Trees For Evaluating Dialog Error Conditions Based on Acoustic Information

Using Model Trees For Evaluating Dialog Error Conditions Based on Acoustic Information. Goal. Use model trees for evaluating user utterances for response to system error. Input: acoustic features from user’s speech signal. Output: a measure representing user activation.

morrison
Download Presentation

Using Model Trees For Evaluating Dialog Error Conditions Based on Acoustic Information

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Using Model Trees For Evaluating Dialog Error Conditions Based on Acoustic Information Goal • Use model trees for evaluating user utterances for response to system error. • Input: acoustic features from user’s speech signal. • Output: a measure representing user activation. • Develop an online, objective, human-centered evaluation metric for spoken dialog systems. Abe Kazemzadeh, Sungbok Lee, and Shrikanth Narayanan Computer Science, Electrical Engineering, and Linguistics SAIL Lab @ Viterbi School of Engineering University of Southern California Motivation • Errors are a prevalent phenomenon in spoken dialog systems. • Evaluate and optimize of dialog systems. • Obtain feedback from user behavior. • Synthesize low-level features into one, real-valued measurement of a user’s activation. Results Histograms of the model tree output for the whole corpus (histogram 1), for error responses (histogram 2), and for non-error responses (histogram 3). Lower left plot shows the precision and recall. Data • Communicator Travel Planning Systems, June 2000 recordings. • Annotated to describe the way that users become aware of and react to errors. • 141 dialogs, 2586 utterances. Model Trees • Machine learning technique, similar to decision trees and model trees. • Outputs a continuous, real-valued number based on a linear regression model for each leaf node. Best correlation with user surveys occurred when model tree output sums were normalized for dialog length and when only the highest 30% were considered. Methodology • Feature extraction: • Train by using annotated data: if there is an error response, set model tree target to 1, else, 0. • Analysis Conclusion • Overall ability to pick out error responses is 65% precision, 63% recall. • The model tree approach allows for a threshold that can shift preferents toward precision or recall. • Correlation between model tree analysis and survey results was moderate. • Different questions showed different levels of correlation. • Model tree output can be interpreted as an indicator of user state and can show a dialog activation landscape which can be used in user emotion tracking, e.g., to identify dialog hotspots. • Future work will aim to further this study by: • Testing other methods of synthesizing lower level features, in particular, Bayesian networks • Examining other corpora. Currently analyzing All My Sons radio play. • Example

More Related