Interfaces between Speech and Non-Speech Audio Technology. Michael Pucher (FTW Vienna, ICSI Berkeley). Contents. Text-to-Speech Synthesis (TTS) Automatic Speech Recognition (ASR, STT) Dialog Systems Multimodal Mobile Applications Resources. Non-linguistic. Sound signals. Music.
Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.
Interfaces between Speech and Non-Speech Audio Technology
Michael Pucher (FTW Vienna, ICSI Berkeley)
Perspectival, spatial cues
Affective states and attitudes
Pragmatics and discourse
Structural prosodic elements
Lexical semantics and syntax
Command & control
Broadcast news transcription
Speaker dependent or speaker independent
„um ah to san francisco from new york“
1. Apply SLM
2. Apply grammar on results of SLM
Paolo Massimino : Loquendo S.p.A.
From Marked Text to Mixed Speech and Sound
Bob Cooper : Avaya Corporation
A Case Study on the planned and actual Use of Auditory Feedback and Audio Cues in the Realization of a Personal Virtual Assistant
<prosody contour="(0%,+20Hz) (10%,+30%) (40%,+10Hz)">
Any female voice here.
<voice age="6"> A female child voice here. </voice>
The Importance of Multimodality for Mobile Applications
Minhua Ma : University of Ulster Paul Mc Kevitt : University of Ulster
Lexical Semantics and Auditory Display in Virtual Storytelling
Thank you for your attention