Speech Recognition and Coversational Interfaces Larry Rudolph content

  • Slides: 25
Download presentation
Speech Recognition and Coversational Interfaces Larry Rudolph (content adapted from Jim Glass) April 2006

Speech Recognition and Coversational Interfaces Larry Rudolph (content adapted from Jim Glass) April 2006 1 Pervasive Computing MIT 6. 883 SMA 5508 Spring 2006 Larry Rudolph

The Space of Recognition Domain Speaker Dependent Independent Transcription Dependent not interesting (training) Ultimate

The Space of Recognition Domain Speaker Dependent Independent Transcription Dependent not interesting (training) Ultimate Goal (requires Independent We are here knowledge)

The Space of Recognition • Speaker dependent • • • First train the system

The Space of Recognition • Speaker dependent • • • First train the system to recognize your speaking Better recognition rates -- can learn idiosynchroses Domain dependent: • Only recognize what is in the domain • • Better recognition rates Domain can be large. How is it specified? Pervasive Computing MIT 6. 883 SMA 5508 Spring 2006 Larry Rudolph

Why Speech? • • No special training -- naive users(? ) Leaves hands and

Why Speech? • • No special training -- naive users(? ) Leaves hands and eyes free -- but must know when to start recognition High data rate -- assuming low errors Inexpensive I/O -- microphone, speaker, button • • speaker needed for feedback Some things are easier to specify with speech Pervasive Computing MIT 6. 883 SMA 5508 Spring 2006 Larry Rudolph

Communication via Spoken Language Human Input Output Speech Recognition Synthesis Computer Text Generation Understanding

Communication via Spoken Language Human Input Output Speech Recognition Synthesis Computer Text Generation Understanding Meaning

Components of Conversational Systems Language Generation Dialogue Management Speech Synthesis Audio Database Speech Recognition

Components of Conversational Systems Language Generation Dialogue Management Speech Synthesis Audio Database Speech Recognition Context Resolution Language Understanding

Galaxy -- MIT SLS group • • SLS: Spoken Language Systems We will be

Galaxy -- MIT SLS group • • SLS: Spoken Language Systems We will be making use of some of there technology • • There are similar components developed by other groups (and some are public domain). The Galaxy System is organized around this cycle for conversational interfaces Pervasive Computing MIT 6. 883 SMA 5508 Spring 2006 Larry Rudolph

Components of MIT Conversational Systems GALAXY Language GENESIS Generation Dialogue Management Manager Speech ENVOICE

Components of MIT Conversational Systems GALAXY Language GENESIS Generation Dialogue Management Manager Speech ENVOICE Synthesis Audio Hub Speech SUMMIT Recognition Database Context Discourse Resolution Language TINA Understanding

Segment-Based Speech Recognition Waveform Frame-based measurements (every 5 ms) Segment network created by interconnecting

Segment-Based Speech Recognition Waveform Frame-based measurements (every 5 ms) Segment network created by interconnecting spectral landmarks p - k ax m - computers uw d x e r ao z d h a e that - talk Probabilistic search finds most likely phone & word strings k

Segment-Based Speech Recognition

Segment-Based Speech Recognition

Natural Language Understanding Some syntactic nodes carry semantic tags for creating semantic frame sentence

Natural Language Understanding Some syntactic nodes carry semantic tags for creating semantic frame sentence full_parse Clause: DISPLAY Topic: FLIGHT Predicate: FROM Topic: CITY Name: "Boston" Predicate: TO Topic: CITY Name: "Denver" command subject display show me topic predicate flight source destination flight_list from city to city flights from boston to denver

Dialogue Modeling Strategies • Effective conversational interface must incorporate extensive and complex dialogue modeling

Dialogue Modeling Strategies • Effective conversational interface must incorporate extensive and complex dialogue modeling • Conversational systems differ in the degree with which human or computer takes the initiative Computer Initiative • Computer maintains tight control • Human is highly restricted C: Please say the departure city. Human • Human takes complete control • Computer is totally passive H: I want to visit my grandmother. • The Galaxy System use a mixed initiative approach, where both the human & the computer play an active role

Different Roles of Dialogue Management • Pre-Retrieval: Ambiguous Input => Unique Query to DB

Different Roles of Dialogue Management • Pre-Retrieval: Ambiguous Input => Unique Query to DB U: C: I need a flight from Boston to San Francisco Did you say Boston or Austin? Boston, Massachusetts I need a date before I can access Travelocity Tomorrow Hold on while I retrieve the flights for you Clarification (recognition errors) Clarification (insufficient info) • Post-Retrieval: Multiple DB Retrievals => Unique Response C: I have found 10 flights meeting your specification. When would you like to leave? U: In the morning. Help the user narrow down the choices C: Do you have a preferred airline? U: United C: I found two non-stop United flights leaving in the morning…

Concatenative Speech Synthesis • Output waveform generated by concatenating segments of pre-recorded speech corpus.

Concatenative Speech Synthesis • Output waveform generated by concatenating segments of pre-recorded speech corpus. • Concatenation at phrase, word or sub-word level. Synthesis Examples The third ad is a 1996 black Acura Integra with 45380 miles. The price is 8970 dollars. Please call (404) 399 -7682. labyrinth abracadabra obligatory laboratory compassion disputed cedar city since giant since computer science Continental flight 4695 from Greensboro is expected in Halifax at 10: 08 pm local time.

Multilingual Conversational Interfaces • Adopts an interlingua approach for multilingual humanmachine interactions • Applications:

Multilingual Conversational Interfaces • Adopts an interlingua approach for multilingual humanmachine interactions • Applications: – Mu. Xing: Mandarin system for weather information – Mokusei: Japanese system for weather information – Spanish systems are also under development – New speech-to-speech translation work (Phrasebook) Language Generation Text-to-Speech Conversion Audio I/O Server Servers Models Rules Dialogue Management Hub Application Back-end Application Back-end Discourse Resolution Speech Recognition Models Language Understanding Models Rules Language Transparent Language Independent Language Dependent

Bilingual Jupiter Demonstration

Bilingual Jupiter Demonstration

Multi-modal Conversational Interfaces • Typing, pointing, clicking can augment/complement speech • A picture (or

Multi-modal Conversational Interfaces • Typing, pointing, clicking can augment/complement speech • A picture (or a map) is worth a thousand words • Applications: – Web. Galaxy – Allows typing and clicking – Includes map-based navigation – With display – Embedded in a web browser – Current exhibit at MIT Museum SPEECH RECOGNITION HANDWRITING RECOGNITION LANGUAGE UNDERSTANDING GESTURE RECOGNITION MOUTH & EYES TRACKING meaning

Web. Galaxy Demonstration

Web. Galaxy Demonstration

Delegating Tasks to Computers • Many information related activities can be done off line

Delegating Tasks to Computers • Many information related activities can be done off line • Off-line delegation frees the user to attend to other matters • Application: Orion system – Task Specification: User interacts with Orion to specify a task “Call me every morning at 6 and tell me the weather in Boston. ” “Send me e-mail any time between 4 and 6 p. m. if the traffic on Route 93 is at a standstill. ” – Task Execution: Orion leverages existing infrastructure to support interaction with humans – Event Notification: Orion calls back to deliver information

Audio Visual Integration • Audio and visual signals both contain information about: – Identity

Audio Visual Integration • Audio and visual signals both contain information about: – Identity of the person: Who is talking? – Linguistic message: What’s (s)he saying? – Emotion, mood, stress, etc. : How does (s)he feel? • The two channels of information – Are often inter-related – Are often complementary – Must be consistent • Integration of these cues can lead to enhanced capabilities for future human computer interfaces

Audio Visual Symbiosis Personal Identity Speaker ID Acoustic Signal Robust ASR Speech Recognition Linguistic

Audio Visual Symbiosis Personal Identity Speaker ID Acoustic Signal Robust ASR Speech Recognition Linguistic Message Lip/Mouth Reading Face ID Robust Person ID Visual Signal Acoustic Paraling. Detection Visual Paraling. Detection Robust Paralinguistic Detection Paralinguistic Information

Multi-modal Interfaces: Beyond Clicking • Inputs need to be understood in the proper context

Multi-modal Interfaces: Beyond Clicking • Inputs need to be understood in the proper context Are there any over here? What does he mean by “any, ” and what is he pointing at? Does this mean “yes, ” “one, ” or something else? • Timing information is a useful way to relate inputs Move this one over there Where is she looking or pointing at while saying “this” and “there”?

Multi-modal Fusion: Initial Progress • All multi-modal inputs are synchronized – Speech recognizer generates

Multi-modal Fusion: Initial Progress • All multi-modal inputs are synchronized – Speech recognizer generates absolute times for words – Mouse and gesture movements generate {x, y, t} triples – Network Time Protocol (NTP) is used for msec time resolution • Speech understanding constrains gesture interpretation – Initial work identifies an object or a location from gesture inputs – Speech constrains what, when, and how items are resolved – Object resolution also depends on information from application Speech: Pointing: “Move this one over here” (object) (location) time

Multi-modal Demonstration • Manipulating planets in a solar-system application • Created w. Speech. Builder

Multi-modal Demonstration • Manipulating planets in a solar-system application • Created w. Speech. Builder utility with small changes • Gestures from vision (Darrell & Demirdjien)

Multi-modal Demonstration • Manipulating planets in a solar-system application • Created w. Speech. Builder

Multi-modal Demonstration • Manipulating planets in a solar-system application • Created w. Speech. Builder utility with small changes • Gestures from vision (Darrell & Demirdjien)