Kisser et al., 2022 - Google Patents
Erroneous reactions of voice assistants “in the wild”–first analysesKisser et al., 2022
View PDF- Document ID
- 1288076087851087657
- Author
- Kisser L
- Siegert I
- Publication year
- Publication venue
- Elektronische Sprachsignalverarbeitung
External Links
Snippet
Voice assistants are increasingly dominating everyday life and represent an easy way to perform various tasks with minimal effort. The areas of application for voice assistants are diverse and range from answering simple information questions to processing complex …
- 238000006243 chemical reaction 0 title abstract description 11
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/1822—Parsing for meaning understanding
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/183—Speech classification or search using natural language modelling using context dependencies, e.g. language models
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L2015/088—Word spotting
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/065—Adaptation
- G10L15/07—Adaptation to the speaker
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/30—Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7139706B2 (en) | System and method of developing automatic speech recognition vocabulary for voice activated services | |
Litman et al. | Designing and evaluating an adaptive spoken dialogue system | |
US20030061029A1 (en) | Device for conducting expectation based mixed initiative natural language dialogs | |
Skantze | Exploring human error handling strategies: Implications for spoken dialogue systems | |
CN112232083A (en) | Man-machine conversation spoken language evaluation system | |
Kisser et al. | Erroneous reactions of voice assistants “in the wild”–first analyses | |
Komatani et al. | User modeling in spoken dialogue systems to generate flexible guidance | |
Chukharev‐Hudilainen et al. | The development and evaluation of interactional competence elicitor for oral language assessments | |
Hone et al. | Designing habitable dialogues for speech-based interaction with computers | |
Johnstone et al. | There was a long pause: influencing turn-taking behaviour in human-human and human-computer spoken dialogues | |
Norman et al. | Studying alignment in a collaborative learning activity via automatic methods: The link between what we say and do | |
Skidmore et al. | Using Alexa for flashcard-based learning | |
KR100450019B1 (en) | Method of service for english training of interactive voice response using internet | |
Gunkel | Computational interpersonal communication: Communication studies and spoken dialogue systems | |
Phukon et al. | Can VUI turn-taking entrain user behaviours? voice user interfaces that disallow overlapping speech present turn-taking challenges | |
Rose | Fluidity: Real-time feedback on acoustic measures of second language speech fluency | |
Spitz | Collection and analysis of data from real users: Implications for speech recognition/understanding systems | |
Patel et al. | Google duplex-a big leap in the evolution of artificial intelligence | |
Wolters et al. | Making it easier for older people to talk to smart homes: The effect of early help prompts | |
Wirén et al. | Experiences of an in-service Wizard-of-Oz data collection for the deployment of a call-routing application | |
Heeman et al. | Can overhearers predict who will speak next? | |
Scovell et al. | Impact of accuracy and latency on mean opinion scores for speech recognition solutions | |
Sloan et al. | Emotional response language education: a first ‘in-the-wild’evaluation | |
Ondáš et al. | Analysis of turn-taking in the Slovak interview corpus | |
Passonneau et al. | Seeing what you said: How wizards use voice search results |