US7548859B2 - Method and system for assisting users in interacting with multi-modal dialog systems - Google Patents
Method and system for assisting users in interacting with multi-modal dialog systems Download PDFInfo
- Publication number
- US7548859B2 US7548859B2 US11/196,112 US19611205A US7548859B2 US 7548859 B2 US7548859 B2 US 7548859B2 US 19611205 A US19611205 A US 19611205A US 7548859 B2 US7548859 B2 US 7548859B2
- Authority
- US
- United States
- Prior art keywords
- modal
- user
- dialog
- utterances
- grammar
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/038—Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/451—Execution arrangements for user interfaces
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/038—Indexing scheme relating to G06F3/038
- G06F2203/0381—Multimodal input, i.e. interface arrangements enabling the user to issue commands by simultaneous use of input devices of different nature, e.g. voice plus gesture on digitizer
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/183—Speech classification or search using natural language modelling using context dependencies, e.g. language models
- G10L15/19—Grammatical context, e.g. disambiguation of the recognition hypotheses based on word sequence rules
- G10L15/193—Formal grammars, e.g. finite state automata, context free grammars or word networks
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/226—Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics
- G10L2015/228—Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics of application context
Definitions
- This invention is in the field of software, and more specifically in the field of software that assists users in interacting with multi-modal dialog systems.
- Dialog systems are systems that allow a user to interact with a computer system to perform tasks such as retrieving information, conducting transactions, and other such problem solving tasks.
- a dialog system can use several modalities for interaction. Examples of modalities include speech, gesture, touch, and handwriting.
- User-computer interactions in dialog systems are enhanced by employing multiple modalities in a coordinated manner. Dialog systems using multiple modalities for user-computer interaction are referred to as multi-modal dialog systems.
- the user interacts with a multi-modal dialog system by using multiple modalities that are coordinated as a dialog-based user interface.
- the user interface is referred to as a multi-modal dialog interface.
- a set of interactions of the user with the dialog system is referred to as a dialog. Each interaction is known as a turn of the dialog.
- the information provided by either the user or the dialog system is referred to as a context of the dialog.
- a user should know which modalities to use and the method of using them, in a coordinated manner in different turns of a dialog. Further, the utilization of the multi-modal user interface of a dialog system can be difficult to use, due either to the lack of experience with using the system, or to changes in the availability and capability of modalities during a dialog.
- dialog-based systems do not help a user to use different modalities in a coordinated manner. Further, these dialog systems do not take into account modality availability information. Moreover, the dialog systems do not provide a response to the questions asked by users in a dialog related to the use of different modalities.
- FIG.1 is a block diagram of a multi-modal dialog system, in accordance with some embodiments of the present invention.
- FIG. 2 is a block diagram of an input processor in a multi-modal dialog system, in accordance with some embodiments of the present invention
- FIG. 3 is a flowchart illustrating a method for assisting a user in interacting with a multi-modal dialog system, in accordance with an embodiment of the present invention
- FIG. 4 is a flowchart illustrating a method for assisting a user in interacting with the multi-modal dialog system, in accordance with another embodiment of the present invention.
- FIG. 5 is a block diagram of an exemplary multi-modal grammar rule, in accordance with one embodiment of the present invention.
- FIG. 6 is a block diagram of an electronic equipment, in accordance with some embodiments of the present invention.
- FIG. 1 a block diagram shows a representative environment in which the present invention may be practiced, in accordance with some embodiments of the present invention.
- the representative environment includes an input-output module 102 and a multi-modal dialog system 104 .
- the input-output module 102 receives user inputs and displays system outputs.
- the input-output module 102 is a user interface such as a computer monitor, a touch screen, or a keyboard.
- a user interacts with the multi-modal dialog system 104 through the input-output module 102 .
- the user provides an input by using various modalities through one or more input-output modules, of which one input-output module 102 is shown.
- Examples of various modalities include, but are not limited to, voice, gesture, and handwriting.
- the user interacts with the multi-modal dialog system 104 by using one or more modalities in a coordinated manner.
- the multi-modal dialog system 104 uses a combined interpretation of user inputs provided through various modalities. For example, a user can say, “I want to find all the restaurants here”, touching an area on a map displayed on a touch screen. In this case, the multi-modal dialog system 104 combines the inputs provided through speech and gesture, to generate a combined interpretation.
- the user may ask the multi-modal dialog system 104 how various modalities can be used in a coordinated manner. This type of question is referred to as a “What Can I Do?” (WCID) question.
- WID What Can I Do?”
- the multi-modal dialog system 104 includes an input processor 106 and an output generator 108 .
- the input processor 106 interprets and processes the inputs as well the WCID questions, and provides the interpretation and answers to the output generator 108 .
- the output generator 108 receives the WCID answers and presents them to the input-output module 102 .
- a block diagram shows the input processor 106 in the multi-modal dialog system 104 for assisting a user, in accordance with some embodiments of the present invention.
- the input processor 106 includes a Multi-Modal Input Fusion (MMIF) component 202 , a dialog manager 204 , a visual context manager 206 , a multi-modal utterance generator 208 , and a multi-modal utterance ranker 210 .
- the MMIF component 202 accepts and interprets a user WCID question in a dialog.
- the MMIF component 202 maintains a record of modality capability information.
- the MMIF component 202 maintains a record of the capabilities of the modalities that were used by the user at each previous turn of the dialog. Further, the MMIF component 202 updates the record of modality capability information at the turn of the dialog.
- the dialog manager 204 generates a multi-modal grammar 212 , based on a current context of the dialog. The information provided by a user in a previous turn or turns of the dialog is referred to as the current context of the dialog.
- a multi-modal grammar refers to a set of rules that are used to interpret a user input in the next turn of the dialog. Further, the dialog manager 204 maintains and updates the contexts of the dialog.
- the visual context manager 206 maintains a list of a set of concepts and objects that are visible to the user, as well as a list of objects that are in the dialog focus but are not visible in the current turn of the dialog. For example, the visual context manager 206 can maintain a place on a map displayed to the user, or a place mentioned in the previous turn of the dialog that is not shown on the user interface of the input-output module 102 .
- the multi-modal utterance generator 208 accepts the multi-modal grammar 212 and the visual context information from the dialog manager 204 and the visual context manager 206 , respectively.
- the multi-modal utterance generator 208 generates one or more user multi-modal utterances based on a user WCID question, the multi-modal grammar 212 and the visual context information.
- a user multi-modal utterance may be a natural language sentence that the multi-modal dialog system can articulate to a user, using speech, with multi-modal coordination between certain words in the sentence.
- a multi-modal coordination specifies the sequence of modalities to be used and a temporal order between the modalities.
- a temporal order specifies the time gap between the use of the modalities.
- the multi-modal utterance ranker 210 ranks one or more user multi-modal utterances, based on the current context of the dialog, a task model, and the modality availability information.
- a task model is a data structure used to model a task that is to be performed. The task model elaborates on the knowledge necessary for completing the task.
- One or more ranked user multi-modal utterances are conveyed to the user through the output generator 108 .
- a flowchart illustrates a method for assisting a user in interacting with the multi-modal dialog system 104 , in accordance with an embodiment of the present invention.
- the MMIF component 202 interprets a question, such as a WCID question, from a user interacting with the multi-modal dialog system 104 .
- a question such as a WCID question
- one or more user multi-modal utterances are generated by the multi-modal utterance generator 208 .
- the one or more user multi-modal user utterances are based on the WCID question and the multi-modal grammar 212 generated by the dialog manager 204 .
- the one or more user multi-modal utterances are further based on the visual context information.
- the multi-modal grammar 212 is generated by the dialog manager 204 based on the current context of the dialog. Thereafter, at step 306 , the one or more user multi-modal utterances are conveyed to the user through the output generator 108 .
- a flowchart illustrates a method for assisting a user in interacting with the multi-modal dialog system 104 , in accordance with some embodiments of the present invention.
- the MMIF component 202 interprets a question, such as a WCID question, from a user interacting with the multi-modal dialog system 104 .
- the multi-modal grammar 212 is generated by the dialog manager 204 . Generation of the multi-modal grammar 212 is based on the current context of the dialog. In other words, it is based on information provided by the user in the previous turns of the dialog.
- the multi-modal grammar 212 defines a sequence of words that a user may provide in the multi-modal inputs in the next turn of the dialog.
- An exemplary multi-modal grammar rule is explained in conjunction with FIG. 5 .
- a visual context of the multi-modal dialog is obtained.
- the visual context refers to objects that are visible to the user on the user interface of the input-output module 102 .
- the visual context information may also include objects that are in the dialog focus but are not currently visible. For example, the visual context may include a place mentioned by the user in the previous turn of a dialog.
- one or more user multi-modal utterances are generated, based on the multi-modal grammar 212 and the WCID question.
- the one or more user multi-modal utterances are generated by parsing the multi-modal grammar 212 .
- the one or more user multi-modal utterances are de-referenced, based on the visual context information. De-referencing is carried out to include in a user utterance instructions for any non-speech input that the user has to enter in coordination with the speech input.
- one or more user multi-modal utterances are ranked and a list is generated, based on the rankings. In one embodiment of the invention, the ranking is based on a task model, the modality preferences of the user, and modality availability information.
- the ranked one or more user multi-modal utterances are conveyed to the user through the output generator 108 .
- the output reflects the modalities to be used for providing user inputs.
- the input-output module 102 may direct the user by saying, “You can circle an area on the map”, using speech while showing a circle with ink on the user interface of the input-output module 102 to indicate the area.
- the one or more user multi-modal utterances are generated, based on the multi-modal grammar 212 .
- the dialog manager 204 In a multi-modal dialog, at a particular turn, the dialog manager 204 expects the user to make a request to go to a certain place. Accordingly, the dialog manager 204 generates the multi-modal grammar 212 .
- the multi-modal grammar 212 is illustrated in FIG. 5
- a block diagram represents the multi-modal grammar 212 , generated by the dialog manager 204 , in accordance with some embodiments of the invention.
- the multi-modal grammar 212 is a network of a plurality of non-terminals, hereinafter referred to as non-terminals 502 , 504 , 506 , and 508 .
- Each non-terminal represents semantic information relevant to a dialog.
- non-terminals may represent a place by a combination of street name and suburb name.
- the modality-capability information provided by the MMIF component 202 is also attached to each non-terminal.
- the multimodal grammar 212 further comprises one or more connections, such as a connection 510 and a connection 514 .
- the connection 510 indicates that the modality used is touch, whereas the connection 514 shows that the modality used is speech.
- the dialog information is represented according to defined rules attached to the non-terminals 502 , 504 , 506 and 508 , and the connections 510 , 512 and 514 .
- An example of a rule is the modality capability rule which specifies how the modalities may be used by a user to provide an input.
- a rule is a sequence of non-terminals, to be provided using the same modality.
- speech may be used for the sequence of non-terminals 502 , 504 and 506 .
- touch may generate the semantic information for both place name and suburb, as represented by the non-terminal 508 and the connection 512 .
- the multi-modal grammar 212 is elaborated on with the help of a vocabulary of relevant modalities.
- the user can provide an input by saying, “Go”, followed by the street name and then the suburb name. Alternately, the user can say “Go”, and indicate the street and suburb by touching a point on a map displayed on the user interface.
- the user multi-modal utterances are generated. For the example given above, a user multi-modal utterance would be “you can say ‘go to this place’, followed by touching a point on the map to identify the place.”
- user multi-modal utterances are de-referenced.
- the above-mentioned user multi-modal utterance is modified as “you can say go to the restaurant to go to the Chipotle restaurant.”
- the user's WCID question includes information about the particular modality through which the user wants to provide the input. For example, if the user wants to use a gesture modality to provide the next input, then the multi-modal utterance generator 208 generates the user multi-modal utterances according to the selected modality.
- the electronic equipment 600 includes a means for interpreting 602 , a means for generating multi-modal grammar 604 , a means for storing 606 , a means for generating user multi-modal utterances 608 , and a means for conveying 610 .
- the means for interpreting 602 accepts and interprets user inputs and WCID questions.
- the information provided by the user in the user inputs is referred to as a current context of the dialog.
- the means for interpreting 602 records the modality availability information and the modality combinations used by the user in previous turns of the dialog.
- the means for generating multi-modal grammar 604 obtains interpretations of the WCID question from the means for interpreting 602 .
- the means for generating multi-modal grammar 604 generates the multi-modal grammar 212 , based on the current context of the dialog and the modality availability information.
- the means for storing 606 stores a visual context of the dialog.
- the means for generating user multi-modal utterances 608 generates one or more user multi-modal utterances, based on the multi-modal grammar 612 and the visual context of the dialog. Further, the means for generating user multi-modal utterances 608 ranks one or more user multi-modal utterances, based on the current context of the dialog, the modality availability information, and the modality preferences of the user.
- the means for conveying 610 conveys the ranked user multi-modal utterances to the user.
- the method for assisting a user in interacting with a multi-modal dialog system described herein may comprise one or more conventional processors and unique stored program instructions that control the one or more processors to implement some, most, or all of the functions described herein; as such, the functions of generating multi-modal interpretations and generating combined semantic meaning representations may be interpreted as being steps of the method.
- the same functions could be implemented by a state machine that has no stored program instructions, in which each function or some combinations of certain portions of the functions are implemented as custom logic. A combination of the two approaches could be used. Thus, methods and means for performing these functions have been described herein.
- the method for generating user multi-modal utterances can be used in multi-modal devices, to assist users in interacting with the multi-modal devices, for example, a handset which a user can input with speech, a keypad, or a combination of both.
- the method can also be used in multi-modal applications of personal communication systems (PCS). It can also be used in commercial equipment, ranging from extremely complicated computers and robots to simple pieces of test equipment. Further, its range of applications extends to all areas where access to information and browsing takes place with the aid of a multi-modal interface.
- PCS personal communication systems
- the terms “comprises”, “comprising”, “includes”, “including” or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus.
- a “set” as used herein, means a non-empty set (i.e., for the sets defined herein, comprising at least one member).
- the term “another”, as used herein, is defined as at least a second or more.
- the term “having”, as used herein, is defined as comprising.
- the term “coupled”, as used herein with reference to electro-optical technology, is defined as connected, although not necessarily directly, and not necessarily mechanically.
- program as used herein, is defined as a sequence of instructions designed for execution on a computer system.
- a “program”, or “computer program”, may include a subroutine, a function, a procedure, an object method, an object implementation, an executable application, an applet, a serviet, a source code, an object code, a shared library/dynamic load library and/or other sequence of instructions designed for execution on a computer system. It is further understood that the use of relational terms, if any, such as first and second, top and bottom, and the like are used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Multimedia (AREA)
- Acoustics & Sound (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
Claims (13)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/196,112 US7548859B2 (en) | 2005-08-03 | 2005-08-03 | Method and system for assisting users in interacting with multi-modal dialog systems |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/196,112 US7548859B2 (en) | 2005-08-03 | 2005-08-03 | Method and system for assisting users in interacting with multi-modal dialog systems |
Publications (2)
Publication Number | Publication Date |
---|---|
US20070033526A1 US20070033526A1 (en) | 2007-02-08 |
US7548859B2 true US7548859B2 (en) | 2009-06-16 |
Family
ID=37718963
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/196,112 Active 2026-08-18 US7548859B2 (en) | 2005-08-03 | 2005-08-03 | Method and system for assisting users in interacting with multi-modal dialog systems |
Country Status (1)
Country | Link |
---|---|
US (1) | US7548859B2 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080133228A1 (en) * | 2006-11-30 | 2008-06-05 | Rao Ashwin P | Multimodal speech recognition system |
US20120010875A1 (en) * | 2002-11-28 | 2012-01-12 | Nuance Communications Austria Gmbh | Classifying text via topical analysis, for applications to speech recognition |
US9830912B2 (en) | 2006-11-30 | 2017-11-28 | Ashwin P Rao | Speak and touch auto correction interface |
US9922640B2 (en) | 2008-10-17 | 2018-03-20 | Ashwin P Rao | System and method for multimodal utterance detection |
US10114676B2 (en) | 2015-05-05 | 2018-10-30 | Microsoft Technology Licensing, Llc | Building multimodal collaborative dialogs with task frames |
US11645036B2 (en) | 2019-01-23 | 2023-05-09 | Samsung Electronics Co., Ltd. | Electronic device and operating method for providing feedback information in response to user input |
Families Citing this family (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070157093A1 (en) * | 2005-12-30 | 2007-07-05 | Patrick Karcher | Systems and methods for adaptive help mechanisms for a user |
US8073681B2 (en) | 2006-10-16 | 2011-12-06 | Voicebox Technologies, Inc. | System and method for a cooperative conversational voice user interface |
US7818176B2 (en) | 2007-02-06 | 2010-10-19 | Voicebox Technologies, Inc. | System and method for selecting and presenting advertisements based on natural language processing of voice-based input |
US8140335B2 (en) | 2007-12-11 | 2012-03-20 | Voicebox Technologies, Inc. | System and method for providing a natural language voice user interface in an integrated voice navigation services environment |
US9305548B2 (en) | 2008-05-27 | 2016-04-05 | Voicebox Technologies Corporation | System and method for an integrated, multi-modal, multi-device natural language voice services environment |
WO2010045375A1 (en) * | 2008-10-14 | 2010-04-22 | Honda Motor Co., Ltd. | Improving dialog coherence using semantic features |
US8326637B2 (en) | 2009-02-20 | 2012-12-04 | Voicebox Technologies, Inc. | System and method for processing multi-modal device interactions in a natural language voice services environment |
US8689037B2 (en) | 2009-12-16 | 2014-04-01 | Qualcomm Incorporated | System and method for asynchronously and independently controlling core clocks in a multicore central processing unit |
US20110145559A1 (en) * | 2009-12-16 | 2011-06-16 | Thomson Steven S | System and method for controlling central processing unit power with guaranteed steady state deadlines |
US9563250B2 (en) | 2009-12-16 | 2017-02-07 | Qualcomm Incorporated | System and method for controlling central processing unit power based on inferred workload parallelism |
US8650426B2 (en) * | 2009-12-16 | 2014-02-11 | Qualcomm Incorporated | System and method for controlling central processing unit power in a virtualized system |
US9128705B2 (en) * | 2009-12-16 | 2015-09-08 | Qualcomm Incorporated | System and method for controlling central processing unit power with reduced frequency oscillations |
US9104411B2 (en) | 2009-12-16 | 2015-08-11 | Qualcomm Incorporated | System and method for controlling central processing unit power with guaranteed transient deadlines |
US8909962B2 (en) * | 2009-12-16 | 2014-12-09 | Qualcomm Incorporated | System and method for controlling central processing unit power with guaranteed transient deadlines |
US9176572B2 (en) | 2009-12-16 | 2015-11-03 | Qualcomm Incorporated | System and method for controlling central processing unit power with guaranteed transient deadlines |
US8775830B2 (en) * | 2009-12-16 | 2014-07-08 | Qualcomm Incorporated | System and method for dynamically controlling a plurality of cores in a multicore central processing unit based on temperature |
WO2016044321A1 (en) | 2014-09-16 | 2016-03-24 | Min Tang | Integration of domain information into state transitions of a finite state transducer for natural language processing |
WO2016044290A1 (en) | 2014-09-16 | 2016-03-24 | Kennewick Michael R | Voice commerce |
CN107003999B (en) | 2014-10-15 | 2020-08-21 | 声钰科技 | System and method for subsequent response to a user's prior natural language input |
US10431214B2 (en) | 2014-11-26 | 2019-10-01 | Voicebox Technologies Corporation | System and method of determining a domain and/or an action related to a natural language input |
US10614799B2 (en) | 2014-11-26 | 2020-04-07 | Voicebox Technologies Corporation | System and method of providing intent predictions for an utterance prior to a system detection of an end of the utterance |
WO2018023106A1 (en) | 2016-07-29 | 2018-02-01 | Erik SWART | System and method of disambiguating natural language processing requests |
CN116383365B (en) * | 2023-06-01 | 2023-09-08 | 广州里工实业有限公司 | Learning material generation method and system based on intelligent manufacturing and electronic equipment |
Citations (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5878274A (en) * | 1995-07-19 | 1999-03-02 | Kabushiki Kaisha Toshiba | Intelligent multi modal communications apparatus utilizing predetermined rules to choose optimal combinations of input and output formats |
US5890122A (en) * | 1993-02-08 | 1999-03-30 | Microsoft Corporation | Voice-controlled computer simulateously displaying application menu and list of available commands |
US5983190A (en) * | 1997-05-19 | 1999-11-09 | Microsoft Corporation | Client server animation system for managing interactive user interface characters |
US6094635A (en) * | 1997-09-17 | 2000-07-25 | Unisys Corporation | System and method for speech enabled application |
US6182046B1 (en) | 1998-03-26 | 2001-01-30 | International Business Machines Corp. | Managing voice commands in speech applications |
US6188985B1 (en) * | 1997-01-06 | 2001-02-13 | Texas Instruments Incorporated | Wireless voice-activated device for control of a processor-based host system |
US6208971B1 (en) * | 1998-10-30 | 2001-03-27 | Apple Computer, Inc. | Method and apparatus for command recognition using data-driven semantic inference |
US6208972B1 (en) * | 1998-12-23 | 2001-03-27 | Richard Grant | Method for integrating computer processes with an interface controlled by voice actuated grammars |
US6456972B1 (en) * | 1998-09-30 | 2002-09-24 | Scansoft, Inc. | User interface for speech recognition system grammars |
US20030061029A1 (en) * | 2001-08-29 | 2003-03-27 | Efraim Shaket | Device for conducting expectation based mixed initiative natural language dialogs |
US20030093419A1 (en) * | 2001-08-17 | 2003-05-15 | Srinivas Bangalore | System and method for querying information using a flexible multi-modal interface |
US20030234818A1 (en) | 2002-06-21 | 2003-12-25 | Schmid Philipp Heinz | Speech platform architecture |
US20040006475A1 (en) | 2002-07-05 | 2004-01-08 | Patrick Ehlen | System and method of context-sensitive help for multi-modal dialog systems |
US6735592B1 (en) * | 2000-11-16 | 2004-05-11 | Discern Communications | System, method, and computer program product for a network-based content exchange system |
-
2005
- 2005-08-03 US US11/196,112 patent/US7548859B2/en active Active
Patent Citations (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5890122A (en) * | 1993-02-08 | 1999-03-30 | Microsoft Corporation | Voice-controlled computer simulateously displaying application menu and list of available commands |
US5878274A (en) * | 1995-07-19 | 1999-03-02 | Kabushiki Kaisha Toshiba | Intelligent multi modal communications apparatus utilizing predetermined rules to choose optimal combinations of input and output formats |
US6188985B1 (en) * | 1997-01-06 | 2001-02-13 | Texas Instruments Incorporated | Wireless voice-activated device for control of a processor-based host system |
US5983190A (en) * | 1997-05-19 | 1999-11-09 | Microsoft Corporation | Client server animation system for managing interactive user interface characters |
US6094635A (en) * | 1997-09-17 | 2000-07-25 | Unisys Corporation | System and method for speech enabled application |
US6182046B1 (en) | 1998-03-26 | 2001-01-30 | International Business Machines Corp. | Managing voice commands in speech applications |
US6456972B1 (en) * | 1998-09-30 | 2002-09-24 | Scansoft, Inc. | User interface for speech recognition system grammars |
US6208971B1 (en) * | 1998-10-30 | 2001-03-27 | Apple Computer, Inc. | Method and apparatus for command recognition using data-driven semantic inference |
US6208972B1 (en) * | 1998-12-23 | 2001-03-27 | Richard Grant | Method for integrating computer processes with an interface controlled by voice actuated grammars |
US6735592B1 (en) * | 2000-11-16 | 2004-05-11 | Discern Communications | System, method, and computer program product for a network-based content exchange system |
US20030093419A1 (en) * | 2001-08-17 | 2003-05-15 | Srinivas Bangalore | System and method for querying information using a flexible multi-modal interface |
US20030061029A1 (en) * | 2001-08-29 | 2003-03-27 | Efraim Shaket | Device for conducting expectation based mixed initiative natural language dialogs |
US20030234818A1 (en) | 2002-06-21 | 2003-12-25 | Schmid Philipp Heinz | Speech platform architecture |
US20040006475A1 (en) | 2002-07-05 | 2004-01-08 | Patrick Ehlen | System and method of context-sensitive help for multi-modal dialog systems |
Non-Patent Citations (1)
Title |
---|
Hirohiko Sagawa, Correction Grammars for Error Handling in a Speech Dialog System , Jun. 30 2004,Carnegie mellon University, pp. 1-4. * |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120010875A1 (en) * | 2002-11-28 | 2012-01-12 | Nuance Communications Austria Gmbh | Classifying text via topical analysis, for applications to speech recognition |
US8612209B2 (en) * | 2002-11-28 | 2013-12-17 | Nuance Communications, Inc. | Classifying text via topical analysis, for applications to speech recognition |
US8965753B2 (en) | 2002-11-28 | 2015-02-24 | Nuance Communications, Inc. | Method to assign word class information |
US9996675B2 (en) | 2002-11-28 | 2018-06-12 | Nuance Communications, Inc. | Method to assign word class information |
US10515719B2 (en) | 2002-11-28 | 2019-12-24 | Nuance Communications, Inc. | Method to assign world class information |
US10923219B2 (en) | 2002-11-28 | 2021-02-16 | Nuance Communications, Inc. | Method to assign word class information |
US20080133228A1 (en) * | 2006-11-30 | 2008-06-05 | Rao Ashwin P | Multimodal speech recognition system |
US8355915B2 (en) * | 2006-11-30 | 2013-01-15 | Rao Ashwin P | Multimodal speech recognition system |
US9830912B2 (en) | 2006-11-30 | 2017-11-28 | Ashwin P Rao | Speak and touch auto correction interface |
US9922640B2 (en) | 2008-10-17 | 2018-03-20 | Ashwin P Rao | System and method for multimodal utterance detection |
US10114676B2 (en) | 2015-05-05 | 2018-10-30 | Microsoft Technology Licensing, Llc | Building multimodal collaborative dialogs with task frames |
US11645036B2 (en) | 2019-01-23 | 2023-05-09 | Samsung Electronics Co., Ltd. | Electronic device and operating method for providing feedback information in response to user input |
Also Published As
Publication number | Publication date |
---|---|
US20070033526A1 (en) | 2007-02-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7548859B2 (en) | Method and system for assisting users in interacting with multi-modal dialog systems | |
US11532306B2 (en) | Detecting a trigger of a digital assistant | |
EP3701520B1 (en) | Multi-turn canned dialog | |
US10445429B2 (en) | Natural language understanding using vocabularies with compressed serialized tries | |
US10482874B2 (en) | Hierarchical belief states for digital assistants | |
US10332518B2 (en) | User interface for correcting recognition errors | |
US7177815B2 (en) | System and method of context-sensitive help for multi-modal dialog systems | |
US7451088B1 (en) | System and method of handling problematic input during context-sensitive help for multi-modal dialog systems | |
JP6204982B2 (en) | Contextual query tuning using natural motion input | |
CN110019752A (en) | Multi-direction dialogue | |
Mankoff et al. | OOPS: a toolkit supporting mediation techniques for resolving ambiguity in recognition-based interfaces | |
Nagao et al. | Ubiquitous talker: Spoken language interaction with real world objects | |
WO2006107586A2 (en) | Method and system for interpreting verbal inputs in a multimodal dialog system | |
Vu et al. | Gptvoicetasker: Llm-powered virtual assistant for smartphone | |
DK179930B1 (en) | Detecting a trigger of a digital assistant | |
CN112071318A (en) | User interface for correcting recognition errors | |
Seipel et al. | Speak to your software visualization—exploring component-based software architectures in augmented reality with a conversational interface | |
US20060085414A1 (en) | System and methods for reference resolution | |
KR20190134794A (en) | Hierarchical Belief States for Digital Assistants | |
EP4377773A2 (en) | Gaze based dictation | |
West et al. | A context inference and multi-modal approach to mobile information access | |
CN115083415A (en) | Continuous dialogue with digital assistant | |
Potamianos et al. | Human-computer interfaces to multimedia content a review | |
Ismail et al. | Designing and usability evaluation of multimodal input modalities for visually impaired persons on mobile platforms | |
CN115083414A (en) | Multi-state digital assistant for continuous conversation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MOTOROLA, INC., ILLINOIS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:THOMPSON, WILLIAM K.;GUPTA, ANURAG K.;LEE, HANG S.;REEL/FRAME:017495/0851;SIGNING DATES FROM 20051014 TO 20060101 |
|
AS | Assignment |
Owner name: MOTOROLA, INC., ILLINOIS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:THOMPSON, WILLIAM K.;LEE, HANG S.;GUPTA, ANURAG K.;REEL/FRAME:017333/0374;SIGNING DATES FROM 20051014 TO 20060110 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: MOTOROLA MOBILITY, INC, ILLINOIS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MOTOROLA, INC;REEL/FRAME:025673/0558 Effective date: 20100731 |
|
AS | Assignment |
Owner name: MOTOROLA MOBILITY LLC, ILLINOIS Free format text: CHANGE OF NAME;ASSIGNOR:MOTOROLA MOBILITY, INC.;REEL/FRAME:029216/0282 Effective date: 20120622 |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: GOOGLE TECHNOLOGY HOLDINGS LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MOTOROLA MOBILITY LLC;REEL/FRAME:034419/0001 Effective date: 20141028 |
|
FPAY | Fee payment |
Year of fee payment: 8 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 12 |