Assisted multi-modal dialogue
First Claim
Patent Images
1. A client-server system for providing assisted multi-modal dialogue, comprising:
- a web server for generating client-side markups having recognition and audible prompting for execution on a client having recognition capabilities, the web server further including controls for generating the client side markups, the controls including speech controls inheritance for setting values to properties associated with the controls and organized in collections to construct a dialog for obtaining information pertaining to a plurality of topics, each collection of controls configured to create a separate dialog associated with a separate topic;
a recognition server, coupled to the web server, for providing speech recognition processing to received voice data based on a grammar or language model provided with the received voice data to produce speech recognition results, the speech recognition results being provided to the web server; and
a telephone voice browser, coupled to the web server, for processing voice data, the telephone voice browser including a media server for providing a telephony interface and a voice browser;
wherein the controls of the web server include companion controls associated with corresponding primary controls for providing recognition and audible prompting, the companion controls including a semantic map, wherein the semantic map includes semantic items and forms an association between a visual domain of the primary controls and a non-visual recognition domain of the companion controls and wherein the dialog includes at least one question provided by a prompt object and at least one answer, a grammar object is provided to define a grammar for recognition of input data and related processing on the input and an answer property associates a recognized result with a semantic item in the semantic map.
2 Assignments
0 Petitions
Accused Products
Abstract
Controls are provided for a web server to generate client side markups that include recognition and/or audible prompting. The controls are organized in collections to obtain information pertaining to different topics. Each collection of controls create a separate dialog. In this manner, the collections can be selectively specified to execute the corresponding dialog.
-
Citations
18 Claims
-
1. A client-server system for providing assisted multi-modal dialogue, comprising:
-
a web server for generating client-side markups having recognition and audible prompting for execution on a client having recognition capabilities, the web server further including controls for generating the client side markups, the controls including speech controls inheritance for setting values to properties associated with the controls and organized in collections to construct a dialog for obtaining information pertaining to a plurality of topics, each collection of controls configured to create a separate dialog associated with a separate topic; a recognition server, coupled to the web server, for providing speech recognition processing to received voice data based on a grammar or language model provided with the received voice data to produce speech recognition results, the speech recognition results being provided to the web server; and a telephone voice browser, coupled to the web server, for processing voice data, the telephone voice browser including a media server for providing a telephony interface and a voice browser; wherein the controls of the web server include companion controls associated with corresponding primary controls for providing recognition and audible prompting, the companion controls including a semantic map, wherein the semantic map includes semantic items and forms an association between a visual domain of the primary controls and a non-visual recognition domain of the companion controls and wherein the dialog includes at least one question provided by a prompt object and at least one answer, a grammar object is provided to define a grammar for recognition of input data and related processing on the input and an answer property associates a recognized result with a semantic item in the semantic map. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9)
-
-
10. A computer implemented method for performing recognition and/or audible prompting on a client device in a client/server system, the method comprising:
-
generating client-side markups having recognition and audible prompting at a web server for execution on a client having recognition capabilities; providing controls, organized in collections, for generating the client side markups at the web server, wherein the providing controls includes setting values to properties associated with the controls according to speech controls inheritance, wherein the providing controls for the web server includes providing companion controls associated with corresponding primary controls and including a semantic map to provide recognition and audible prompting; obtaining information pertaining to a plurality of topics using a dialog constructed from the controls, each collection of controls configured to create a separate dialog associated with a separate topic; providing speech recognition processing to received voice data at a recognition server based on a grammar or language model provided with the received voice data to produce speech recognition results; providing the speech recognition results to the web server; processing voice data at a telephone voice browser; providing semantic items for the semantic map; forming an association between a visual domain of the primary controls and a non-visual recognition domain of the companion controls using the semantic items; providing at least one question for the dialog via a prompt object; providing at least one answer; providing a grammar object to define a grammar for recognition of input data and related processing on the input; and providing an answer property associating a recognized result with a semantic item in the semantic map. - View Dependent Claims (11, 12, 13, 14, 15, 16, 17)
-
-
18. A client-server system for providing assisted multi-modal dialogue, comprising:
-
a web server for generating client-side markups having recognition and audible prompting for execution on a client having recognition capabilities, the web server also includes controls for generating the client side markups, the controls including speech controls inheritance for setting values to properties associated with the controls and organized in collections to construct a dialog for obtaining information pertaining to a plurality of topics, each collection of controls configured to create a separate dialog associated with a separate topic, the web server further includes an authoring tool for dynamically generating the client-side markups and a specific form of markup for the type of client accessing the web server and a library for providing visual, recognition and audible prompting markup information; a recognition server, coupled to the web server, for providing speech recognition processing to received voice data based on a grammar or language model provided with the received voice data to produce speech recognition results, the speech recognition results being provided to the web server; a telephone voice browser, coupled to the web server, for processing voice data, the telephone voice browser including a media server for providing a telephony interface and a voice browser; and at least one client device for receiving the dialog and providing voice input in response to the dialog; wherein the controls of the web server include companion controls associated with corresponding primary controls for providing recognition and audible prompting, the companion controls including a semantic map, wherein the semantic map includes semantic items and forms an association between a visual domain of the primary controls and a non-visual recognition domain of the companion controls; wherein the dialog includes at least one question provided by a prompt object and at least one answer, a grammar object is provided to define a grammar for recognition of input data and related processing on the input and an answer property associates a recognized result with a semantic item in the semantic map; and wherein the recognized input is associated with one of the primary controls and when input is received through a graphical user interface, the received input is replicated in a corresponding semantic item and status information for the input is set as being confirmed via a companion control, and wherein a reset is provided by the companion controls to expose semantic items and status information for resetting selected portions of the dialog to remove input associated with semantic items corresponding to the selected portions of the dialog.
-
Specification