Synchronizing visual and speech events in a multimodal application
First Claim
1. A method for synchronizing visual and speech events in a multimodal application, the method comprising:
- calling a voice form of the multimodal application, wherein the multimodal application is run using at least one computer processor, wherein the multimodal application provides a multimodal web page to a client device over a network;
receiving speech from a user;
determining a semantic interpretation of at least a portion of the speech using the voice form;
calling a global application update handler of the multimodal application and exiting the voice form;
identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation, wherein the additional processing function is independent of the voice form; and
executing the additional processing function to synchronize visual and speech events in the multimodal application,wherein determining a semantic interpretation of at least a portion of the speech comprises determining a plurality of semantic interpretations of the at least a portion of the speech, andwherein identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation comprises identifying, by the global application update handler, an additional processing function for each of the plurality of semantic interpretations.
1 Assignment
0 Petitions
Accused Products
Abstract
Exemplary methods, systems, and products are disclosed for synchronizing visual and speech events in a multimodal application, including receiving from a user speech; determining a semantic interpretation of the speech; calling a global application update handler; identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation; and executing the additional function. Typical embodiments may include updating a visual element after executing the additional function. Typical embodiments may include updating a voice form after executing the additional function. Typical embodiments also may include updating a state table after updating the voice form. Typical embodiments also may include restarting the voice form after executing the additional function.
-
Citations
17 Claims
-
1. A method for synchronizing visual and speech events in a multimodal application, the method comprising:
-
calling a voice form of the multimodal application, wherein the multimodal application is run using at least one computer processor, wherein the multimodal application provides a multimodal web page to a client device over a network; receiving speech from a user; determining a semantic interpretation of at least a portion of the speech using the voice form; calling a global application update handler of the multimodal application and exiting the voice form; identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation, wherein the additional processing function is independent of the voice form; and executing the additional processing function to synchronize visual and speech events in the multimodal application, wherein determining a semantic interpretation of at least a portion of the speech comprises determining a plurality of semantic interpretations of the at least a portion of the speech, and wherein identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation comprises identifying, by the global application update handler, an additional processing function for each of the plurality of semantic interpretations. - View Dependent Claims (2, 3, 4, 5, 6)
-
-
7. A system for synchronizing visual and speech events in a multimodal application, the system comprising:
-
at least one computer processor; at least one computer memory operatively coupled to the computer processor; and computer program instructions disposed within the computer memory that, when executed, cause the at least one computer processor to; call a voice form of the multimodal application, wherein the multimodal application provides a multimodal web page to a client device over a network; receive speech from a user; determine a plurality of semantic interpretations of at least a portion of the speech using the voice form; call a global application update handler of the multimodal application and exit the voice form; identify, by the global application update handler, an additional processing function in dependence upon the semantic interpretation for each of the plurality of semantic interpretations, wherein the additional processing function is independent of the voice form; and execute the additional processing function to synchronize visual and speech events in the multimodal application. - View Dependent Claims (8, 9, 10, 11, 12)
-
-
13. A non-transitory computer-readable storage medium comprising instructions that, when executed on at least one processor in a computer, perform a method of synchronizing visual and speech events in a multimodal application, the method comprising:
-
calling a voice form of the multimodal application, wherein the multimodal application provides a multimodal web page to a client device over a network; receiving speech from a user; determining a semantic interpretation of at least a portion of the speech using the voice form; calling a global application update handler of the multimodal application and exiting the voice form; identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation, wherein the additional processing function is independent of the voice form; and executing the additional processing function to synchronize visual and speech events in the multimodal application, wherein determining a semantic interpretation of at least a portion of the speech comprises determining a plurality of semantic interpretations of the at least a portion of the speech; and wherein identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation comprises identifying, by the global application update handler, an additional processing function for each of the plurality of semantic interpretations. - View Dependent Claims (14, 15, 16, 17)
-
Specification