Synchronizing visual and speech events in a multimodal application
First Claim
1. A method, comprising:
- receiving, by a multimodal application executing on a computer processor, multimodal input from a multimodal browser of a device, wherein the multimodal input comprises speech from a user;
determining a semantic interpretation of at least a portion of the speech using a voice form;
calling a global application update handler of the multimodal application;
identifying, by the global application update handler, an additional processing function based at least in part upon the semantic interpretation and a geographical location, wherein the additional processing function is independent of the voice form; and
executing the additional processing function, wherein the additional processing function executed depends on the semantic interpretation of the at least a portion of the speech,wherein determining a semantic interpretation of at least a portion of the speech comprises determining a plurality of semantic interpretations of the at least a portion of the speech, andwherein identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation comprises identifying, by the global application update handler, an additional processing function for each of the plurality of semantic interpretations.
2 Assignments
0 Petitions
Accused Products
Abstract
Exemplary methods, systems, and products are disclosed for synchronizing visual and speech events in a multimodal application, including receiving from a user speech; determining a semantic interpretation of the speech; calling a global application update handler; identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation; and executing the additional function. Typical embodiments may include updating a visual element after executing the additional function. Typical embodiments may include updating a voice form after executing the additional function. Typical embodiments also may include updating a state table after updating the voice form. Typical embodiments also may include restarting the voice form after executing the additional function.
-
Citations
20 Claims
-
1. A method, comprising:
-
receiving, by a multimodal application executing on a computer processor, multimodal input from a multimodal browser of a device, wherein the multimodal input comprises speech from a user; determining a semantic interpretation of at least a portion of the speech using a voice form; calling a global application update handler of the multimodal application; identifying, by the global application update handler, an additional processing function based at least in part upon the semantic interpretation and a geographical location, wherein the additional processing function is independent of the voice form; and executing the additional processing function, wherein the additional processing function executed depends on the semantic interpretation of the at least a portion of the speech, wherein determining a semantic interpretation of at least a portion of the speech comprises determining a plurality of semantic interpretations of the at least a portion of the speech, and wherein identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation comprises identifying, by the global application update handler, an additional processing function for each of the plurality of semantic interpretations. - View Dependent Claims (2, 3, 4, 5, 6)
-
-
7. A system, comprising:
-
at least one computer processor; at least one computer memory operatively coupled to the computer processor; and computer program instructions disposed within the computer memory that, when executed, cause the at least one computer processor to; receive, by a multimodal application executing on a computer processor, multimodal input from a multimodal browser of a device, wherein the multimodal input comprises speech from a user; determine a semantic interpretation of at least a portion of the speech using a voice form, the semantic interpretation comprising a plurality of semantic interpretations of the at least a portion of the speech; call a global application update handler of the multimodal application; identify, by the global application update handler, an additional processing function based at least in part upon the semantic interpretation and a geographical location, for each of the plurality of semantic interpretations, wherein the additional processing function is independent of the voice form; and execute the additional processing function, wherein the additional processing function executed depends on the semantic interpretation of the at least a portion of the speech. - View Dependent Claims (8, 9, 10, 11, 12)
-
-
13. A non-transitory computer-readable storage medium comprising instructions that, when executed on at least one computer processor, perform a method, comprising:
-
receiving, by a multimodal application executing on a computer processor, multimodal input from a multimodal browser of a device, wherein the multimodal input comprises speech from a user; determining a semantic interpretation of at least a portion of the speech using a voice form; calling a global application update handler of the multimodal application; identifying, by the global application update handler, an additional processing function based at least in part upon the semantic interpretation and a geographical location, wherein the additional processing function is independent of the voice form; and executing the additional processing function, wherein the additional processing function executed depends on the semantic interpretation of the at least a portion of the speech, wherein determining a semantic interpretation of at least a portion of the speech comprises determining a plurality of semantic interpretations of the at least a portion of the speech, and wherein identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation comprises identifying, by the global application update handler, an additional processing function for each of the plurality of semantic interpretations. - View Dependent Claims (14, 15, 16, 17, 18)
-
-
19. A non-transitory computer-readable storage medium comprising instructions that, when executed on at least one computer processor, perform a method, comprising:
-
receiving speech from a user; determining a semantic interpretation of at least a portion of the speech; and identifying, by a global application update handler, an additional processing function in dependence upon the semantic interpretation; executing the additional processing function to provide an advertisement based at least in part upon the semantic interpretation and a geographical location, wherein determining a semantic interpretation of at least a portion of the speech comprises determining a plurality of semantic interpretations of the at least a portion of the speech, and wherein identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation comprises identifying, by the global application update handler, an additional processing function for each of the plurality of semantic interpretations. - View Dependent Claims (20)
-
Specification