Configuring a speech engine for a multimodal application based on location
First Claim
1. A method of configuring a speech engine for a multimodal application based on location, the multimodal application operating on a multimodal device supporting multiple modes of user interaction with the multimodal application, the modes of user interaction including a voice mode and one or more non-voice modes, the multimodal application operatively coupled to a speech engine, the method comprising:
- receiving a location change notification in a location change monitor from a device location manager, the device location manager operatively coupled to a position detection component of the multimodal device, the location change notification specifying a current location of the multimodal device;
identifying in a configuration parameter repository, location-based configuration parameters for the speech engine in dependence upon the current location of the multimodal device, the location-based configuration parameters specifying a configuration for the speech engine at the current location, wherein the location-based configuration parameters include an identifier for an acoustic model from among a plurality of acoustic models including a first acoustic model and a second acoustic model, an identifier for a lexicon from among a plurality of lexicons including a first lexicon and a second lexicon, speech transition times, silence detection times, speech timeouts, gain maps, and a configuration for use by a text-to-speech (‘
TTS’
) engine including a voice used in synthesizing speech from text, wherein each of the first acoustic model and the second acoustic model associates acoustic features with phonemes, wherein the first lexicon and the second lexicon specify a different phoneme representation for a same word; and
updating, by the location change monitor, a current configuration for the speech engine according to the identified location-based configuration parameters.
3 Assignments
0 Petitions
Accused Products
Abstract
Methods, apparatus, and products are disclosed for configuring a speech engine for a multimodal application based on location. The multimodal application operates on a multimodal device supporting multiple modes of user interaction with the multimodal application. The multimodal application is operatively coupled to a speech engine. Configuring a speech engine for a multimodal application based on location includes: receiving a location change notification in a location change monitor from a device location manager, the location change notification specifying a current location of the multimodal device; identifying, by the location change monitor, location-based configuration parameters for the speech engine in dependence upon the current location of the multimodal device, the location-based configuration parameters specifying a configuration for the speech engine at the current location; and updating, by the location change monitor, a current configuration for the speech engine according to the identified location-based configuration parameters.
153 Citations
11 Claims
-
1. A method of configuring a speech engine for a multimodal application based on location, the multimodal application operating on a multimodal device supporting multiple modes of user interaction with the multimodal application, the modes of user interaction including a voice mode and one or more non-voice modes, the multimodal application operatively coupled to a speech engine, the method comprising:
-
receiving a location change notification in a location change monitor from a device location manager, the device location manager operatively coupled to a position detection component of the multimodal device, the location change notification specifying a current location of the multimodal device; identifying in a configuration parameter repository, location-based configuration parameters for the speech engine in dependence upon the current location of the multimodal device, the location-based configuration parameters specifying a configuration for the speech engine at the current location, wherein the location-based configuration parameters include an identifier for an acoustic model from among a plurality of acoustic models including a first acoustic model and a second acoustic model, an identifier for a lexicon from among a plurality of lexicons including a first lexicon and a second lexicon, speech transition times, silence detection times, speech timeouts, gain maps, and a configuration for use by a text-to-speech (‘
TTS’
) engine including a voice used in synthesizing speech from text, wherein each of the first acoustic model and the second acoustic model associates acoustic features with phonemes, wherein the first lexicon and the second lexicon specify a different phoneme representation for a same word; andupdating, by the location change monitor, a current configuration for the speech engine according to the identified location-based configuration parameters. - View Dependent Claims (2, 3, 4)
-
-
5. An apparatus for configuring a speech engine for a multimodal application based on location, the multimodal application operating on a multimodal device supporting multiple modes of user interaction with the multimodal application, the modes of user interaction including a voice mode and one or more non-voice modes, the multimodal application operatively coupled to a speech engine, the apparatus comprising a computer processor and a computer memory operatively coupled to the computer processor, the computer memory having disposed within it computer program instructions that, when executed by the computer processor, perform a method of:
-
receiving a location change notification in a location change monitor from a device location manager, the device location manager operatively coupled to a position detection component of the multimodal device, the location change notification specifying a current location of the multimodal device; identifying in a configuration parameter repository, location-based configuration parameters for the speech engine in dependence upon the current location of the multimodal device, the location-based configuration parameters specifying a configuration for the speech engine at the current location, wherein the location-based configuration include an identifier for an acoustic model from among a plurality of acoustic models including a first acoustic model and a second acoustic model, an identifier for a lexicon from among a plurality of lexicons including a first lexicon and a second lexicon, speech transition times, silence detection times, speech timeouts, gain maps, and a configuration for use by a text-to-speech (‘
TTS’
) engine including a voice used in synthesizing speech from text, wherein each of the first acoustic model and the second acoustic model associates acoustic features with phonemes, wherein the first lexicon and the second lexicon specify a different phoneme representation for a same word; andupdating, by the location change monitor, a current configuration for the speech engine according to the identified location-based configuration parameters. - View Dependent Claims (6, 7, 8)
-
-
9. A non-transitory computer readable recordable medium encoded with a plurality of instructions that, when executed on a computer, perform a method of configuring a speech engine for a multimodal application based on location, the multimodal application operating on a multimodal device supporting multiple modes of user interaction with the multimodal application, the modes of user interaction including a voice mode and one or more non-voice modes, the multimodal application operatively coupled to a speech engine, the method comprising:
-
receiving a location change notification in a location change monitor from a device location manager, the device location manager operatively coupled to a position detection component of the multimodal device, the location change notification specifying a current location of the multimodal device, wherein the location change notification includes a timestamp specifying a time at which the multimodal device arrived at the current location; identifying in a configuration parameter repository, location-based configuration parameters for the speech engine in dependence upon the current location of the multimodal device, the location-based configuration parameters specifying a configuration for the speech engine at the current location, wherein the location-based configuration parameters include an identifier for an acoustic model from among a plurality of acoustic models including a first acoustic model and a second acoustic model, an identifier for a lexicon from among a plurality of lexicons including a first lexicon and a second lexicon, speech transition times, silence detection times, speech timeouts, gain maps, and a configuration for use by a text-to-speech (‘
TTS’
) engine including a voice used in synthesizing speech from text, wherein each of the first acoustic model and the second acoustic model associates acoustic features with phonemes, wherein the first lexicon and the second lexicon specify a different phoneme representation for a same word; andupdating, by the location change monitor, a current configuration for the speech engine according to the identified location-based configuration parameters. - View Dependent Claims (10, 11)
-
Specification