Method and apparatus for providing global voice-based entry of geographic information in a device
First Claim
1. A method for automatic speech recognition in a device, comprising:
- partitioning a global speech decoding graph into one or more spatial partitions according to a geographic topology of one or more geographic entities, one or more geographic terms, or a combination thereof, wherein each of the spatial partitions contains a decoding graph comprising a sub-set of the one or more geographic entities, the one or more geographic terms, or a combination thereof associated with a geographic area;
determining one or more of key entities occurring in each of the one or more spatial partitions, wherein at least one of the one or more of key entities includes at least one of the one or more geographic entities and the one or more geographic terms;
constructing a combined set of key entities comprising the one or more key entities from said each spatial partition;
creating a retrieval index to map the one or more key entities in the combined set of key entities to a corresponding partition from among the one or more spatial partitions,wherein a voice input signal associated with a request for one or more navigation or mapping related services is processed, using automatic speech recognition and a first partition associated with a first geographic area from among the one or more spatial partitions, the combined set of key entities, and the retrieved index are stored in a memory of the device, andwherein a second partition that is associated with a second geographic area and not in the memory of the device is retrieved based on the combined set of key entities and the retrieval index to automatically re-process the voice input signal when an out-of-vocabulary result is obtained with respect to the first partition and to provide the one or more navigation or mapping related services via a user interface based, at least in part, on the re-processing.
1 Assignment
0 Petitions
Accused Products
Abstract
An approach is provided for global voice-based entry of location information. The approach involves partitioning a global speech decoding graph into spatial partitions. The approach also involves determining key entities occurring in each spatial partitions to construct a combined set of key entities. The approach further involves creating a retrieval index to map the key entities in the combined set of key entities to a corresponding partition. A first partition, the combined set of key entities, and the retrieved index are stored in a memory of a device for processing a voice input signal. A second partition that is not in the memory of the device is retrieved based on the combined set of key entities and the retrieval index to automatically re-process the voice input signal when an out-of-vocabulary result is obtained from the first partition.
21 Citations
22 Claims
-
1. A method for automatic speech recognition in a device, comprising:
-
partitioning a global speech decoding graph into one or more spatial partitions according to a geographic topology of one or more geographic entities, one or more geographic terms, or a combination thereof, wherein each of the spatial partitions contains a decoding graph comprising a sub-set of the one or more geographic entities, the one or more geographic terms, or a combination thereof associated with a geographic area; determining one or more of key entities occurring in each of the one or more spatial partitions, wherein at least one of the one or more of key entities includes at least one of the one or more geographic entities and the one or more geographic terms; constructing a combined set of key entities comprising the one or more key entities from said each spatial partition; creating a retrieval index to map the one or more key entities in the combined set of key entities to a corresponding partition from among the one or more spatial partitions, wherein a voice input signal associated with a request for one or more navigation or mapping related services is processed, using automatic speech recognition and a first partition associated with a first geographic area from among the one or more spatial partitions, the combined set of key entities, and the retrieved index are stored in a memory of the device, and wherein a second partition that is associated with a second geographic area and not in the memory of the device is retrieved based on the combined set of key entities and the retrieval index to automatically re-process the voice input signal when an out-of-vocabulary result is obtained with respect to the first partition and to provide the one or more navigation or mapping related services via a user interface based, at least in part, on the re-processing. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15)
-
-
16. An apparatus for automatic speech recognition in a device, comprising:
-
at least one processor; and at least one memory including computer program code for one or more programs, the at least one memory and the computer program code configured to, with the at least one processor, cause the apparatus to perform at least the following, partition a global speech decoding graph into one or more spatial partitions according to a geographic topology of one or more geographic entities, one or more geographic terms, or a combination thereof, wherein each of the spatial partitions contains a decoding graph comprising a sub-set of the one or more geographic entities, the one or more geographic terms, or a combination thereof associated with a geographic area; determine one or more of key entities occurring in each of the one or more spatial partitions, wherein at least one of the one or more of key entities includes at least one of the one or more geographic entities and the one or more geographic terms; construct a combined set of key entities comprising the one or more key entities from said each spatial partition; create a retrieval index to map the one or more key entities in the combined set of key entities to a corresponding partition from among the one or more spatial partitions, wherein a voice input signal associated with a request for one or more navigation or mapping related services is processed, using automatic speech recognition and a first partition associated with a first geographic area from among the one or more spatial partitions, the combined set of key entities, and the retrieved index are stored in a memory of the device, and wherein a second partition that is associated with a second geographic area and not in the memory of the device is retrieved based on the combined set of key entities and the retrieval index to automatically re-process the voice input signal when an out-of-vocabulary result is obtained with respect to the first partition and to provide the one or more navigation or mapping related services via a user interface based, at least in part, on the re-processing. - View Dependent Claims (17, 18, 19)
-
-
20. A non-transitory computer-readable storage medium for automatic speech recognition in a device, carrying one or more sequences of one or more instructions which, when executed by one or more processors, cause an apparatus to at least perform the following steps:
-
partitioning a global speech decoding graph into one or more spatial partitions according to a geographic topology of one or more geographic entities, one or more geographic terms, or a combination thereof, wherein each of the spatial partitions contains a decoding graph comprising a sub-set of the one or more geographic entities, the one or more geographic terms, or a combination thereof associated with a geographic area; determining one or more of key entities occurring in each of the one or more spatial partitions, wherein at least one of the one or more of key entities includes at least one of the one or more geographic entities and the one or more geographic terms; constructing a combined set of key entities comprising the one or more key entities from said each spatial partition; creating a retrieval index to map the one or more key entities in the combined set of key entities to a corresponding partition from among the one or more spatial partitions, wherein a voice input signal associated with a request for one or more navigation or mapping related services is processed, using automatic speech recognition and a first partition associated with a first geographic area from among the one or more spatial partitions, the combined set of key entities, and the retrieved index are stored in a memory of the device, and wherein a second partition that is associated with a second geographic area and not in the memory of the device is retrieved based on the combined set of key entities and the retrieval index to automatically re-process the voice input signal when an out-of-vocabulary result is obtained with respect to the first partition and to provide the one or more navigation or mapping related services via a user interface based, at least in part, on the re-processing. - View Dependent Claims (21, 22)
-
Specification