MOBILE SYSTEMS AND METHODS OF SUPPORTING NATURAL LANGUAGE HUMAN-MACHINE INTERACTIONS
5 Assignments
0 Petitions
Accused Products
Abstract
A mobile system is provided that includes speech-based and non-speech-based interfaces for telematics applications. The mobile system identifies and uses context, prior information, domain knowledge, and user specific profile data to achieve a natural environment for users that submit requests and/or commands in multiple domains. The invention creates, stores and uses extensive personal profile information for each user, thereby improving the reliability of determining the context and presenting the expected results for a particular question or command. The invention may organize domain specific behavior and information into agents, that are distributable or updateable over a wide area network.
-
Citations
48 Claims
-
1-32. -32. (canceled)
-
33. A device for processing natural language inputs, comprising one or more processors configured to:
-
receive a multi-modal natural language input from a user, the multi-modal natural language input including a natural language utterance and a non-speech input; generate a non-speech transcription from the non-speech input; identify the user who provided the multi-modal natural language input; generate a speech-based transcription based on a cognitive model associated with the user, wherein the cognitive model includes information on one or more prior interactions between the user and the device; generate a merged transcription from the speech-based transcription and the non-speech transcription; identify an entry in a context stack that matches information in the merged transcription; identify a domain agent associated with the entry in the context stack; determine a request based on the merged transcription; and communicate the request to the domain agent, wherein the domain agent is configured to generate a response to the user. - View Dependent Claims (34, 35, 36, 37, 38, 39, 40, 41, 42, 43)
-
-
44. A method for processing natural language inputs, comprising:
-
receiving, by one or more processors, a multi-modal natural language input from a user, the multi-modal natural language input including a natural language utterance and a non-speech input; generating, by the one or more processors, a non-speech transcription from the non-speech input; identifying, by the one or more processors, the user who provided the multi-modal natural language input; generating, by the one or more processors, a speech-based transcription based on a cognitive model associated with the user, wherein the cognitive model includes information on one or more prior interactions between the user and the device; generating, by the one or more processors, a merged transcription from the speech-based transcription and the non-speech transcription; identifying, by the one or more processors, an entry in a context stack that matches information in the merged transcription; identifying, by the one or more processors, a domain agent associated with the entry in the context stack; determining, by the one or more processors, a request based on the merged transcription; and communicating, by the one or more processors, the request to the domain agent, wherein the domain agent is configured to generate a response to the user. - View Dependent Claims (45)
-
-
46. A device for processing natural language inputs, comprising one or more processors configured to:
-
receive a natural language utterance from a user; identify the user who provided the natural language utterance; generate a speech-based transcription based on a personal cognitive model associated with the user and a general cognitive model associated with the user, wherein the personal cognitive model includes information on one or more prior interactions between the device and the user, and wherein the general cognitive model includes information on one or more prior interactions between the device and a plurality of users; identify an entry indicative of a context of the natural language utterance; identify a domain agent associated with the entry in the stack; determine a request based on the speech-based transcription; and communicate the request to the domain agent, wherein the domain agent is configured to generate a response to the user. - View Dependent Claims (47, 48)
-
Specification