Interactive manual, system and method for vehicles and other complex equipment
First Claim
1. A method of interacting with a user to provide information regarding an operation of a device, comprising:
- receiving speech input from the user;
converting the speech into a word sequence;
identifying a structure to which the word sequence conforms;
providing a structured manual including information related to an operation of a device;
providing a visual model to relate a visual representation of the information;
interpreting the structure in a context and extracting information and associated visual representation from the structured manual and the visual model, wherein the structure is categorized into one of a “
how-to”
meaning structure category and a “
what-is”
meaning structure category, each of the meaning structure categories characterizing the structures of its members as being of a single respective underlying meaning structure, different information being deemed pertinent for the extraction depending upon the meaning category into which the structure of the word sequence has been categorized; and
outputting the information to the user,wherein the structure manual and the visual model form a model package that includes a grammar package and a grammar table defined using a grammar specification language,wherein the model package further includes an objects file to define device model information that includes a model of at least one device component,wherein the object file is associated with at least one of a surrounding table, a slide table, a how-to table, and a what-is table,wherein the grammar package includes a set of phrases and sentences as grammars for the speech recognition engine to recognize, and the grammar table includes animation clips associated with the specified grammars, andwherein a categorization into the “
what-is”
meaning structure category causes the dialog arrangement to;
determine an object in the “
what-is”
meaning structure;
determine an object identification of the object;
search the surrounding table for information of adjacent objects based on the object identification;
search the what-is table based on the object identification for a textual description associated with a feature of the object; and
display the textual description when the associated feature is displayed.
1 Assignment
0 Petitions
Accused Products
Abstract
A method and system of providing an interactive manual, including a speech engine to receive and process speech from a user, convert the speech into a word sequence, and identify meaning structures from the word sequence, a structured manual including information related to an operation of a device, a visual model to relate visual representation of the information, a dialog management arrangement to interpret the meaning structures in a context and to extract pertinent information and the visual representation from the structured manual and the visual model, and an output arrangement to output the information and visual representation.
23 Citations
38 Claims
-
1. A method of interacting with a user to provide information regarding an operation of a device, comprising:
-
receiving speech input from the user; converting the speech into a word sequence; identifying a structure to which the word sequence conforms; providing a structured manual including information related to an operation of a device; providing a visual model to relate a visual representation of the information; interpreting the structure in a context and extracting information and associated visual representation from the structured manual and the visual model, wherein the structure is categorized into one of a “
how-to”
meaning structure category and a “
what-is”
meaning structure category, each of the meaning structure categories characterizing the structures of its members as being of a single respective underlying meaning structure, different information being deemed pertinent for the extraction depending upon the meaning category into which the structure of the word sequence has been categorized; andoutputting the information to the user, wherein the structure manual and the visual model form a model package that includes a grammar package and a grammar table defined using a grammar specification language, wherein the model package further includes an objects file to define device model information that includes a model of at least one device component, wherein the object file is associated with at least one of a surrounding table, a slide table, a how-to table, and a what-is table, wherein the grammar package includes a set of phrases and sentences as grammars for the speech recognition engine to recognize, and the grammar table includes animation clips associated with the specified grammars, and wherein a categorization into the “
what-is”
meaning structure category causes the dialog arrangement to;determine an object in the “
what-is”
meaning structure;determine an object identification of the object; search the surrounding table for information of adjacent objects based on the object identification; search the what-is table based on the object identification for a textual description associated with a feature of the object; and display the textual description when the associated feature is displayed. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9)
-
-
10. A storage medium having a set of instructions executable by a processor to perform a method of interacting with a user to provide information regarding an operation of a device, the method comprising:
-
receiving speech input from the user; converting the speech into a word sequence; identifying a structure to which the word sequence conforms; providing a structured manual including information related to an operation of a device; providing a visual model to relate a visual representation of the information; interpreting the structure in a context and extracting information and associated visual representation from the structured manual and the visual model, wherein the structure is categorized into one of a “
how-to”
meaning structure category and a “
what-is”
meaning structure category, each of the meaning structure categories characterizing the structures of its members as being of a single respective underlying meaning structure, different information being deemed pertinent for the extraction depending upon the meaning category into which the structure of the word sequence has been categorized; andoutputting the information to the user, wherein the structure manual and the visual model form a model package that includes a grammar package and a grammar table defined using a grammar specification language, wherein the model package further includes an objects file to define device model information that includes a model of at least one device component, wherein the object file is associated with at least one of a surrounding table, a slide table, a how-to table, and a what-is table, wherein the grammar package includes a set of phrases and sentences as grammars for the speech recognition engine to recognize, and the grammar table includes animation clips associated with the specified grammars, and wherein a categorization into the “
what-is”
meaning structure category causes the dialog arrangement to;determine an object in the “
what-is”
meaning structure;determine an object identification of the object; search the surrounding table for information of adjacent objects based on the object identification; search the what-is table based on the object identification for a textual description associated with a feature of the object; and display the textual description when the associated feature is displayed.
-
-
11. An interactive manual system, comprising:
-
a speech engine to receive and process speech from a user, convert the speech into a word sequence, and identify from the word sequence at least one of a spatial meaning structure, a geometric meaning structure, and a color meaning structure from the word sequence; a structured manual including textual information related to an operation of a device; a visual model to relate portions of a visual representation of the device to portions of the textual information; a dialog management arrangement that uses a context to identify at least one component of the device based on the identified at least one meaning structure and that extracts at least one portion of the textual information and at least one portion of the visual representation that pertain to the identified at least one component, wherein the structure is categorized into one of a “
how-to”
meaning structure category and a “
what-is”
meaning structure category; andan output arrangement to output the extracted portions of the textual information and the visual representation, wherein the structure manual and the visual model form a model package that includes a grammar package and a grammar table defined using a grammar specification language, wherein the model package further includes an objects file to define device model information that includes a model of at least one device component, wherein the object file is associated with at least one of a surrounding table, a slide table, a how-to table, and a what-is table, wherein the grammar package includes a set of phrases and sentences as grammars for the speech recognition engine to recognize, and the grammar table includes animation clips associated with the specified grammars, and wherein a categorization into the “
what-is”
meaning structure category causes the dialog arrangement to;determine an object in the “
what-is”
meaning structure;determine an object identification of the object; search the surrounding table for information of adjacent objects based on the object identification; search the what-is table based on the object identification for a textual description associated with a feature of the object; and display the textual description when the associated feature is displayed.
-
-
12. An interactive manual system, comprising:
-
a speech engine to receive and process speech from a user, convert the speech into a word sequence, and identify a structure to which the word sequence conforms; a structured manual including information related to an operation of a device; a visual model to relate a visual representation of the information; a dialog management arrangement to interpret the structure in a context and to extract pertinent information and the visual representation from the structured manual and the visual model, wherein the structure is categorized into one of a “
how-to”
meaning structure category and a “
what-is”
meaning structure category, each of the meaning structure categories characterizing the structures of its members as being of a single respective underlying meaning structure, different information being deemed pertinent for the extraction depending upon the meaning structure category into which the structure of the word sequence has been categorized; andan output arrangement to output the information and visual representation, wherein the structure manual and the visual model form a model package that includes a grammar package and a grammar table defined using a grammar specification language, wherein the model package further includes an objects file to define device model information that includes a model of at least one device component, wherein the object file is associated with at least one of a surrounding table, a slide table, a how-to table, and a what-is table, wherein the grammar package includes a set of phrases and sentences as grammars for the speech recognition engine to recognize, and the grammar table includes animation clips associated with the specified grammars, and wherein a categorization into the “
what-is”
meaning structure category causes the dialog arrangement to;determine an object in the “
what-is”
meaning structure;determine an object identification of the object; search the surrounding table for information of adjacent objects based on the object identification; search the what-is table based on the object identification for a textual description associated with a feature of the object; and display the textual description when the associated feature is displayed. - View Dependent Claims (13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28)
-
-
29. An interactive manual system, comprising:
-
a speech engine to receive and process speech from a user, convert the speech into a word sequence, and identify a structure to which the word sequence conforms; a structured manual including information related to an operation of a device; a visual model to relate a visual representation of the information; a dialog management arrangement to interpret the structure in a context and to extract pertinent information and the visual representation from the structured manual and the visual model, wherein the structure is categorized into one of a “
how-to”
meaning structure category and a “
what-is”
meaning structure category, each of the meaning structure categories characterizing the structures of its members as being of a single respective underlying meaning structure, different information being deemed pertinent for the extraction depending upon the meaning structure category into which the structure of the word sequence has been categorized; andan output arrangement to output the information and visual representation, wherein the structure manual and the visual model form a model package that includes a grammar package and a grammar table defined using a grammar specification language, wherein the model package further includes an objects file to define device model information that includes a model of at least one device component, wherein the object file is associated with at least one of a surrounding table, a slide table, a how-to table, and a what-is table, wherein the grammar package includes a set of phrases and sentences as grammars for the speech recognition engine to recognize, and the grammar table includes animation clips associated with the specified grammars, and wherein a categorization into the “
how-to”
meaning structure category causes the dialog arrangement to;determine an object in the “
how-to”
meaning structure;determine an object identification of the object; search the surrounding table for information of adjacent objects based on the object identification; search the slide table for animation clips associated with the object identification; link the animation clips associated with the object identification and the animation clips in the grammar table; and display the linked animation clips. - View Dependent Claims (30)
-
-
31. An interactive manual system, comprising:
-
a speech engine configured for converting an utterance from a user into a word sequence and identifying a meaning structure category to which the word sequence conforms based on one of a statistical method and rule-based method, a determination being made whether the utterance is a command or a question, the meaning structure category being identified as either a “
how-to”
meaning structure category or a “
what-is”
meaning structure category;a display manager arrangement configured for displaying views of a current object based on the command; a model manager arrangement configured for storing manual data regarding at least one object, and an object file, the manual data including text descriptions, static display frames, animation clips, and a lookup table for searching the manual data, the object file including at least one of spatial and graphical information of the object; and an interaction manager arrangement, programmed with a plurality of rule and action pairs, configured for controlling a display via the display manager arrangement and based on an identified meaning structure category; wherein the interaction manager arrangement identifies an object in one of the identified meaning structure categories, determines a rule based on the identified object and the identified meaning structure category, and generates a display via the display manager arrangement based on manual data and object file data associated with the identified object and the identified meaning structure category. - View Dependent Claims (32, 33, 34, 35)
-
-
36. A method of interacting with a user to provide information regarding an operation of a device, comprising:
-
receiving speech input from the user; converting the speech into a word sequence; identifying a structure to which the word sequence conforms; providing a structured manual including information related to an operation of a device; providing a visual model to relate a visual representation of the information; interpreting the structure in a context; extracting information and associated visual representation from the structured manual and the visual model, wherein the structure is categorized into one of a “
how-to”
meaning structure category and a “
what-is”
meaning structure category, each of the meaning structure categories characterizing the structures of its members as being of a single respective underlying meaning structure, different information being deemed pertinent for the extraction depending upon the meaning category into which the structure of the word sequence has been categorized; andoutputting the information to the user, wherein the structure manual and the visual model form a model package that includes a grammar package and a grammar table defined using a grammar specification language, wherein the model package further includes an objects file to define device model information that includes a model of at least one device component, wherein the object file is associated with at least one of a surrounding table, a slide table, a how-to table, and a what-is table, wherein the grammar package includes a set of phrases and sentences as grammars for the speech recognition engine to recognize, and the grammar table includes animation clips associated with the specified grammars, and wherein a categorization into the “
how-to”
meaning structure category causes the dialog arrangement to;determine an object in the “
how-to”
meaning structure;determine an object identification of the object; search the surrounding table for information of adjacent objects based on the object identification; search the slide table for animation clips associated with the object identification; link the animation clips associated with the object identification and the animation clips in the grammar table; and display the linked animation clips.
-
-
37. A storage medium having a set of instructions executable by a processor to perform a method of interacting with a user to provide information regarding an operation of a device, the method comprising:
-
receiving speech input from the user; converting the speech into a word sequence; identifying a structure to which the word sequence conforms; providing a structured manual including information related to an operation of a device; providing a visual model to relate a visual representation of the information; interpreting the structure in a context; extracting information and associated visual representation from the structured manual and the visual model, wherein the structure is categorized into one of a “
how-to”
meaning structure category and a “
what-is”
meaning structure category, each of the meaning structure categories characterizing the structures of its members as being of a single respective underlying meaning structure, different information being deemed pertinent for the extraction depending upon the meaning category into which the structure of the word sequence has been categorized; andoutputting the information to the user, wherein the structure manual and the visual model form a model package that includes a grammar package and a grammar table defined using a grammar specification language, wherein the model package further includes an objects file to define device model information that includes a model of at least one device component, wherein the object file is associated with at least one of a surrounding table, a slide table, a how-to table, and a what-is table, wherein the grammar package includes a set of phrases and sentences as grammars for the speech recognition engine to recognize, and the grammar table includes animation clips associated with the specified grammars, and wherein a categorization into the “
how-to”
meaning structure category causes the dialog arrangement to;determine an object in the “
how-to”
meaning structure;determine an object identification of the object; search the surrounding table for information of adjacent objects based on the object identification; search the slide table for animation clips associated with the object identification; link the animation clips associated with the object identification and the animation clips in the grammar table; and display the linked animation clips.
-
-
38. An interactive manual system, comprising:
-
a speech engine to receive and process speech from a user, convert the speech into a word sequence, and identify from the word sequence at least one of a spatial meaning structure, a geometric meaning structure, and a color meaning structure from the word sequence; a structured manual including textual information related to an operation of a device; a visual model to relate portions of a visual representation of the device to portions of the textual information; a dialog management arrangement that uses a context to identify at least one component of the device based on the identified at least one meaning structure and that extracts at least one portion of the textual information and at least one portion of the visual representation that pertain to the identified at least one component, wherein the structure is categorized into one of a “
how-to”
meaning structure category and a “
what-is”
meaning structure category; andan output arrangement to output the extracted portions of the textual information and the visual representation, wherein the structure manual and the visual model form a model package that includes a grammar package and a grammar table defined using a grammar specification language, wherein the model package further includes an objects file to define device model information that includes a model of at least one device component, wherein the object file is associated with at least one of a surrounding table, a slide table, a how-to table, and a what-is table, wherein the grammar package includes a set of phrases and sentences as grammars for the speech recognition engine to recognize, and the grammar table includes animation clips associated with the specified grammars, and wherein a categorization into the “
how-to”
meaning structure category causes the dialog arrangement to;determine an object in the “
how-to”
meaning structure;determine an object identification of the object; search the surrounding table for information of adjacent objects based on the object identification; search the slide table for animation clips associated with the object identification; link the animation clips associated with the object identification and the animation clips in the grammar table; and display the linked animation clips.
-
Specification