Method of performing multi-modal dialogue between a humanoid robot and user, computer program product and humanoid robot for implementing said method
First Claim
1. A method of performing a dialogue between a humanoid robot and at least one user comprising the following steps, carried out iteratively by said humanoid robot:
- i) acquiring a plurality of input signals from respective sensors, at least one said sensor being a sound sensor and at least one other sensor being a motion or image sensor;
ii) interpreting the acquired signals to recognize a plurality of events generated by said user, selected from a group comprising;
the utterance of at least a word or sentence, an intonation of voice, a gesture, a body posture, a facial expression;
iii) determining a response of said humanoid robot, comprising at least one event selected from a group comprising;
the utterance of at least a word or sentence, an intonation of voice, a gesture, a body posture, a facial expression, said determining being performed by applying a set of rules, each said rule associating a set of input events to a response of the robot;
iv) generating said or each said event;
wherein at least some of said rules applied at said step iii) associate a response to a combination of at least two events jointly generated by said user and recognized at said step ii), of which at least one is not a word or sentence uttered by said user, andif the response determined during step iii) is or comprises at least the utterance of a word or sentence, executing a step iii-a) of performing a syntactic analysis of a sentence to be uttered by the robot to determine at least one word to be animated depending on a function of the at least one word within a structure of said sentence and determining an animation accompanying said response as a function of said analysis.
0 Assignments
0 Petitions
Accused Products
Abstract
A method of performing dialog between a humanoid robot and user comprises: i) acquiring input signals from respective sensors, at least one being a sound sensor and another being a motion or image sensor; ii) interpreting the signals to recognize events generated by the user, including: the utterance of a word or sentence, an intonation of voice, a gesture, a body posture, a facial expression; iii) determining a response of the humanoid robot, comprising an event such as: the utterance of a word or sentence, an intonation of voice, a gesture, a body posture, a facial expression; iv) generating, an event by the humanoid robot; wherein step iii) comprises determining the response from events jointly generated by the user and recognized at step ii), of which at least one is not words uttered by the user. A computer program product and humanoid robot for carrying out the method is provided.
76 Citations
25 Claims
-
1. A method of performing a dialogue between a humanoid robot and at least one user comprising the following steps, carried out iteratively by said humanoid robot:
-
i) acquiring a plurality of input signals from respective sensors, at least one said sensor being a sound sensor and at least one other sensor being a motion or image sensor; ii) interpreting the acquired signals to recognize a plurality of events generated by said user, selected from a group comprising;
the utterance of at least a word or sentence, an intonation of voice, a gesture, a body posture, a facial expression;iii) determining a response of said humanoid robot, comprising at least one event selected from a group comprising;
the utterance of at least a word or sentence, an intonation of voice, a gesture, a body posture, a facial expression, said determining being performed by applying a set of rules, each said rule associating a set of input events to a response of the robot;iv) generating said or each said event; wherein at least some of said rules applied at said step iii) associate a response to a combination of at least two events jointly generated by said user and recognized at said step ii), of which at least one is not a word or sentence uttered by said user, and if the response determined during step iii) is or comprises at least the utterance of a word or sentence, executing a step iii-a) of performing a syntactic analysis of a sentence to be uttered by the robot to determine at least one word to be animated depending on a function of the at least one word within a structure of said sentence and determining an animation accompanying said response as a function of said analysis. - View Dependent Claims (2, 3, 4, 5, 6, 7, 10, 11, 23, 24, 25)
wherein said or at least one said processor is programmed or configured to carry out the method according to claim 1.
-
-
25. The humanoid robot comprising:
-
at least one embedded processor; a sensor assembly operatively connected to said or at least one said processor and comprising at least one sound sensor and at least one image or movement sensor, to acquire respective input signals; a speech synthesis module driven by said or at least one said processor to utter words or sentences, and a set of actuators driven by said or at least one said processor enabling said robot to perform a plurality of movements or gestures, further comprising a device for connection to at least one remote server, said or at least one said processor being programmed or configured to cooperate with said or at least one said remote server to carry out the method according to claim 1.
-
-
8. A method of performing a dialogue between a humanoid robot and at least one user comprising the following steps, carried out iteratively by said humanoid robot:
-
i) acquiring a plurality of input signals from respective sensors, at least one said sensor being a sound sensor and at least one other sensor being a motion or image sensor; ii) interpreting the acquired signals to recognize a plurality of events generated by said user, selected from a group comprising;
the utterance of at least a word or sentence, an intonation of voice, a gesture, a body posture, a facial expression;iii) determining a response of said humanoid robot, comprising at least one event selected from a group comprising;
the utterance of at least a word or sentence, an intonation of voice, a gesture, a body posture, a facial expression, said determining being performed by applying a set of rules, each said rule associating a set of input events to a response of the robot;iv) generating said or each said event; a step of phonetic transcription of a set of sounds acquired by a sound sensor; a step of simplifying and smoothing the resulting phonetic transcription; calculating an edit distance between said simplified and smoothed phonetic transcription and a plurality of entries, obtained by simplifying and smoothing a predefined set of words in natural language; and choosing a natural language word of said predefined set, corresponding to the entry with the lowest edit distance from said simplified and smoothed phonetic transcription, wherein at least some of said rules applied at said step iii) associate a response to a combination of at least two events jointly generated by said user and recognized at said step ii), of which at least one is not a word or sentence uttered by said user, said step ii) comprises searching a match between an acquired signal and an event belonging to a list of expected events stored in a memory of said humanoid robot, or accessible by it, said searching being carried out by successively using a plurality of matching methods with increasing complexity until an event is recognized with a confidence score greater than a predetermined value, or after the highest complexity recognition method has been used, and said matching methods include, by order of increasing complexity;
the search for an exact match, the search for an approximate match, the search for a phonetic correspondence—
only in the case of voice recognition—
and the search for a semantic correspondence. - View Dependent Claims (9)
-
-
12. A method of performing a dialogue between a humanoid robot and at least one user comprising the following steps, carried out iteratively by said humanoid robot:
-
i) acquiring a plurality of input signals from respective sensors, at least one said sensor being a sound sensor and at least one other sensor being a motion or image sensor; ii) interpreting the acquired signals to recognize a plurality of events generated by said user, selected from a group comprising;
the utterance of at least a word or sentence, an intonation of voice, a gesture, a body posture, a facial expression;iii) determining a response of said humanoid robot, comprising at least one event selected from a group comprising;
the utterance of at least a word or sentence, an intonation of voice, a gesture, a body posture, a facial expression, said determining being performed by applying a set of rules, each said rule associating a set of input events to a response of the robot;iv) generating said or each said event; and if the response determined during step iii) is or comprises at least the utterance of a word or sentence, the execution of a step iii-a) of performing linguistic analysis of the words or sentences to be uttered and determining an animation accompanying said response as a function of said analysis, said step iii-a comprises the substeps of; α
) identifying at least one word of the response to be animated;β
) determining a concept and expressiveness, called one-off expressiveness, associated with said or each said word to be animated; andγ
) choosing from a list of animations stored in a memory of said humanoid robot, or accessible by it, an animation based on said concept and said one-off expressiveness, whereinat least some of said rules applied at said step iii) associate a response to a combination of at least two events jointly generated by said user and recognized at said step ii), of which at least one is not a word or sentence uttered by said user. - View Dependent Claims (13, 14, 15, 16, 17)
-
-
18. A method of performing a dialogue between a humanoid robot and at least one user comprising the following steps, carried out iteratively by said humanoid robot:
-
i) acquiring a plurality of input signals from respective sensors, at least one said sensor being a sound sensor and at least one other sensor being a motion or image sensor; ii) interpreting the acquired signals to recognize a plurality of events generated by said user, selected from a group comprising;
the utterance of at least a word or sentence, an intonation of voice, a gesture, a body posture, a facial expression;iii) determining a response of said humanoid robot, comprising at least one event selected from a group comprising;
the utterance of at least a word or sentence, an intonation of voice, a gesture, a body posture, a facial expression, said determining being performed by applying a set of rules, each said rule associating a set of input events to a response of the robot;iv) generating said or each said event; and the following steps, implemented iteratively by said robot simultaneously with said steps i) to iv); A) determining the position of at least a portion of the body of said user relative to a reference frame fixed to the said robot; and B) driving at least one actuator of said robot to maintain the distance between said robot or an element thereof and said at least one or said body part of said user within a predefined range of values, wherein at least some of said rules applied at said step iii) associate a response to a combination of at least two events jointly generated by said user and recognized at said step ii), of which at least one is not a word or sentence uttered by said user. - View Dependent Claims (19, 20, 21, 22)
-
Specification