Method and system for aligning natural and synthetic video to speech synthesis
First Claim
1. A method for encoding a facial animation comprising the steps of:
- a) creating a data stream;
b) creating a facial mimic stream including a plurality of facial animation parameters;
c) inserting a plurality of time stamps in the data stream pointing to said plurality of facial animation parameters, wherein said plurality of time stamps establishes a synchronization relationship with said data stream and said facial mimic stream; and
d) encoding said data stream and said facial mimic stream.
4 Assignments
0 Petitions
Accused Products
Abstract
According to MPEG-4'"'"'s TTS architecture, facial animation can be driven by two streams simultaneously—text, and Facial Animation Parameters. In this architecture, text input is sent to a Text-To-Speech converter at a decoder that drives the mouth shapes of the face. Facial Animation Parameters are sent from an encoder to the face over the communication channel. The present invention includes codes (known as bookmarks) in the text string transmitted to the Text-to-Speech converter, which bookmarks are placed between words as well as inside them. According to the present invention, the bookmarks carry an encoder time stamp. Due to the nature of text-to-speech conversion, the encoder time stamp does not relate to real-world time, and should be interpreted as a counter. In addition, the Facial Animation Parameter stream carries the same encoder time stamp found in the bookmark of the text. The system of the present invention reads the bookmark and provides the encoder time stamp as well as a real-time time stamp to the facial animation system. Finally, the facial animation system associates the correct facial animation parameter with the real-time time stamp using the encoder time stamp of the bookmark as a reference.
42 Citations
12 Claims
-
1. A method for encoding a facial animation comprising the steps of:
-
a) creating a data stream;
b) creating a facial mimic stream including a plurality of facial animation parameters;
c) inserting a plurality of time stamps in the data stream pointing to said plurality of facial animation parameters, wherein said plurality of time stamps establishes a synchronization relationship with said data stream and said facial mimic stream; and
d) encoding said data stream and said facial mimic stream. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8)
-
-
9. A method for encoding a facial animation including at least one facial mimic and speech in the form of a text stream, comprising the steps of:
-
a) assigning a predetermined code to the at least one facial mimic;
b) placing the predetermined code within the text stream, wherein said predetermined code indicates a presence of a particular facial mimic and wherein said predetermined code points to a stream of facial mimics, thereby indicating a synchronization relationship between the text stream and the facial mimic stream;
c) encoding said text stream; and
d) placing the predetermined code in between letters in the text stream.
-
-
10. A method for encoding a facial animation including at least one facial mimic and speech in the form of a text stream, comprising the steps of:
-
a) assigning a predetermined code to the at least one facial mimic;
b) placing the predetermined code within the text stream, wherein said predetermined code indicates a presence of a particular facial mimic and wherein said predetermined code points to a stream of facial mimics, thereby indicating a synchronization relationship between the text stream and the facial mimic stream;
c) encoding said text stream; and
d) placing the predetermined code inside words in the text stream.
-
-
11. A method for decoding a facial animation including speech and at least one facial mimic, comprising the steps of:
-
a) monitoring a text stream for a set of predetermined codes corresponding to a set of facial mimics, wherein the predetermined code points to a stream of facial mimics established during an encoding process of said text stream, thereby indicating a synchronization relationship between the text stream and the facial mimic stream;
b) sending a signal to a visual decoder to start a particular facial mimic upon detecting the presence of the set of predetermined codes; and
c) placing the predetermined code in between phonemes in the text stream.
-
-
12. A method for decoding a facial animation including speech and at least one facial mimic, comprising the steps of:
-
a) monitoring a text stream for a set of predetermined codes corresponding to a set of facial mimics, wherein the predetermined code points to a stream of facial mimics established during an encoding process of said text stream, thereby indicating a synchronization relationship between the text stream and the facial mimic stream;
b) sending a signal to a visual decoder to start a particular facial mimic upon detecting the presence of the set of predetermined codes; and
c) placing the predetermined code inside words in the text stream.
-
Specification