Coarticulation method for audio-visual text-to-speech synthesis
First Claim
Patent Images
1. A method of synchronizing synthesized speech and animation, the method comprising:
- associating, by a computing device, a received stimulus with a phoneme having corresponding mouth parameters in a coarticulation library;
selecting, by the computing device, a parameter set corresponding to the mouth parameters from an animation library, the parameter set representing frame segments; and
generating, via a noise producing entity, speech associated with the stimulus that is synchronized with the frame segments and overlaying the frame segments on a larger entity to synthesize a whole animated image.
4 Assignments
0 Petitions
Accused Products
Abstract
A method for generating animated sequences of talking heads in text-to-speech applications wherein a processor samples a plurality of frames comprising image samples. The processor reads first data comprising one or more parameters associated with noise-producing orifice images of sequences of at least three concatenated phonemes which correspond to an input stimulus. The processor reads, based on the first data, second data comprising images of a noise-producing entity. The processor generates an animated sequence of the noise-producing entity.
19 Citations
17 Claims
-
1. A method of synchronizing synthesized speech and animation, the method comprising:
-
associating, by a computing device, a received stimulus with a phoneme having corresponding mouth parameters in a coarticulation library; selecting, by the computing device, a parameter set corresponding to the mouth parameters from an animation library, the parameter set representing frame segments; and generating, via a noise producing entity, speech associated with the stimulus that is synchronized with the frame segments and overlaying the frame segments on a larger entity to synthesize a whole animated image. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10)
-
-
11. A system for synchronizing synthesized speech and animation, the system comprising:
-
a processor; a first module controlling the processor to associate a received stimulus with a phoneme having corresponding mouth parameters in a coarticulation library; a second module controlling the processor to select a parameter set corresponding to the mouth parameters from an animation library, the parameter set representing frame segments; and a third module controlling the processor to generate, via a noise producing entity, speech associated with the stimulus that is synchronized with the frame segments and to overlay the frame segments on a larger entity to synthesize a whole animated image. - View Dependent Claims (12, 13, 14, 15, 16)
-
-
17. A method of synchronizing synthesized speech and animation, the method comprising:
-
associating, by a computing device, a received stimulus with a phoneme having corresponding mouth parameters in a coarticulation library; selecting, by the computing device, a parameter set corresponding to the mouth parameters from an animation library, the parameter set representing frame segments; and generating, via a noise producing entity, speech associated with the stimulus that is synchronized with the frame segments and overlaying the frame segments on a larger entity to synthesize a whole animated image.
-
Specification