Application of emotion-based intonation and prosody to speech in text-to-speech systems
First Claim
Patent Images
1. A method of converting text to speech, said method comprising the steps of:
- accepting text input;
providing synthetic speech output corresponding to the text input;
imparting emotion-based features to synthetic speech output;
said step of imparting emotion-based features comprising;
accepting instruction for imparting at least one emotion-based paradigm to synthetic speech output, wherein said step of accepting instruction further comprises accepting emotion-based commands from a user interface; and
applying at least one emotion-based paradigm to synthetic speech output, said step of applying at least one emotion-based paradigm to synthetic speech output comprising;
altering at least one segment to be used in synthetic speech output, whereby emotion in speech is reflected in how individual words or syllables are stressed;
altering at least one prosodic pattern to be used in synthetic speech output, whereby emotion in speech is reflected in prosodic patterns; and
selectably applying a single emotion-based paradigm over a single utterance of synthetic speech output;
orapplying a variable emotion-based paradigm over individual segments of an utterance of synthetic speech output.
4 Assignments
0 Petitions
Accused Products
Abstract
A text-to-speech system that includes an arrangement for accepting text input, an arrangement for providing synthetic speech output, and an arrangement for imparting emotion-based features to synthetic speech output. The arrangement for imparting emotion-based features includes an arrangement for accepting instruction for imparting at least one emotion-based paradigm to synthetic speech output, as well as an arrangement for applying at least one emotion-based paradigm to synthetic speech output.
35 Citations
4 Claims
-
1. A method of converting text to speech, said method comprising the steps of:
-
accepting text input; providing synthetic speech output corresponding to the text input; imparting emotion-based features to synthetic speech output; said step of imparting emotion-based features comprising; accepting instruction for imparting at least one emotion-based paradigm to synthetic speech output, wherein said step of accepting instruction further comprises accepting emotion-based commands from a user interface; and applying at least one emotion-based paradigm to synthetic speech output, said step of applying at least one emotion-based paradigm to synthetic speech output comprising; altering at least one segment to be used in synthetic speech output, whereby emotion in speech is reflected in how individual words or syllables are stressed; altering at least one prosodic pattern to be used in synthetic speech output, whereby emotion in speech is reflected in prosodic patterns; and selectably applying a single emotion-based paradigm over a single utterance of synthetic speech output;
orapplying a variable emotion-based paradigm over individual segments of an utterance of synthetic speech output. - View Dependent Claims (2, 3, 4)
-
Specification