Method for detecting emotions involving subspace specialists
First Claim
Patent Images
1. A method for detecting emotions from speech input comprising:
- evaluating, deriving and/or extracting at least a first feature class and a second feature class of features from a given speech input, wherein the first feature class does not include features of the second feature class and the second feature class does not include features of the first feature class, and the first feature class includes prosodic features and the second feature class includes voice quality features;
associating said first and second feature classes with dimensions of an underlying emotional space including a first dimension of activation or arousal and a second dimension of evaluation or pleasure, respectively;
using for each dimension of the underlying emotional space a specialized classifier system, each of which being configured to classify features from an assigned feature class associated with a respective feature class, wherein each specialized classifier system operates independently from each other, and each of the specialized classifier systems uses as input only features of a respectively assigned feature class; and
combining outputs of said specialized classifier systems for each feature class to form a global classifier system configured to output a current emotional state.
2 Assignments
0 Petitions
Accused Products
Abstract
To detect and determine a current emotional state (CES) of a human being from a spoken speech input (SI), it is suggested in a method for detecting emotions to identify first and second feature classes (A, E) with, in particular distinct, dimensions of an underlying emotional manifold (EM) or emotional space (ES) and/or with subspaces thereof.
18 Citations
10 Claims
-
1. A method for detecting emotions from speech input comprising:
-
evaluating, deriving and/or extracting at least a first feature class and a second feature class of features from a given speech input, wherein the first feature class does not include features of the second feature class and the second feature class does not include features of the first feature class, and the first feature class includes prosodic features and the second feature class includes voice quality features; associating said first and second feature classes with dimensions of an underlying emotional space including a first dimension of activation or arousal and a second dimension of evaluation or pleasure, respectively; using for each dimension of the underlying emotional space a specialized classifier system, each of which being configured to classify features from an assigned feature class associated with a respective feature class, wherein each specialized classifier system operates independently from each other, and each of the specialized classifier systems uses as input only features of a respectively assigned feature class; and combining outputs of said specialized classifier systems for each feature class to form a global classifier system configured to output a current emotional state. - View Dependent Claims (2, 3, 4)
-
-
5. A method of detecting emotions, comprising:
-
providing a first feature class and a second feature class of features of a speech input, wherein said first feature class comprises prosodic features and corresponds to a first dimension of an emotional space, and wherein said second feature class comprises voice quality features and corresponds to a second dimension of said emotional space, wherein the first feature class does not include features of the second feature class and the second feature class does not include features of the first feature class; using for said first and second feature classes specialized classifier systems including a first and second classifier, respectively, wherein said first and second classifiers are configured to classify features of said first and second feature classes, respectively, said second classifier includes a plurality of single classifiers, said first classifier and said second classifier operating independently from one another, and each of the specialized classifier systems uses as input only features of a respectively assigned feature class; and combining outputs of said first and second classifiers to form a global classifier configured to output a current emotional state. - View Dependent Claims (6, 7, 8, 10)
-
-
9. A method of detecting emotions, comprising:
-
providing a first feature class and a second feature class of features of a speech input, wherein said first feature class comprises prosodic features and corresponds to a first dimension of an emotional space, and wherein said second feature class comprises voice quality features and corresponds to a second dimension of said emotional space, wherein the first feature class does not include features of the second feature class and the second feature class does not include features of the first feature class; using for said first and second feature classes specialized classifier systems including a first and second classifier, respectively, wherein said first and second classifiers are configured to classify features of said first and second feature classes, respectively, said first and second classifiers operate independently from each other, and each of the specialized classifier systems uses as input only features of a respectively assigned feature class; and combining outputs of said first and second classifiers to form a global classifier configured to output a current emotional state.
-
Specification