Detecting emotions using voice signal analysis
First Claim
1. A method of detecting an emotional state, the method comprising:
- providing a speech signal;
dividing the speech signal into at least one of segments, frames, and subframes;
extracting at least one acoustic feature from the speech signal;
calculating statistics from the at least one acoustic feature;
classifying the speech with at least one neural network classifier as belonging to at least one emotional state; and
storing in memory and outputting in a human-recognizable format an indication of the at least one emotional state, wherein the speech is classified by a classifier taught to recognize at least one emotional state from a finite number of emotional states.
3 Assignments
0 Petitions
Accused Products
Abstract
A system and method are provided for detecting emotional states using statistics. First, a speech signal is received. At least one acoustic parameter is extracted from the speech signal. Then statistics or features from samples of the voice are calculated from extracted speech parameters. The features serve as inputs to a classifier, which can be a computer program, a device or both. The classifier assigns at least one emotional state from a finite number of possible emotional states to the speech signal. The classifier also estimates the confidence of its decision. Features that are calculated may include a maximum value of a fundamental frequency, a standard deviation of the fundamental frequency, a range of the fundamental frequency, a mean of the fundamental frequency, and a variety of other statistics.
331 Citations
59 Claims
-
1. A method of detecting an emotional state, the method comprising:
-
providing a speech signal;
dividing the speech signal into at least one of segments, frames, and subframes;
extracting at least one acoustic feature from the speech signal;
calculating statistics from the at least one acoustic feature;
classifying the speech with at least one neural network classifier as belonging to at least one emotional state; and
storing in memory and outputting in a human-recognizable format an indication of the at least one emotional state, wherein the speech is classified by a classifier taught to recognize at least one emotional state from a finite number of emotional states. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59)
-
-
18. A system for classifying speech, the system comprising:
-
a computer system comprising a central processing unit, an input device, at least one memory for storing data indicative of a speech signal, and an output device;
logic for receiving and analyzing a speech signal;
logic for dividing the speech signal;
logic for extracting at least one feature from the speech signal;
logic for calculating statistics of the speech;
logic for at least one neural network for classifying the speech as belonging to at least one of a finite number of emotional states; and
logic for storing in memory and outputting an indication of the at least one emotional state.
-
-
35. A method of recognizing emotional states in a voice, the method comprising:
-
providing a first plurality and a second plurality of voice samples;
identifying each sample of said pluralities of samples as belonging to a predominant emotional state;
dividing each sample into at least one of frames, subframes, and segments;
extracting at least one acoustic feature for each sample of the pluralities of samples;
calculating statistics of the speech samples from the at least one feature;
classifying an emotional state in the first plurality of samples with at least one neural network;
training the at least one neural network to recognize an emotional state from the statistics by comparing the results of identifying and recognizing for the first plurality of samples;
classifying an emotion in the second plurality of voice samples with the at least one trained neural network; and
storing in memory and outputting in a human-recognizable format an indication of the emotional state.
-
-
46. A system for detecting an emotional state in a voice signal, the system comprising:
-
a speech reception device;
at least one computer connected to the speech reception device;
at least one memory operably connected to the at least one computer;
a computer program including at least one neural network for dividing the voice signal into a plurality of segments, and for analyzing the segments according to features of the segments to detect the emotional state in the voice signal;
a database of speech signal features and statistics accessible to the computer for comparison with features of the voice signal; and
an output device coupled to the computer for notifying a user of the emotional state detected in the voice signal.
-
Specification