High-order entropy error functions for neural classifiers
First Claim
Patent Images
1. An artificial neural network classifier comprising:
- an input layer of one or more nodes to receive training data;
an output layer L of one or more nodes to provide an actual output indicating a level of training of the artificial neural network classifier based on the training data; and
at least a hidden layer of one or more nodes intermediate said input layer and said output layer, each node to receive input values via the input layer, each node to perform a transformation of the received input values based on a set of weights, each transformation to determine in part the actual output of output layer L, the set of weights to be updated based at least in part on an error function having an operator of the form where yjL is an actual output at node j of output layer L, tj is a target output at the node j, and n is greater than or equal to two, wherein the updated set of weights is to be used in determining confidence level information for a feature vector.
1 Assignment
0 Petitions
Accused Products
Abstract
An automatic speech recognition system comprising a speech decoder to resolve phone and word level information, a vector generator to generate information vectors on which a confidence measure is based by a neural network classifier (ANN). An error signal is designed which is not subject to false saturation or over specialization. The error signal is integrated into an error function which is back propagated through the ANN.
-
Citations
12 Claims
-
1. An artificial neural network classifier comprising:
-
an input layer of one or more nodes to receive training data; an output layer L of one or more nodes to provide an actual output indicating a level of training of the artificial neural network classifier based on the training data; and at least a hidden layer of one or more nodes intermediate said input layer and said output layer, each node to receive input values via the input layer, each node to perform a transformation of the received input values based on a set of weights, each transformation to determine in part the actual output of output layer L, the set of weights to be updated based at least in part on an error function having an operator of the form where yjL is an actual output at node j of output layer L, tj is a target output at the node j, and n is greater than or equal to two, wherein the updated set of weights is to be used in determining confidence level information for a feature vector. - View Dependent Claims (2, 3)
-
-
4. An automatic speech recognition system comprising:
-
a speech decoder to decode acoustic data; a vector generator to calculate a feature vector based on the decoded acoustic data; and an adaptive neural network classifier to determine a confidence in the calculated feature vector, the determining based on a training of the adaptive neural classifier according to an error function having an operator of the form where yjL is an actual output at node j of an output layer L of the adaptive neural network classifier, tj is a target output at the node j, and n is greater than or equal to two. - View Dependent Claims (5, 6, 7, 8)
where “
min” and
“
max”
are the minimum and maximum value of feature x respectively in a set of training data.
-
-
6. The automatic speech recognition system according to claim 4 wherein said vector generator comprises calculation circuits to produce an 8-dimensional feature vector.
-
7. The automatic speech recognition system according to claim 6 wherein said 8 dimensions comprise A-stabil, LogAWE-end, NscoreQ, N-active-leafs, NScore, Score-per-frame, Duration.
-
8. The automatic speech recognition system according to claim 4 wherein said classifier comprises a multi-layer perceptron (MLP).
-
9. A method for training a neural network classifier to produce an output indicative of a confidence level for a decoded word, the method comprising:
-
providing a first training pattern to the neural network classifier having an output layer L of one or more nodes, the output layer L to provide an output indicating a level of training of the neural network classifier; forward propagating the training pattern through the neural network classifier, the propagating based in part on a set of weights; determining an error signal based on the propagated training pattern, the error based on an error function having an operator of the form where yjL is an actual output at node j of output layer L of the neural network classifier, tj is the value of a target output at the node j, and n is greater than or equal to two; and updating the set of weights based on the determined error signal, the updated set of weights to be used in producing the output indicative of the confidence level for the decoded word. - View Dependent Claims (10)
-
-
11. A computer-readable medium storing instructions thereon which when executed by one or more processors cause the one or more processors to perform the method of:
-
implementing a multiple layer perceptron having an output layer L of one or more nodes, the output layer L to provide an output indicating a level of training of the multiple layer perceptron, performing, via the implemented multiple layer perceptron, a series of linear transformations of a set of training data, the linear transformations based on a set of weights; determining an error signal for an output of the linear transformations, the error signal based on an error function having an operator of the form where yjL is an actual output at node j of output layer L, tj is a target output at a node j, and n is greater than or equal to two; updating the set of weights based at least in part on the error signal; receiving a multiple-dimension feature vector based on decoded acoustic data; and determining, via the multiple layer perceptron, a confidence level of the feature vector, the determining based on the updated set of weights. - View Dependent Claims (12)
-
Specification