Adaptive bayes feature extraction
First Claim
1. A computer-implemented method for extracting discriminately informative features from input patterns, which provide discrimination between two classes, a class-of-interest and a class-other, while reducing the number of features, comprising the steps of:
- receiving a training set of class-of-interest patterns, a set of unlabeled patterns from an input-data-set, and an estimate of a class-of-interest a priori probability in said input-data-set, said input-data-set being at least one of an image, video or speech data set;
selecting elements of a predetermined polynomial function;
executing a training stage using said class-of-interest a priori probability, said training set of class-of-interest patterns, and said unlabeled patterns from said input-data-set, said training stage including a step of selecting a set of weights for said polynomial function that ensure a least squares approximation of a class-of-interest posterior distribution function using said polynomial function;
classifying said pattern from said input-data-set as being either said class-of-interest or said class-other in accordance with a conditional test defined by an adaptive Bayes decision rule;
extracting a predetermined percent of said classified patterns that lie near a decision boundary;
locating points lying on said decision boundary using said extracted patterns that lie near said decision boundary;
calculating normal vectors to said decision boundary using said points lying on said decision boundary;
calculating an effective decision boundary feature matrix;
calculating eigenvalues, eigenvectors, and a rank of said effective decision boundary feature matrix;
selecting a set of said eigenvectors for use in a feature extraction matrix; and
extracting a reduced set of features using said feature extraction matrix,whereby said discriminately informative features are extracted from input patterns which provide discrimination between a class-of-interest and a class-other while reducing the number of features, using only said training set of class-of-interest patterns, and said unlabeled patterns from said input-data-set, and without any a priori knowledge of said class-other.
0 Assignments
0 Petitions
Accused Products
Abstract
A system and method for extracting “discriminately informative features” from input patterns which provide accurate discrimination between two classes, a class-of-interest and a class-other, while reducing the number of features under the condition where training samples or otherwise, are provided a priori only for the class-of-interest thus eliminating the requirement for any a priori knowledge of the other classes in the input-data-set while exploiting the potentially robust and powerful feature extraction capability provided by fully supervised feature extraction approaches. The system and method extracts discriminate features by exploiting the ability of the adaptive Bayes classifier to define an optimal Bayes decision boundary between the class-of-interest and class-other using only labeled samples from the class-of-interest and unlabeled samples from the data to be classified. Optimal features are derived from vectors normal to the decision boundary defined by the adaptive Bayes classifier.
-
Citations
27 Claims
-
1. A computer-implemented method for extracting discriminately informative features from input patterns, which provide discrimination between two classes, a class-of-interest and a class-other, while reducing the number of features, comprising the steps of:
-
receiving a training set of class-of-interest patterns, a set of unlabeled patterns from an input-data-set, and an estimate of a class-of-interest a priori probability in said input-data-set, said input-data-set being at least one of an image, video or speech data set; selecting elements of a predetermined polynomial function; executing a training stage using said class-of-interest a priori probability, said training set of class-of-interest patterns, and said unlabeled patterns from said input-data-set, said training stage including a step of selecting a set of weights for said polynomial function that ensure a least squares approximation of a class-of-interest posterior distribution function using said polynomial function; classifying said pattern from said input-data-set as being either said class-of-interest or said class-other in accordance with a conditional test defined by an adaptive Bayes decision rule; extracting a predetermined percent of said classified patterns that lie near a decision boundary; locating points lying on said decision boundary using said extracted patterns that lie near said decision boundary; calculating normal vectors to said decision boundary using said points lying on said decision boundary; calculating an effective decision boundary feature matrix; calculating eigenvalues, eigenvectors, and a rank of said effective decision boundary feature matrix; selecting a set of said eigenvectors for use in a feature extraction matrix; and extracting a reduced set of features using said feature extraction matrix, whereby said discriminately informative features are extracted from input patterns which provide discrimination between a class-of-interest and a class-other while reducing the number of features, using only said training set of class-of-interest patterns, and said unlabeled patterns from said input-data-set, and without any a priori knowledge of said class-other. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12)
-
-
13. A computer-implemented method for extracting discriminately informative features from input patterns, which provide discrimination between a class-of-interest and a class-other while reducing the number of features, comprising the steps of:
-
receiving a training set of class-of-interest patterns, a set of unlabeled patterns from an input-data-set, and an estimate of a class-of-interest a priori probability in said input-data-set, said input-data-set being at least one of an image, video or speech data set; selecting a predetermined number of Gaussian kernel densities functions; selecting parameter values for said Gaussian kernel densities functions where said selected parameter values cause said Gaussian kernel densities to approximate the probability density function of said input-data-set; executing a training stage using said a priori probability of said class-of-interest, said training set of class-of-interest patterns, and said unlabeled patterns from said input-data-set, said training stage including a step of least squares approximation of a class-of-interest posterior distribution function using a linear combination of weighted said Gaussian kernel density functions; classifying said pattern from said input-data-set as being either said class-of-interest or said class-other in accordance with a conditional test defined by a adaptive Bayes decision rule; extracting a predetermined percent of said classified patterns that lie near a decision boundary; locating points lying on said decision boundary using said extracted patterns that lie near said decision boundary; calculating normal vectors to said decision boundary using said points lying on said decision boundary; calculating an effective decision boundary feature matrix; calculating eigenvalues, eigenvectors, and rank of said effective decision boundary feature matrix; selecting a set of eigenvectors for use in a feature extraction matrix; and extracting a reduced set of features using said feature extraction matrix, whereby said discriminately informative features are extracted from input patterns which provide discrimination between a class-of-interest and a class-other while reducing the number of features, using only said training set of class-of-interest patterns, and said unlabeled patterns from said input-data-set, and without any a priori knowledge of said class-other. - View Dependent Claims (14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27)
-
Specification