×

Classification method and apparatus based on boosting and pruning of multiple classifiers

  • US 6,456,991 B1
  • Filed: 09/01/1999
  • Issued: 09/24/2002
  • Est. Priority Date: 09/01/1999
  • Status: Expired due to Fees
First Claim
Patent Images

1. A neural network classifier boosting and pruning method including the steps of:

  • (a) providing a set of training data having inputs with correct classifications corresponding to the inputs;

    (b) ordering the set of training data into a plurality Y of differently ordered data sets, where Y is the total number of differently ordered data sets and y represents a particular one of the Y differently ordered data sets;

    (c) dividing each of the Y differently ordered data sets into a set of N particular data portions Dn,y where N represents the total number of data portions into which the particular data set y was divided and n identifies a particular data portion among the N data portions derived from a particular data set y;

    (d) associating a plurality X,y of booster classifiers Bx,y for each particular data set y, where X represents the total number of booster types associated with each particular data set y and x represents a particular booster type, the plurality X,y of booster classifiers Bx,y for each particular data set y including a terminal booster;

    (e) training one of each plurality X,y of booster classifiers Bx,y with a particular data portion Dn,y, resulting in a plurality of rules;

    (f) testing the booster classifier Bx,y, trained in step (e) utilizing another particular data portion, Dn,y, said testing resulting in correctly classified data and mistakenly classified data;

    (g) creating a particular training data set Tx,y corresponding to each particular data set y, the contents of which are defined by;

    Tx,y=(w1,x,y)

    M

    (Bx,y

    (Dx+1,y)
    )
    +(w2,x,y)

    C

    (B1,y

    (D1,y)
    )
    +

    j=1x-1

    (x-1>

    0
    )






    (w2+j,x,y)

    C

    (Bx-j+1,y

    (Dx-j+1,y)
    )
    embedded image

    where M(Bx,y (•

    )) identifies data upon which mistakes were made by the associated booster classifier Bx,y, C(Bx,y (•

    )) identifies data for which correct classifications were made by the associated booster classifier Bx,y, and w represents independently selectable apportionment factors which apportion the amount of mistakenly classified data and the amount of correctly classified data, and j is a summation index which ranges from 1 to x−

    1 where x−

    1>

    0 and represents the total number booster classifiers Bx,y from which correct examples are used for the particular training set Tx,y;

    (h) training one of the plurality X,y of booster classifiers Bx,y associated with each particular data set y with the particular training data set Tx,y created in step (g) which corresponds to the same particular data set y, resulting in a plurality of rules;

    (i) testing the booster classifier Bx,y trained in step (h) utilizing another particular data portion Dn,y from the same particular data set y, said testing resulting in correctly classified data and mistakenly classified data;

    (j) repeating steps (g) through (i) X−

    3 additional times until a total of X−

    1 booster classifiers Bx,y have been trained and tested for each particular data set y;

    (k) testing the X−

    1 booster classifiers Bx,y trained in steps (e) through (j) for each particular data set y utilizing the particular data portion y associated with each X−

    1 booster classifiers Bx,y, said testing resulting in correctly classified data and mistakenly classified data;

    (l) creating a residual training set TR,y for each particular data set y including the mistakenly classified data resulting from step (k);

    (m) training the terminal booster for each particular data set y utilizing the residual training set TR,y created in step (l);

    resulting in a plurality of rules for use in a classification apparatus for receiving data for classification and, based on the classficiation data, outputting classfications.

View all claims
  • 1 Assignment
Timeline View
Assignment View
    ×
    ×