System for constructing decision tree classifiers using structure-driven induction
First Claim
1. A computer-implemented method for constructing a decision tree for computer-implemented information processing, comprising the steps of:
- constructing a tree structure of a predetermined size with empty internal nodes, including at least one terminal node and at least one internal node;
using training vectors of predetermined classification to substantially simultaneously determine splits for each decision tree node, wherein;
(a) said training vectors include a back propagation component that determines said splits for each internal node, and (b) said training vectors include a competitive learning component that controls the number of terminal nodes thereby determining the effective size of said decision tree.
1 Assignment
0 Petitions
Accused Products
Abstract
A computer-implemented apparatus and method for designing decision tree classifiers for use in artificial applications. A tree structure of fixed size with empty internal nodes, i.e. nodes without any splitting function, and labeled terminal nodes is first constructed. Using a collection of training vectors of known classification, a neural learning scheme combining backpropagation and soft competitive learning is then used to simultaneously determine the splits for each decision tree node. Compact trees are generated that have multifeature splits at each internal node which are determined on global rather than local basis. The computer-implemented apparatus and method consequently produces decision trees yielding better classification and interpretation of the underlying relationships in the data.
58 Citations
7 Claims
-
1. A computer-implemented method for constructing a decision tree for computer-implemented information processing, comprising the steps of:
-
constructing a tree structure of a predetermined size with empty internal nodes, including at least one terminal node and at least one internal node;
using training vectors of predetermined classification to substantially simultaneously determine splits for each decision tree node, wherein;
(a) said training vectors include a back propagation component that determines said splits for each internal node, and (b) said training vectors include a competitive learning component that controls the number of terminal nodes thereby determining the effective size of said decision tree. - View Dependent Claims (2, 3, 4)
using a computer-implemented neural learning scheme to determine the splits for each internal node through back propagation and to control the number of terminal nodes through competitive learning.
-
-
3. The computer-implemented method of claim 1 further comprising the step of:
generating a map between components of a neural network and the nodes of the decision tree for use in constructing the decision tree.
-
4. The computer-implemented method of claim 1 further comprising the steps of:
-
generating the training vectors through the neural network;
transforming a plurality of the neural network components into nodes of the decision tree based upon the generated map.
-
-
5. A computer-implemented apparatus for constructing a decision tree for computer-implemented information processing, comprising:
-
a tree structure data structure of a predetermined size with empty internal nodes, including at least one terminal node and at least one internal node;
a neural network for generating training vectors, said training vectors include a back propagation component that determines said splits for each internal node, and said training vectors include a competitive learning component that controls the number of terminal nodes thereby determining the effective size of said decision tree. - View Dependent Claims (6, 7)
a computer-implemented neural learning scheme to determine the splits for each internal node through back propagation and to control the number of terminal nodes through competitive learning.
-
-
7. The computer-implemented apparatus of claim 5 further comprising:
a map between components of a neural network and the nodes of the decision tree for use in constructing the decision tree.
Specification