DECISION TREE TRAINING IN MACHINE LEARNING
First Claim
1. A machine learning device comprising:
- a communications interface arranged to receive training data;
a tree training logic arranged to train a random decision forest using the received training data and on the basis of uncertainty measures of at least some of the received training data computed using an uncertainty measurement logic;
the uncertainty measurement logic arranged to either correct for bias in the uncertainty measurement or to use a non-parametric estimate of uncertainty.
3 Assignments
0 Petitions
Accused Products
Abstract
Improved decision tree training in machine learning is described, for example, for automated classification of body organs in medical images or for detection of body joint positions in depth images. In various embodiments, improved estimates of uncertainty are used when training random decision forests for machine learning tasks in order to give improved accuracy of predictions and fewer errors. In examples, bias corrected estimates of entropy or Gini index are used or non-parametric estimates of differential entropy. In examples, resulting trained random decision forests are better able to perform classification or regression tasks for a variety of applications without undue increase in computational load.
92 Citations
20 Claims
-
1. A machine learning device comprising:
-
a communications interface arranged to receive training data; a tree training logic arranged to train a random decision forest using the received training data and on the basis of uncertainty measures of at least some of the received training data computed using an uncertainty measurement logic; the uncertainty measurement logic arranged to either correct for bias in the uncertainty measurement or to use a non-parametric estimate of uncertainty. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12)
-
-
13. A machine learning method comprising:
-
receiving training data at a communications interface; training, at a processor, a random decision forest using the received training data and on the basis of a measure of uncertainty of at least some of the received training data; computing, at the processor, the measure of the uncertainty so as to either correct for bias in the measurement of the uncertainty or to use a non-parametric estimate of the uncertainty. - View Dependent Claims (14, 15, 16, 17)
-
-
18. A machine learning method comprising:
-
receiving training data at a communications interface the training data comprising examples of data to be classified into one of a plurality of possible classes; training, at a processor, a random decision forest to classify data into the possible classes, the training carried out using the received training data and on the basis of a measure of uncertainty of at least some of the received training data; computing, at the processor, the measure of the uncertainty so as to either correct for bias in the measurement of the uncertainty or to use a non-parametric estimate of the uncertainty; and
where the number of possible classes is such that it is difficult to estimate empirical class frequencies reliably. - View Dependent Claims (19, 20)
-
Specification