Alternating optimization of decision trees, with application to learning sparse oblique trees
Carreira-Perpinan, Miguel A., Tavallali, Pooya
–Neural Information Processing Systems
Learning a decision tree from data is a difficult optimization problem. The most widespread algorithm in practice, dating to the 1980s, is based on a greedy growth of the tree structure by recursively splitting nodes, and possibly pruning back the final tree. The parameters (decision function) of an internal node are approximately estimated by minimizing an impurity measure. We give an algorithm that, given an input tree (its structure and the parameter values at its nodes), produces a new tree with the same or smaller structure but new parameter values that provably lower or leave unchanged the misclassification error. This can be applied to both axis-aligned and oblique trees and our experiments show it consistently outperforms various other algorithms while being highly scalable to large datasets and trees.
Neural Information Processing Systems
Feb-14-2020, 07:42:40 GMT
- Technology: