Skip to yearly menu bar Skip to main content


Poster

Smaller, more accurate regression forests using tree alternating optimization

Arman Zharmagambetov · Miguel Carreira-Perpinan

Virtual

Keywords: [ Other ] [ Boosting / Ensemble Methods ] [ Supervised Learning ]


Abstract:

Regression forests, based on ensemble approaches such as bagging or boosting, have long been recognized as the leading off-the-shelf method for regression. However, forests rely on a greedy top-down procedure such as CART to learn each tree. We extend a recent algorithm for learning classification trees, Tree Alternating Optimization (TAO), to the regression case, and use it with bagging to construct regression forests of oblique trees, having hyperplane splits at the decision nodes. In a wide range of datasets, we show that the resulting forests exceed the accuracy of state-of-the-art algorithms such as random forests, AdaBoost or gradient boosting, often considerably, while yielding forests that have usually fewer and shallower trees and hence fewer parameters and faster inference overall. This result has an immense practical impact and advocates for the power of optimization in ensemble learning.

Chat is not available.