Title
BDT: Gradient Boosted Decision Tables for High Accuracy and Scoring Efficiency
Abstract
In this paper we present gradient boosted decision tables (BDTs). A d-dimensional decision table is essentially a mapping from a sequence of d boolean tests to a real value in {R}. We propose novel algorithms to fit decision tables. Our thorough empirical study suggests that decision tables are better weak learners in the gradient boosting framework and can improve the accuracy of the boosted ensemble. In addition, we develop an efficient data structure to represent decision tables and propose a novel fast algorithm to improve the scoring efficiency for boosted ensemble of decision tables. Experiments on public classification and regression datasets demonstrate that our method is able to achieve 1.5x to 6x speedups over the boosted regression trees baseline. We complement our experimental evaluation with a bias-variance analysis that explains how different weak models influence the predictive power of the boosted ensemble. Our experiments suggest gradient boosting with randomly backfitted decision tables distinguishes itself as the most accurate method on a number of classification and regression problems. We have deployed a BDT model to LinkedIn news feed system and achieved significant lift on key metrics.
Year
DOI
Venue
2017
10.1145/3097983.3098175
KDD
Keywords
Field
DocType
classification,regression,decision table,gradient boosting
Data mining,Data structure,Decision table,Predictive power,Regression,Computer science,Artificial intelligence,Regression problems,Empirical research,Machine learning,Gradient boosting
Conference
ISBN
Citations 
PageRank 
978-1-4503-4887-4
4
0.42
References 
Authors
11
2
Name
Order
Citations
PageRank
Yin Lou150628.82
Mikhail Obukhov280.84