Amichai Painsky , School of Mathematical Sciences, Tel Aviv University, Ramat Aviv, Israel, 6997801.(email:email@example.com)
Recursive partitioning methods producing tree-like models are a long standing staple of predictive modeling. However, a fundamental flaw in the partitioning (or splitting) rule of commonly used tree building methods precludes them from treating different types of variables equally. This most clearly manifests in these methods’ inability to properly utilize categorical variables with a large number of categories, which are ubiquitous in the new age of big data. We propose a framework to splitting using leave-one-out (LOO) cross validation (CV) for selecting the splitting variable, then performing a regular split (in our case, following CART’s approach) for the selected variable. The most important consequence of our approach is that categorical variables with many categories can be safely used in tree building and are only chosen if they contribute to predictive power. We demonstrate in extensive simulation and real data analysis that our splitting approach significantly improves the performance of both single tree models and ensemble methods that utilize trees. Importantly, we design an algorithm for LOO splitting variable selection which under reasonable assumptions does not substantially increase the overall computational complexity compared to CART for two-class classification.
Gradient Boosting, Classification and regression trees, Random Forests
Amichai Painsky, Saharon Rosset, "Cross-Validated Variable Selection in Tree-Based Methods Improves Predictive Performance", IEEE Transactions on Pattern Analysis & Machine Intelligence, vol. , no. , pp. 1, 5555, doi:10.1109/TPAMI.2016.2636831