Improved Multi-Class Cost-Sensitive Boosting via Estimation of the Minimum-Risk Class [article]

Ron Appel, Xavier Burgos-Artizzu, Pietro Perona
2016 arXiv   pre-print
We present a simple unified framework for multi-class cost-sensitive boosting. The minimum-risk class is estimated directly, rather than via an approximation of the posterior distribution. Our method jointly optimizes binary weak learners and their corresponding output vectors, requiring classes to share features at each iteration. By training in a cost-sensitive manner, weak learners are invested in separating classes whose discrimination is important, at the expense of less relevant
more » ... tion boundaries. Additional contributions are a family of loss functions along with proof that our algorithm is Boostable in the theoretical sense, as well as an efficient procedure for growing decision trees for use as weak learners. We evaluate our method on a variety of datasets: a collection of synthetic planar data, common UCI datasets, MNIST digits, SUN scenes, and CUB-200 birds. Results show state-of-the-art performance across all datasets against several strong baselines, including non-boosting multi-class approaches.
arXiv:1607.03547v2 fatcat:y5ebtcywyrgtdly3y2mxbydkny