The Internet Archive has a preservation copy of this work in our general collections.
The file type is application/pdf
.
Max-Margin Stacking and Sparse Regularization for Linear Classifier Combination and Selection
[article]
2011
arXiv
pre-print
The main principle of stacked generalization (or Stacking) is using a second-level generalizer to combine the outputs of base classifiers in an ensemble. In this paper, we investigate different combination types under the stacking framework; namely weighted sum (WS), class-dependent weighted sum (CWS) and linear stacked generalization (LSG). For learning the weights, we propose using regularized empirical risk minimization with the hinge loss. In addition, we propose using group sparsity for
arXiv:1106.1684v1
fatcat:nlhcz5m3ejdnxotat25ov7s6qu