# Item

ITEM ACTIONSEXPORT

Released

Journal Article

#### Sparse Regression Ensembles in Infinite and Finite Hypothesis Spaces

##### MPS-Authors

There are no MPG-Authors in the publication available

##### External Resource

https://link.springer.com/content/pdf/10.1023/A:1013907905629.pdf

(Publisher version)

##### Fulltext (restricted access)

There are currently no full texts shared for your IP range.

##### Fulltext (public)

There are no public fulltexts stored in PuRe

##### Supplementary Material (public)

There is no public supplementary material available

##### Citation

Rätsch, G., Demiriz, A., & Bennett, K. (2002). Sparse Regression Ensembles in Infinite
and Finite Hypothesis Spaces.* Machine Learning,* *48*,
193-221. doi:10.1023/A:1013907905629.

Cite as: https://hdl.handle.net/11858/00-001M-0000-0013-E0BE-E

##### Abstract

We examine methods for constructing regression ensembles based on a linear program (LP). The ensemble regression function consists of linear combinations of base hypotheses generated by some boosting-type base learning algorithm. Unlike the classification case, for regression the set of possible hypotheses producible by the base learning algorithm may be infinite. We explicitly tackle the issue of how to define and solve ensemble regression when the hypothesis space is infinite. Our approach is based on a semi-infinite linear program that has an infinite number of constraints and a finite number of variables. We show that the regression problem is well posed for infinite hypothesis spaces in both the primal and dual spaces. Most importantly, we prove there exists an optimal solution to the infinite hypothesis space problem consisting of a finite number of hypothesis. We propose two algorithms for solving the infinite and finite hypothesis problems. One uses a column generation simplex-type algorithm and the other adopts an exponential barrier approach. Furthermore, we give sufficient conditions for the base learning algorithm and the hypothesis set to be used for infinite regression ensembles. Computational results show that these methods are extremely promising.