Published January 1, 2017
| Version v1
Journal article
Open
Highly efficient nonlinear regression for big data with lexicographical splitting
Creators
- 1. Bilkent Univ, Dept Elect & Elect Engn, TR-06800 Ankara, Turkey
- 2. Turk Telekom Commun Serv Inc, Istanbul, Turkey
Description
This paper considers the problem of online piecewise linear regression for big data applications. We introduce an algorithm, which sequentially achieves the performance of the best piecewise linear (affine) model with optimal partition of the space of the regressor vectors in an individual sequence manner. To this end, our algorithm constructs a class of sequential piecewise linear models over a set of partitions of the regressor space and efficiently combines them in the mixture-of-experts setting. We show that the algorithm is highly efficient with computational complexity of only , where m is the dimension of the regressor vectors. This efficient computational complexity is achieved by efficiently representing all of the models using a "lexicographical splitting graph." We analyze the performance of our algorithm without any statistical assumptions, i.e., our results are guaranteed to hold. Furthermore, we demonstrate the effectiveness of our algorithm over the well-known data sets in the machine learning literature with computational complexity fraction of the state of the art.
Files
bib-9b5a5784-b3d1-4919-88f8-fbbbef70bebb.txt
Files
(195 Bytes)
| Name | Size | Download all |
|---|---|---|
|
md5:d964c7e3512a734f3b2ed25760af6a13
|
195 Bytes | Preview Download |