Gradient Weights help Nonparametric Regressors
Kpotufe, Samory, Boularias, Abdeslam
–Neural Information Processing Systems
In regression problems over $\real^d$, the unknown function $f$ often varies more in some coordinates than in others. We show that weighting each coordinate $i$ with the estimated norm of the $i$th derivative of $f$ is an efficient way to significantly improve the performance of distance-based regressors, e.g. kernel and $k$-NN regressors. We propose a simple estimator of these derivative norms and prove its consistency. Moreover, the proposed estimator is efficiently learned online.
Neural Information Processing Systems
Dec-31-2012
- Country:
- Europe > Germany
- Baden-Württemberg > Tübingen Region > Tübingen (0.04)
- North America > United States
- California > Orange County
- Irvine (0.04)
- Illinois > Cook County
- Chicago (0.04)
- New York > New York County
- New York City (0.04)
- California > Orange County
- South America > Paraguay
- Europe > Germany
- Technology: