Learning rates for kernel-based expectile regression

作者:Muhammad Farooq, Ingo Steinwart

摘要

Conditional expectiles are becoming an increasingly important tool in finance as well as in other areas of applications. We analyse a support vector machine type approach for estimating conditional expectiles and establish learning rates that are minimax optimal modulo a logarithmic factor if Gaussian RBF kernels are used and the desired expectile is smooth in a Besov sense. As a special case, our learning rates improves the best known rates for kernel-based least squares regression in aforementioned scenario. Key ingredients of our statistical analysis are a general calibration inequality for the asymmetric least squares loss, a corresponding variance bound as well as an improved entropy number bound for Gaussian RBF kernels.

论文关键词:Support vector machines, Self-calibration inequality, Variance bound, Entropy number bound, Learning rates

论文评审过程:

论文官网地址:https://doi.org/10.1007/s10994-018-5762-9