Efficient learning with robust gradient descent

作者:Matthew J. Holland, Kazushi Ikeda

摘要

Minimizing the empirical risk is a popular training strategy, but for learning tasks where the data may be noisy or heavy-tailed, one may require many observations in order to generalize well. To achieve better performance under less stringent requirements, we introduce a procedure which constructs a robust approximation of the risk gradient for use in an iterative learning routine. Using high-probability bounds on the excess risk of this algorithm, we show that our update does not deviate far from the ideal gradient-based update. Empirical tests using both controlled simulations and real-world benchmark data show that in diverse settings, the proposed procedure can learn more efficiently, using less resources (iterations and observations) while generalizing better.

论文关键词:Robust learning, Stochastic optimization, Statistical learning theory

论文评审过程:

论文官网地址:https://doi.org/10.1007/s10994-019-05802-5