An accelerated stochastic variance-reduced method for machine learning problems

作者:

Highlights:

摘要

Variance reduction techniques provide simple and fast algorithms for solving machine learning problems. In this paper, we present a novel stochastic variance-reduced method. The proposed method relies on the mini-batch version of stochastic recursive gradient algorithm (MB-SARAH), which updates stochastic gradient estimates by using a simple recursive scheme. However, facing the challenge of the step size sequence selection in MB-SARAH, we introduce an online step size sequence based on the hypergradient descent (HD) method, which only requires little additional computation. For the proposed method, referred to as MB-SARAH-HD, we provide a general convergence analysis and prove linear convergence for strongly convex problems in expectation. Specifically, we prove that the proposed method has sublinear convergence rate in a single outer loop. We also prove that the iteration complexity outperforms several variants of the state-of-the-art stochastic gradient descent (SGD) method under suitable conditions. Numerical experiments on standard datasets are provided to demonstrate the efficacy and superiority of our MB-SARAH-HD method over existing approaches in the literature.

论文关键词:Stochastic optimization,Variance reduction,Hypergradient,Recursive gradient

论文评审过程:Received 5 July 2019, Revised 16 April 2020, Accepted 18 April 2020, Available online 23 April 2020, Version of Record 23 April 2020.

论文官网地址:https://doi.org/10.1016/j.knosys.2020.105941