Balanced Gradient Training of Feed Forward Networks

作者:Son Nguyen, Michael T. Manry

摘要

We show that there are infinitely many valid scaled gradients which can be used to train a neural network. A novel training method is proposed that finds the best scaled gradients in each training iteration. The method’s implementation uses first order derivatives which makes it scalable and suitable for deep learning and big data. In simulations, the proposed method has similar or less testing error than conjugate gradient and Levenberg Marquardt. The method reaches the final network utilizing fewer multiplies than the other two algorithms. It also works better than conjugate gradient in convolutional neural networks.

论文关键词:Back propagation, Vanishing gradient, Balanced gradient

论文评审过程:

论文官网地址:https://doi.org/10.1007/s11063-021-10474-1