Balanced Gradient Training of Feed Forward Networks
作者:Son Nguyen, Michael T. Manry
摘要
We show that there are infinitely many valid scaled gradients which can be used to train a neural network. A novel training method is proposed that finds the best scaled gradients in each training iteration. The method’s implementation uses first order derivatives which makes it scalable and suitable for deep learning and big data. In simulations, the proposed method has similar or less testing error than conjugate gradient and Levenberg Marquardt. The method reaches the final network utilizing fewer multiplies than the other two algorithms. It also works better than conjugate gradient in convolutional neural networks.
论文关键词:Back propagation, Vanishing gradient, Balanced gradient
论文评审过程:
论文官网地址:https://doi.org/10.1007/s11063-021-10474-1