Classification methodologies of multilayer perceptrons with sigmoid activation functions

作者:

Highlights:

摘要

This paper studies the classification mechanisms of multilayer perceptrons (MLPs) with sigmoid activation functions (SAFs). The viewpoint is presented that in the input space the hyperplanes determined by the hidden basis functions with values 0's do not play the role of decision boundaries, and such hyperplanes do not certainly go through the marginal regions between different classes. For solving an n-class problem, a single-hidden-layer perceptron with at least log2(n-1)⩾2 hidden nodes is needed. The final number of hidden neurons is still related to the sample distribution shapes and regions, but not to the number of samples and input dimensions. As a result, an empirical formula for optimally selecting the initial number of hidden nodes is proposed. The ranks of response matrixes of hidden layers should be taken as a main basis for pruning or growing the existing hidden neurons. A structure-fixed perceptron ought to learn more than one round from different starting weight points for one classification task, and only the group of weights and biases that has the best generalization performance should be reserved. Finally, three examples are given to verify the above viewpoints.

论文关键词:Multilayer perceptrons,Sigmoid activation functions,Hidden nodes,Empirical formula,Generalization

论文评审过程:Received 28 July 2003, Revised 22 March 2005, Accepted 22 March 2005, Available online 20 June 2005.

论文官网地址:https://doi.org/10.1016/j.patcog.2005.03.024