Abstract
We study the simplest possible order one single-layer perceptron with two inputs, using the delta rule with online learning, in order to derive closed form expressions for the mean convergence rates. We investigate the rate of convergence in weight space of the weight vectors corresponding to each of the 14 out of 16 linearly separable rules. These vectors follow zigzagging lines through the piecewise constant vector field to their respective attractors. Based on our studies, we conclude that a single-layer perceptron with inputs will converge in an average number of steps given by an order polynomial in , where is the threshold, and is the size of the initial weight distribution. Exact values for these averages are provided for the five linearly separable classes with . We also demonstrate that the learning rate is determined by the attractor size, and that the attractors of a single-layer perceptron with inputs partition .
- Received 12 August 2006
DOI:https://doi.org/10.1103/PhysRevE.75.026704
©2007 American Physical Society