Share Email Print
cover

Proceedings Paper

Feature competition and domain of attraction in artificial-perceptron pattern recognizer
Author(s): Chia-Lun John Hu
Format Member Price Non-Member Price
PDF $14.40 $18.00
cover GOOD NEWS! Your organization subscribes to the SPIE Digital Library. You may be able to download this paper for free. Check Access

Paper Abstract

As we reported previously, learning of a multi-layered hard-limited perceptron can be formulated into a set of simultaneous linear inequalities. Solving these inequalities under a given training set would then allow us to achieve the goal of learning in this system. If the dimension N of the input vector is much larger than the number M of different patterns to be learned, then there is considerable freedom for the system to select a proper solution of the connection matrix. In most cases, even a single layer perceptron will do the learning satisfactorily. This paper reports the results of some theoretical and experimental studies of this one-layered, hard-limited perceptron trained under the novel, one-step, noniterative learning scheme. Particularly, the analysis of some important properties of this novel learning system, such as automatic feature competition, domain of convergence, and robustness of recognition, are discussed in detail.

Paper Details

Date Published: 29 October 1993
PDF: 4 pages
Proc. SPIE 2032, Neural and Stochastic Methods in Image and Signal Processing II, (29 October 1993); doi: 10.1117/12.162025
Show Author Affiliations
Chia-Lun John Hu, Southern Illinois Univ./Carbondale (United States)


Published in SPIE Proceedings Vol. 2032:
Neural and Stochastic Methods in Image and Signal Processing II
Su-Shing Chen, Editor(s)

© SPIE. Terms of Use
Back to Top