Abstract
Back-Propagation (BP)[Rumelhart et al, 1986] is a popular algorithm employed for training multilayer connectionist learning systems with nonlinear activation function (sigmoid). However, BP is plagued by excruciatingly slow convergence for many applications, and this drawback has been partly attributed to the Flat-Spots problem. Literature defines flat-spots as regions where the derivative of the sigmoid activation function approaches zero, and in these regions the weight changes become negligible, despite the presence of considerable classification error. Thus learning slows down dramatically. Several researchers have addressed this problem posed by flat-spots [Fahlman, 1988, Balakrishnan & Honavar, 1992]. In this paper we present a new way of dealing with the flat-spots in the output layer. This new method uses a Perceptron-like weight-modification strategy to complement BP in the output layer. We also report an empirical evaluation of the comparative performances of these techniques on some data-sets that have been used extensively in bench-marking inductive learning algorithms.
Original language | English (US) |
---|---|
Title of host publication | Proceedings of SPIE - The International Society for Optical Engineering |
Editors | Marry Lou Padgett |
Publisher | Publ by Society of Photo-Optical Instrumentation Engineers |
Pages | 55-60 |
Number of pages | 6 |
Volume | 1721 |
ISBN (Print) | 1565550072 |
State | Published - 1993 |
Event | Proceedings of the 3rd Workshop on Neural Networks: Academic/Industrial/NASA/Defense - Alabama, AL, USA Duration: Feb 10 1992 → Feb 12 1992 |
Other
Other | Proceedings of the 3rd Workshop on Neural Networks: Academic/Industrial/NASA/Defense |
---|---|
City | Alabama, AL, USA |
Period | 2/10/92 → 2/12/92 |
All Science Journal Classification (ASJC) codes
- Electrical and Electronic Engineering
- Condensed Matter Physics