Abstract
In this paper, we propose a new type of recurrent neural-network architecture, in which each output unit is connected to itself and is also fully connected to other output units and all hidden units. The proposed recurrent neural network differs from Jordan's and Elman's recurrent neural networks with respect to function and architecture, because it has been originally extended from being a mere multilayer feedforward neural network, to improve discrimination and generalization powers. We also prove the convergence properties of learning algorithm in the proposed recurrent neural network, and analyze the performance of the proposed recurrent neural network by performing recognition experiments with the totally unconstrained handwritten numeric database of Concordia University, Montreal, Canada. Experimental results have confirmed that the proposed recurrent neural network improves discrimination and generalization powers in the recognition of visual patterns.
Original language | English |
---|---|
Pages (from-to) | 331-340 |
Number of pages | 10 |
Journal | IEEE Transactions on Neural Networks |
Volume | 8 |
Issue number | 2 |
DOIs | |
Publication status | Published - 1997 |
Keywords
- Convergence properties
- Recurrent neural network
- Visual pattern recognition
ASJC Scopus subject areas
- Software
- Computer Science Applications
- Computer Networks and Communications
- Artificial Intelligence