Back to Search
Start Over
Sequential RAM-based Neural Networks: Learnability, Generalisation, Knowledge Extraction, and Grammatical Inference.
- Source :
-
International Journal of Neural Systems . Jun1999, Vol. 9 Issue 3, p203. 8p. - Publication Year :
- 1999
-
Abstract
- A fundamental question in the field of artificial neural networks is what set of problems a given class of networks can perform (computability). Such a problem can be made less general, but no less important, by asking what these networks could learn by using a given training procedure (learnability). The basic purpose of this paper is to address the learnability problem. Specifically, it analyses the learnability of sequential RAM-based neural networks. The analytical tools used are those of Automata Theory. In this context, this paper establishes which class of problems and under what conditions such networks, together with their existing learning rules, can learn and generalise. This analysis also yields techniques for both extracting knowledge from and inserting knowledge into the networks. The results presented here, besides helping in a better understanding of the temporal behaviour of sequential RAM-based networks, could also provide useful insights for the integration of the symbolic/connectionist paradigms. [ABSTRACT FROM AUTHOR]
- Subjects :
- *ARTIFICIAL neural networks
*RANDOM access memory
*ARTIFICIAL intelligence
Subjects
Details
- Language :
- English
- ISSN :
- 01290657
- Volume :
- 9
- Issue :
- 3
- Database :
- Academic Search Index
- Journal :
- International Journal of Neural Systems
- Publication Type :
- Academic Journal
- Accession number :
- 10236394
- Full Text :
- https://doi.org/10.1142/S0129065799000198