Title
A Recurrent Neural Network That Learns To Count
Abstract
Parallel distributed processing (PDP) architectures demonstrate a potentially radical alternative to the traditional theories of language processing that are based on serial computational models. However, learning complex structural relationships in temporal data presents a serious challenge to PDP systems. For example, automata theory dictates that processing strings from a context-free language (CFL) requires a stack or counter memory device. While some PDP models have been hand-crafted to emulate such a device, it is not clear how a neural network might develop such a device when learning a CFL. This research employs standard backpropagation training techniques for a recurrent neural network (RNN) in the task of learning to predict the next character in a simple deterministic CFL (DCFL). We show that an RNN can learn to recognize the structure of a simple DCFL. We use dynamical systems theory to identify how network states reflect that structure by building counters in phase space. The work is an empirical investigation which is complementary to theoretical analyses of network capabilities, yet original in its specific configuration of dynamics involved. The application of dynamical systems theory helps us relate the simulation results to theoretical results, and the learning task enables us to highlight some issues for understanding dynamical systems that process language with counters.
Year
DOI
Venue
1999
10.1080/095400999116340
CONNECTION SCIENCE
Keywords
Field
DocType
recurrent neural network, dynamical systems, context-free languages
Context-free language,Automata theory,Computer science,Recurrent neural network,Dynamical systems theory,Computational model,Artificial intelligence,Backpropagation,Artificial neural network,Connectionism,Machine learning
Journal
Volume
Issue
ISSN
11
1
0954-0091
Citations 
PageRank 
References 
67
3.99
9
Authors
3
Name
Order
Citations
PageRank
Paul Rodriguez1673.99
Janet Wiles2726.12
Jeffrey Elman3673.99