Simple recurrent networks learn context-free and context-sensitive languages by counting

Authors
Citation
P. Rodriguez, Simple recurrent networks learn context-free and context-sensitive languages by counting, NEURAL COMP, 13(9), 2001, pp. 2093-2118
Citations number
39
Categorie Soggetti
Neurosciences & Behavoir","AI Robotics and Automatic Control
Journal title
NEURAL COMPUTATION
ISSN journal
08997667 → ACNP
Volume
13
Issue
9
Year of publication
2001
Pages
2093 - 2118
Database
ISI
SICI code
0899-7667(200109)13:9<2093:SRNLCA>2.0.ZU;2-B
Abstract
It has been shown that if a recurrent neural network (RNN) learns to proces s a regular language, one can extract a finite-state machine (FSM) by treat ing regions of phase-space as FSM states. However, it has also been shown t hat one can construct an RNN to implement Turing machines by using RNN dyna mics as counters. But how does a network learn languages that require count ing? Rodriguez, Wiles, and Elman (1999) showed that a simple recurrent netw ork (SRN) can learn to process a simple context-free language (CFL) by coun ting up and down. This article extends that to show a range of language tas ks in which an SRN develops solutions that not only count but also copy and store counting information. In one case, the network stores information li ke an explicit storage mechanism. In other cases, the network stores inform ation more indirectly in trajectories that are sensitive to slight displace ments that depend on context. In this sense, an SRN can learn analog comput ation as a set of interdependent counters. This demonstrates how SRNs may b e an alternative psychological model of language or sequence processing.