Abstract

In recent years, gradient-based LSTM recurrent neural networks (RNNs) solved many previously RNN-unlearnable tasks. Sometimes, however, gradient information is of little use for training RNNs, due to numerous local minima. For such cases, we present a novel method: EVOlution of systems with LINear Outputs (Evolino). Evolino evolves weights to the nonlinear, hidden nodes of RNNs while computing optimal linear mappings from hidden state to output, using methods such as pseudo-inverse-based linear regression. If we instead use quadratic programming to maximize the margin, we obtain the first evolutionary recurrent support vector machines. We show that Evolino-based LSTM can solve tasks that Echo State nets (Jaeger, 2004a) cannot and achieves higher accuracy in certain continuous function generation tasks than conventional gradient descent RNNs, including gradient-based LSTM.

Keywords

Recurrent neural networkGradient descentMargin (machine learning)Echo state networkComputer scienceMaxima and minimaArtificial intelligenceQuadratic programmingArtificial neural networkAlgorithmMachine learningMathematicsMathematical optimization

MeSH Terms

AnimalsArtificial IntelligenceHumansMemoryModelsNeurologicalNeural NetworksComputerNeuronsNonlinear DynamicsSerial LearningSignal ProcessingComputer-Assisted

Affiliated Institutions

Related Publications

Publication Info

Year
2007
Type
article
Volume
19
Issue
3
Pages
757-779
Citations
251
Access
Closed

Social Impact

Altmetric

Social media, news, blog, policy document mentions

Citation Metrics

251
OpenAlex
22
Influential
162
CrossRef

Cite This

Jürgen Schmidhuber, Daan Wierstra, Matteo Gagliolo et al. (2007). Training Recurrent Networks by Evolino. Neural Computation , 19 (3) , 757-779. https://doi.org/10.1162/neco.2007.19.3.757

Identifiers

DOI
10.1162/neco.2007.19.3.757
PMID
17298232

Data Quality

Data completeness: 90%