First- and Second-Order Methods for Learning: Between Steepest Descent and Newton's Method

1992 Neural Computation 1,187 citations

Abstract

On-line first-order backpropagation is sufficiently fast and effective for many large-scale classification problems but for very high precision mappings, batch processing may be the method of choice. This paper reviews first- and second-order optimization methods for learning in feedforward neural networks. The viewpoint is that of optimization: many methods can be cast in the language of optimization techniques, allowing the transfer to neural nets of detailed results about computational complexity and safety procedures to ensure convergence and to avoid numerical problems. The review is not intended to deliver detailed prescriptions for the most appropriate methods in specific applications, but to illustrate the main characteristics of the different methods and their mutual relations.

Keywords

Computer scienceBackpropagationArtificial neural networkConvergence (economics)Gradient descentArtificial intelligenceFeedforward neural networkMathematical optimizationFeed forwardAlgorithmMathematics

Affiliated Institutions

Related Publications

Optimization for training neural nets

Various techniques of optimizing criterion functions to train neural-net classifiers are investigated. These techniques include three standard deterministic techniques (variable...

1992 IEEE Transactions on Neural Networks 210 citations

Publication Info

Year
1992
Type
article
Volume
4
Issue
2
Pages
141-166
Citations
1187
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

1187
OpenAlex

Cite This

Roberto Battiti (1992). First- and Second-Order Methods for Learning: Between Steepest Descent and Newton's Method. Neural Computation , 4 (2) , 141-166. https://doi.org/10.1162/neco.1992.4.2.141

Identifiers

DOI
10.1162/neco.1992.4.2.141