Abstract

The SGD-QN algorithm is a stochastic gradient descent algorithm that makes careful use of second-order information and splits the parameter update into independently scheduled components. Thanks to this design, SGD-QN iterates nearly as fast as a first-order stochastic gradient descent but requires less iterations to achieve the same accuracy. This algorithm won the Wild Track of the first PASCAL Large Scale Learning Challenge (Sonnenburg et al., 2008).

Keywords

Descent (aeronautics)Stochastic gradient descentApplied mathematicsGradient descentMathematicsStatistical physicsComputer sciencePhysicsMeteorologyArtificial intelligenceArtificial neural network

Related Publications

Squeeze-and-Excitation Networks

Convolutional neural networks are built upon the convolution operation, which extracts informative features by fusing spatial and channel-wise information together within local ...

2018 2018 IEEE/CVF Conference on Computer ... 25361 citations

Publication Info

Year
2009
Type
preprint
Citations
342
Access
Closed

External Links

Citation Metrics

342
OpenAlex

Cite This

Antoine Bordes, Léon Bottou, Patrick Gallinari (2009). SGD-QN: Careful Quasi-Newton Stochastic Gradient Descent. HAL (Le Centre pour la Communication Scientifique Directe) .