Learning Rates for Stochastic Gradient Descent With Nonconvex Objectives
Stochastic gradient descent (SGD) has become the method of choice for training highly complex and nonconvex models since it can not only recover good solutions to minimize training errors but also generalize well. Computational and statistical properties are separately studied to understand the beha...
Veröffentlicht in: | IEEE transactions on pattern analysis and machine intelligence. - 1979. - 43(2021), 12 vom: 22. Dez., Seite 4505-4511 |
---|---|
1. Verfasser: | |
Weitere Verfasser: | |
Format: | Online-Aufsatz |
Sprache: | English |
Veröffentlicht: |
2021
|
Zugriff auf das übergeordnete Werk: | IEEE transactions on pattern analysis and machine intelligence |
Schlagworte: | Journal Article |
Online verfügbar |
Volltext |