|
Stephan Rave

Adrian Riekert (Uni Münster): Adaptive stochastic gradient methods and training of artificial neural networks

Wednesday, 30.10.2024 14:15 im Raum M5

Mathematik und Informatik

We consider stochastic gradient descent (SGD) methods with adaptive learning learning rates that are changed dynamically during the training process. For a particular class of coercive target functions we show convergence of such SGD methods to global minima. Afterwards, we specialize to the situation of optimizing artificial neural networks (ANNs) and discuss convergence results for the gradient flow, the continuous-time analogue of gradient descent.



Angelegt am 18.09.2024 von Stephan Rave
Geändert am 28.10.2024 von Stephan Rave
[Edit | Vorlage]

Oberseminar Numerik