Stochastic gradient descent (SGD): Revision history

Diff selection: Mark the radio buttons of the revisions to compare and hit enter or the button at the bottom.
Legend: (cur) = difference with latest revision, (prev) = difference with preceding revision, m = minor edit.

18 March 2023

  • curprev 13:2913:29, 18 March 2023Walle talk contribs 3,672 bytes +3,672 Created page with "{{see also|Machine learning terms}} ==Introduction== '''Stochastic gradient descent''' ('''SGD''') is an optimization algorithm commonly used in machine learning and deep learning to minimize a given objective function. It is a variant of the gradient descent algorithm that performs updates on a randomly selected subset of the data, rather than the entire dataset, at each iteration. This approach offers several advantages, including faster convergence and the abi..."