Go back to the [[AI Glossary]]
A gradient descent algorithm in which the batch size is one. In other words, SGD relies on a single example chosen uniformly at random from a dataset to calculate an estimate of the gradient at each step.
Expanding this section will automatically generate an AI synthesis of the contributions in this node.