Skip to main content
. 2018 Jun 22;9(4):611–629. doi: 10.1007/s13244-018-0639-9

Fig. 7.

Fig. 7

Gradient descent is an optimization algorithm that iteratively updates the learnable parameters so as to minimize the loss, which measures the distance between an output prediction and a ground truth label. The gradient of the loss function provides the direction in which the function has the steepest rate of increase, and all parameters are updated in the negative direction of the gradient with a step size determined based on a learning rate