Why should we adjust the parameter alpha when using gradient descent?

Why should we adjust the parameter alpha when using gradient descent?




Why should we adjust the parameter alpha when using gradient descent?
Why should we adjust the parameter alpha when using gradient descent?


  • To ensure that the gradient descent algorithm converges in a reasonable time.
  • Failure to converge or too much time to obtain the minimum value implies that our step size is wrong.


Learn More :