Why should we adjust the parameter alpha when using gradient descent?
Why should we adjust the parameter alpha when using gradient descent? |
- To ensure that the gradient descent algorithm converges in a reasonable time.
- Failure to converge or too much time to obtain the minimum value implies that our step size is wrong.