What is the role of the "learning rate" in gradient descent optimization?
The size of the steps taken during each iteration
The regularization strength applied to the mod
The number of iterations in the optimization process
The speed at which the algorithm converges
This question is part of this quiz :
Data Science Quiz