Limitations of Gradient Descent
Explore the core limitations of gradient descent when applied to non-convex optimization problems in machine learning. Learn why gradient descent can be slow, get stuck in local optima, and how learning rate affects convergence. This lesson helps you understand practical challenges to improve optimization strategies.
We have seen how well gradient descent works in the case of convex optimization because of the presence of a single global optimal solution. We will now look at some of the limitations of gradient descent and address them in this chapter.
Intractability
Consider a machine learning problem where we want to minimize the discrepancy between the model prediction
Here,
To compute the gradient