Optimisation techniques
- Iteratively updates parameters using the gradient of the Cost Function with respect to the parameters.
- Requires careful tuning of the Learning Rate (), which controls the size of each update.
Optimization Solvers in Sklearn : Scikit-learn solvers improve on Gradient Descent by leveraging advanced techniques:
- Use second-order information, such as approximations to the Hessian matrix.
- Achieve faster and more reliable convergence compared to Gradient Descent.
- Automatically adapt step sizes Adaptive Learning Rates, eliminating the need for manual tuning.