Optimization functions adjust the Model Parameters to minimize the Loss function, which measures how well the model performs. This is a fundamental step in training machine learning models.
General Optimization Process:
The Optimisation function (e.g., LBFGS, Newton-CG) iteratively updates the Model Parameters by:
- Calculating the gradient of the loss function with respect to the parameters.
- Updating the parameters in the direction of the negative gradient (as described in Gradient Descent).
This process is repeated until:
- The cost function converges (i.e., the change in the loss function becomes negligible), or
- The maximum number of iterations is reached.