Mini-Batch Gradient Descent is an optimization algorithm that serves as a compromise between Batch Gradient Descent and Stochastic Gradient Descent (SGD). Instead of using the entire dataset or a single data point, it updates model parameters using small, random subsets (mini-batches) of the training data.

Pros: