Training & Optimization

Mini-Batch Gradient Descent

Computing gradients on small batches of data, balancing SGD’s noise with full-batch GD’s stability.

This concept is essential for understanding training & optimization and forms a key part of modern AI systems.

  • SGD
  • Gradient Descent
  • Batch Size

Tags

training-optimization sgd gradient-descent batch-size

Related Terms

Added: November 18, 2025