Training & Optimization

Gradient Accumulation

Summing gradients over multiple batches before updating, simulating larger effective batch sizes.

This concept is essential for understanding training & optimization and forms a key part of modern AI systems.

  • Training
  • Batch Size
  • Memory Efficiency

Tags

training-optimization training batch-size memory-efficiency

Related Terms

Added: November 18, 2025