Training & Optimization

Gradient Checkpointing

Trading computation for memory by recomputing activations during backprop instead of storing them.

This concept is essential for understanding training & optimization and forms a key part of modern AI systems.

  • Memory Efficiency
  • Training
  • Backpropagation

Tags

training-optimization memory-efficiency training backpropagation

Related Terms

Added: November 18, 2025