Training & Optimization
Gradient Checkpointing
Trading computation for memory by recomputing activations during backprop instead of storing them.
This concept is essential for understanding training & optimization and forms a key part of modern AI systems.
Related Concepts
- Memory Efficiency
- Training
- Backpropagation
Tags
training-optimization memory-efficiency training backpropagation