Training & Optimization

Weight Decay

A regularization technique that shrinks weights toward zero during optimization. Equivalent to L2 regularization in standard SGD, but differs when using adaptive optimizers like Adam.

  • L2 Regularization: Explore how L2 Regularization relates to Weight Decay
  • Regularization: Explore how Regularization relates to Weight Decay
  • Optimization: Explore how Optimization relates to Weight Decay

Why It Matters

Understanding Weight Decay is crucial for anyone working with training & optimization. This concept helps build a foundation for more advanced topics in AI and machine learning.

Learn More

This term is part of the comprehensive AI/ML glossary. Explore related terms to deepen your understanding of this interconnected field.

Tags

training-optimization l2-regularization regularization optimization

Related Terms

Added: November 18, 2025