Large Language Models

Masked Language Modeling

A pre-training objective where random tokens are masked and the model learns to predict them from context.

  • BERT: Explore how BERT relates to Masked Language Modeling
  • Pre-training: Explore how Pre-training relates to Masked Language Modeling
  • Masking: Explore how Masking relates to Masked Language Modeling

Why It Matters

Understanding Masked Language Modeling is crucial for anyone working with large language models. This concept helps build a foundation for more advanced topics in AI and machine learning.

Learn More

This term is part of the comprehensive AI/ML glossary. Explore related terms to deepen your understanding of this interconnected field.

Tags

large-language-models bert pre-training masking

Related Terms

Added: November 18, 2025