Large Language Models

Attention Mask

A binary mask indicating which tokens should be attended to, used to handle padding and causal masking.

This concept is essential for understanding large language models and forms a key part of modern AI systems.

  • Attention
  • Padding
  • Masking

Tags

large-language-models attention padding masking

Related Terms

Added: November 18, 2025