Large Language Models
Attention Mask
A binary mask indicating which tokens should be attended to, used to handle padding and causal masking.
This concept is essential for understanding large language models and forms a key part of modern AI systems.
Related Concepts
- Attention
- Padding
- Masking
Tags
large-language-models attention padding masking