Large Language Models
Padding Token
A special token used to make sequences the same length in a batch, typically ignored during computation.
This concept is essential for understanding large language models and forms a key part of modern AI systems.
Related Concepts
- Special Token
- Batching
- Attention Mask
Tags
large-language-models special-token batching attention-mask