Large Language Models

Padding Token

A special token used to make sequences the same length in a batch, typically ignored during computation.

This concept is essential for understanding large language models and forms a key part of modern AI systems.

  • Special Token
  • Batching
  • Attention Mask

Tags

large-language-models special-token batching attention-mask

Related Terms

Added: November 18, 2025