Large Language Models

Vocabulary Size

The number of distinct tokens a language model can process, typically 30K-100K+ tokens.

This concept is essential for understanding large language models and forms a key part of modern AI systems.

  • Tokenization
  • Token
  • Embedding

Tags

large-language-models tokenization token embedding

Related Terms

Added: November 18, 2025