Large Language Models

Cross-Attention

Attention between two different sequences, where queries come from one and keys/values from another.

This concept is essential for understanding large language models and forms a key part of modern AI systems.

  • Attention Mechanism
  • Encoder-Decoder
  • Self-Attention

Tags

large-language-models attention-mechanism encoder-decoder self-attention

Related Terms

Added: November 18, 2025