Large Language Models

Bidirectional Attention

Allowing tokens to attend to both past and future context, used in encoder models like BERT.

This concept is essential for understanding large language models and forms a key part of modern AI systems.

  • Attention
  • BERT
  • Encoder

Tags

large-language-models attention bert encoder

Related Terms

Added: November 18, 2025