Neural Networks & Deep Learning

GELU

Gaussian Error Linear Unit - a smooth activation function combining properties of dropout and ReLU, used in BERT and GPT.

This concept is essential for understanding neural networks & deep learning and forms a key part of modern AI systems.

  • Activation Function
  • ReLU
  • Swish

Tags

neural-networks-deep-learning activation-function relu swish

Related Terms

Added: November 18, 2025