A plain-English explanation of Attention (Attention Mechanism) — what it means, why it matters, and how it is used in AI.
Also known as: Self-attention, multi-head attention, scaled dot-product attention
Attention is the core mechanism behind all Transformer-based models including GPT, BERT, and Claude.
The best way to remember Attention is to practice unscrambling it. AI Terminology Scrambler uses spaced repetition to help you learn and retain AI vocabulary in just a few minutes a day.
Practice Attention now →