Generated by GPT-5-mini| Attention (machine learning) | |
|---|---|
![]() Zhang, Aston and Lipton, Zachary C. and Li, Mu and Smola, Alexander J. · CC BY-SA 4.0 · source | |
| Name | Attention (machine learning) |
| Caption | Diagram of attention mechanism in neural networks |
| Introduced | 2014 |
| Developers | Google, Facebook AI Research, OpenAI |
| Field | Artificial intelligence, Machine learning, Natural language processing |
| Notable models | Transformer (machine learning model), BERT, GPT-3, T5 (language model), Vision Transformer |
Attention (machine learning) Attention in machine learning is a mechanism that enables models to dynamically weight inputs when producing outputs, improving performance on sequence and structured prediction tasks. Originating from work on neural machine translation, attention has been integrated into architectures used by organizations such as Google, OpenAI, Facebook AI Research, and DeepMind. Attention underpins influential models like Transformer (machine learning model), BERT, and GPT-3 and has driven advances across Natural language processing, Computer vision, and Speech recognition.
Attention mechanisms were popularized by papers from teams at Google Brain and University of Montreal that built on prior work in Cognitive science and computational models from researchers at University of Toronto. Early demonstrations in Neural machine translation showed benefits versus Sequence-to-sequence models such as those developed at Kyoto University and Microsoft Research. Attention computes a context-aware weighted sum of values using compatibility scores from queries and keys, a construction formalized in the Transformer (machine learning model) paper from Google Research. Subsequent variants were adopted by models from OpenAI, DeepMind, Stanford University, and Carnegie Mellon University for tasks ranging from Question answering benchmarks at ACL (conference) to image recognition evaluations at ImageNet competitions.
Scaled dot-product attention, additive attention, and multiplicative attention trace roots to work by researchers at University of Toronto and Microsoft Research; each computes alignment differently using matrices popularized in implementations at TensorFlow and PyTorch. Multi-head attention, introduced in Google Research's Transformer, runs multiple attention functions in parallel and aggregates outputs, a technique employed by BERT, RoBERTa, and T5 (language model). Variants include sparse attention explored by teams at Facebook AI Research and DeepMind, local attention used in models from Apple and NVIDIA, and Relative Position Representations advanced by researchers at Google DeepMind and University of Oxford. Cross-attention links encoder and decoder states in architectures from Amazon Web Services and Alibaba, while self-attention is central to models by OpenAI and Salesforce Research. Hybrid designs combine attention with convolutional modules pioneered at ETH Zurich and Massachusetts Institute of Technology labs.
Attention is a core component of transformers used in models like BERT, GPT-3, T5 (language model), and Vision Transformer. In Natural language processing, attention enables state-of-the-art systems for Machine translation at WMT shared tasks, Summarization evaluated at EMNLP, and Language modeling benchmarks such as GLUE and SuperGLUE. In Computer vision, attention powers models submitted to ImageNet and COCO challenges by teams at Google Research, Facebook AI Research, and Microsoft Research; applications include object detection in Open Images datasets and scene understanding in robotics labs at Toyota Research Institute and Boston Dynamics. Speech and audio projects at DeepMind and Apple use attention for end-to-end recognition and synthesis; healthcare applications at Mayo Clinic and Johns Hopkins University leverage attention for clinical note parsing. Attention also appears in reinforcement learning systems from DeepMind applied to Atari benchmarks and AlphaGo-related research at Google DeepMind.
Training attention-based models leverages optimization algorithms such as Adam introduced by researchers at King's College London and Google, learning-rate schedules from OpenAI and warmup strategies popularized in Google Research implementations, and regularization tactics including dropout from Geoffrey Hinton's group at University of Toronto. Large-scale pretraining regimes used by OpenAI, Google, and Facebook AI Research employ massive web-scale datasets curation methods inspired by work at Common Crawl and dataset filtering approaches from Stanford University. Efficient training practices such as gradient checkpointing by teams at Facebook AI Research, mixed-precision arithmetic promoted by NVIDIA, and distributed training frameworks from Microsoft Research and Amazon Web Services enable scaling to models like GPT-3 and PaLM (language model). Transfer learning and fine-tuning workflows connect pretrained attention models to downstream tasks in pipelines engineered by groups at Hugging Face and Allen Institute for AI.
Performance of attention-based systems is measured using benchmark suites and metrics established by communities at ACL (conference), NeurIPS, and ICML. Common evaluation datasets include GLUE, SuperGLUE, SQuAD, ImageNet, and COCO; metrics such as BLEU from IBM Research for translation, ROUGE for summarization, and accuracy/F1 for classification are widely reported. Model analysis techniques by researchers at Google Research and OpenAI examine attention heads via probing methods from University of Cambridge and attribution tools from MIT-IBM Watson AI Lab. Efficiency metrics like FLOPs popularized by Stanford University groups and memory footprint studies from Facebook AI Research guide practical deployment trade-offs.
Attention models face limitations studied by teams at OpenAI, DeepMind, and University of Oxford: quadratic scaling of self-attention with sequence length affects long-context tasks tackled by sparse and linear attention proposals from Facebook AI Research and Google DeepMind. Concerns about data bias and model behavior highlighted by researchers at AI Now Institute and Partnership on AI affect deployment in systems built by Amazon and Microsoft. Robustness and adversarial vulnerabilities explored by University of California, Berkeley and Carnegie Mellon University raise safety questions for applications in healthcare settings at Mayo Clinic and autonomous systems by Tesla. Interpretability criticisms spur methods from Stanford University and MIT for explaining attention patterns; computational cost and environmental impact debated in studies from University of Massachusetts Amherst and Princeton University motivate efficient architecture research at Intel and NVIDIA.