Generated by GPT-5-mini| Ampere (GPU microarchitecture) | |
|---|---|
| Name | Ampere |
| Caption | NVIDIA Ampere GPU architecture |
| Designed by | NVIDIA |
| Manufacturer | TSMC |
| Process size | 7 nm |
| First release | 2020 |
| Cores | CUDA, RT, Tensor |
| Predecessor | Turing (microarchitecture) |
| Successor | Ada Lovelace (microarchitecture) |
Ampere (GPU microarchitecture) Ampere is a GPU microarchitecture developed by NVIDIA and launched in 2020 for data center and consumer markets. It succeeds Turing (microarchitecture) and emphasizes mixed‑precision compute, ray tracing, and high throughput for workloads from deep learning to real‑time rendering. Ampere underpins product families such as NVIDIA Tesla, NVIDIA A100, and GeForce RTX 30 series, and it influenced later designs used in supercomputing and cloud computing deployments.
Ampere was unveiled amid collaborations with organizations including Microsoft, Google, and research institutions like University of California, Berkeley to address scaling for AI research and HPC workloads. It integrates architectural innovations for matrix multiply, sparsity, and hardware‑accelerated ray tracing to serve use cases in autonomous vehicles, medical imaging, and weather forecasting. Ampere was fabricated on a 7 nm process by TSMC and launched during industry events involving GTC presentations and government procurement discussions.
Ampere’s architecture reorganizes streaming multiprocessors (SMs) with second‑generation RT cores and third‑generation Tensor cores to accelerate matrix operations used in models such as BERT, ResNet, and Transformer (machine learning model). SMs include improved integer pipelines and larger register files inspired by research from Stanford University and MIT. The design employs a multi‑chip module approach in some variants, echoing packaging developments from AMD and collaborations with Samsung Electronics in process roadmaps. Memory subsystems use HBM2e and enhanced GDDR6X interfaces implemented alongside coherence mechanisms influenced by proposals from OpenCAPI and CCIX consortia.
Ampere delivers substantial throughput gains for single‑precision and mixed‑precision tasks, leveraging TensorRT optimizations and support for FP64, FP32, TF32, and INT8 formats. Hardware‑accelerated ray tracing is implemented with dedicated RT cores to reduce traversal and shading costs in engines like Unreal Engine and Unity (game engine). The architecture introduces support for structural sparsity to double effective matrix throughput for compliant models, a technique discussed in papers from DeepMind and OpenAI. Power and thermal management draw on telemetry frameworks used by NVIDIA DGX systems and data center cooling practices adopted by Amazon Web Services and Microsoft Azure.
Ampere appears across data center and consumer SKUs including the NVIDIA A100, NVIDIA RTX 3080, and NVIDIA RTX 3090. Enterprise deployments include systems such as NVIDIA DGX A100, HPE Apollo, and offerings from Dell EMC for HPC clusters. Cloud providers like Google Cloud Platform and Amazon EC2 integrated Ampere‑based instances for training large language models and inference services. Gaming and creator markets received Ampere through the GeForce RTX 30 series, distributed by partners including ASUS, MSI, and Gigabyte Technology.
Software ecosystems around Ampere include CUDA, cuDNN, TensorRT, and frameworks such as TensorFlow, PyTorch, and MXNet. Compiler and driver updates were coordinated with development tools like NVIDIA Nsight and IDEs supported by JetBrains and Visual Studio. Ampere’s feature set is exposed to developers via SDKs for RTX, OptiX, and libraries used in scientific computing, machine learning research from OpenAI and industry partners, and render pipelines in Blender and Autodesk products.
Reception among reviewers and institutions such as Linus Tech Tips labs, AnandTech, and academic benchmarking groups highlighted large gains in throughput for AI training and ray tracing performance relative to Turing (microarchitecture). Independent evaluations from supercomputing centers and publications like IEEE Spectrum examined energy efficiency and performance per watt in systems competing with AMD Instinct accelerators. Critiques focused on supply constraints and pricing noted in market analyses by Gartner and IDC.
Ampere’s lineage continued into successors such as Ada Lovelace (microarchitecture), influencing design decisions in tensor compute, real‑time ray tracing, and packaging. Its deployment in data centers, cloud platforms, and research labs contributed to accelerated development of large‑scale models by organizations like OpenAI and DeepMind, and it shaped procurement and architecture roadmaps at companies including Facebook (Meta Platforms), Apple Inc., and IBM.
Category:NVIDIA microarchitectures