NVIDIA A100 80GB PCIe Graphics Card - Shop Now and Save!

By: NVIDIA

Model: 900-21001-0320-030

Sku: DGBOS8TT1C

Availability: Out of stock New

Availability: Out of stock

  • FP64: 9.7 TFLOPS
  • FP64 Tensor Core: 19.5 TFLOPS
  • FP32: 19.5 TFLOPS
  • Tensor Float 32 (TF32): 156 TFLOPS | 312 TFLOPS
  • BFLOAT16 Tensor Core: 312 TFLOPS | 624 TFLOPS
  • FP16 Tensor Core: 312 TFLOPS | 624 TFLOPS
  • INT8 Tensor Core: 624 TOPS | 1248 TOPS
  • GPU Memory: 80GB HBM2e
  • GPU Memory Bandwidth: 1,935 GB/s
  • Max Thermal Design Power (TDP): 300W
  • Multi-Instance GPU: Up to 7 MIGs @ 10GB
  • Form Factor: PCIe, Dual-slot air-cooled or single-slot liquid-cooled
  • Interconnect: NVIDIA NVLink Bridge, for 2 GPUs: 600 GB/s, PCIe Gen4: 64 GB/s
Stock Request Notify When Available
  • Add to Wishlist
  • Add to Compare

Accelerating the Most Important Work of Our Time

NVIDIA A100 Tensor Core GPU delivers unprecedented acceleration at every scale to power the world’s highest-performing elastic data centers for AI, data analytics, and HPC. Powered by the NVIDIA Ampere Architecture, A100 is the engine of the NVIDIA data center platform. A100 provides up to 20X higher performance over the prior generation and can be partitioned into seven GPU instances to dynamically adjust to shifting demands. The A100 80GB debuts the world’s fastest memory bandwidth at over 2 terabytes per second (TB/s) to run the largest models and datasets.

Enterprise-Ready Software for AI

The NVIDIA EGX platform includes optimized software that delivers accelerated computing across the infrastructure. With NVIDIA AI Enterprise, businesses can access an end-to-end, cloud-native suite of AI and data analytics software that’s optimized, certified, and supported by NVIDIA to run on VMware vSphere with NVIDIA-Certified Systems. NVIDIA AI Enterprise includes key enabling technologies from NVIDIA for rapid deployment, management, and scaling of AI workloads in the modern hybrid cloud.

The Most Powerful End-to-End AI and HPC Data Center Platform

A100 is part of the complete NVIDIA data center solution that incorporates building blocks across hardware, networking, software, libraries, and optimized AI models and applications from NGC. Representing the most powerful end-to-end AI and HPC platform for data centers, it allows researchers to rapidly deliver real-world results and deploy solutions into production at scale.

Up to 3X Higher AI Training on Largest Models

AI models are exploding in complexity as they take on next-level challenges such as conversational AI. Training them requires massive compute power and scalability.

NVIDIA A100 Tensor Cores with Tensor Float (TF32) provide up to 20X higher performance over the NVIDIA Volta with zero code changes and an additional 2X boost with automatic mixed precision and FP16. When combined with NVIDIA NVLink, NVIDIA NVSwitch, PCI Gen4, NVIDIA InfiniBand, and the NVIDIA Magnum IO SDK, it’s possible to scale to thousands of A100 GPUs.

A training workload like BERT can be solved at scale in under a minute by 2,048 A100 GPUs, a world record for time to solution.

Deep Learning Inference

A100 introduces groundbreaking features to optimize inference workloads. It accelerates a full range of precision, from FP32 to INT4. Multi-Instance GPU (MIG) technology lets multiple networks operate simultaneously on a single A100 for optimal utilization of compute resources. And structural sparsity support delivers up to 2X more performance on top of A100’s other inference performance gains.

On state-of-the-art conversational AI models like BERT, A100 accelerates inference throughput up to 249X over CPUs.

High-Performance Computing

To unlock next-generation discoveries, scientists look to simulations to better understand the world around us.

2X Faster than A100 40GB on Big Data Analytics Benchmark

Data scientists need to be able to analyze, visualize, and turn massive datasets into insights. But scale-out solutions are often bogged down by datasets scattered across multiple servers.

Accelerated servers with A100 provide the needed compute power—along with massive memory, over 2 TB/sec of memory bandwidth, and scalability with NVIDIA NVLink and NVSwitch, —to tackle these workloads. Combined with InfiniBand, NVIDIA Magnum IO and the RAPIDS suite of open-source libraries, including the RAPIDS Accelerator for Apache Spark for GPU-accelerated data analytics, the NVIDIA data center platform accelerates these huge workloads at unprecedented levels of performance and efficiency.

In the Box

  • NVIDIA A100 80GB PCIe Graphics Card