NVIDIA A2
Unprecedented Acceleration for World’s Highest-Performing Elastic Data Centers

The NVIDIA A2 Tensor Core GPU provides entry-level inference with low power, a small footprint, and high performance for intelligent video analytics (IVA) or NVIDIA AI at the edge. Featuring a low-profile PCIe Gen4 card and a low 40–60 watt (W) configurable thermal design power (TDP) capability, the A2 brings versatile inference acceleration to any server.

A2’s versatility, compact size, and low power exceed the demands for edge deployments at scale, instantly upgrading existing entry-level CPU servers to handle inference. Servers accelerated with A2 GPUs deliver up to 20X higher inference performance versus CPUs and 1.3x more efficient IVA deployments than previous GPU generations — all at an entry-level price point.

NVIDIA-Certified systems with the NVIDIA A2, A30, and A100 Tensor Core GPUs and NVIDIA AI—including the NVIDIA Triton Inference Server, open source inference service software—deliver breakthrough inference performance across edge, data center, and cloud. They ensure that AI-enabled applications deploy with fewer servers and less power, resulting in easier deployments and faster insights with dramatically lower costs.

Highlights

GPU Architecture NVIDIA Ampere
CUDA Cores 1280
Tensor Cores 40 | Gen 3
RT Cores 108 Gen 2
Peak FP32 4.5 TFLOPS
Peak TF32 Tensor Core 9 TFLOPS | 18 TFLOPS Sparsity
Peak FP16 Tensor Core 18 TFLOPS | 36 TFLOPS Sparsity
INT8 36 TOPS | 72 TOPS Sparsity
INT4 72 TOPS | 144 TOPS Sparsity
GPU Memory 16 GB GDDR6 ECC
Memory Bandwidth 200 GB/s
Thermal Solution Passive
Maximum Power Consumption 40-60 Watt | Configurable
System Interface PCIe Gen 4.0 x8



General

Brand
NVIDIA

Memory

Video memory
64 GB
Memory Type
GDDR6

Paketin tiedot

Paketin mitat
403 x 205 x 78 mm
Paketin paino
1,42 kg