Subscribe to EDOM TECH Newsletter
NVIDIA
NVIDIA A30 - Tensor Core GPU
A30
Versatile compute acceleration for mainstream enterprise servers.
AI Inference and Mainstream Compute for Every Enterprise
Bring accelerated performance to every enterprise workload with NVIDIA A30 Tensor Core GPUs. With NVIDIA Ampere architecture Tensor Cores and Multi-Instance GPU (MIG), it delivers speedups securely across diverse workloads, including AI inference at scale and high-performance computing (HPC) applications. By combining fast memory bandwidth and low-power consumption in a PCIe form factor—optimal for mainstream servers—A30 enables an elastic data center and delivers maximum value for enterprises.
Product Specifications
FP64 | 5.2 teraFLOPS | |
FP64 Tensor Core | 10.3 teraFLOPS | |
FP32 | 10.3 teraFLOPS | |
TF32 Tensor Core | 82 teraFLOPS | 165 teraFLOPS* | |
BFLOAT16 Tensor Core | 165 teraFLOPS | 330 teraFLOPS* | |
FP16 Tensor Core | 165 teraFLOPS | 330 teraFLOPS* | |
INT8 Tensor Core | 330 TOPS | 661 TOPS* | |
INT4 Tensor Core | 661 TOPS | 1321 TOPS* | |
Media engines | 1 optical flow accelerator (OFA) 1 JPEG decoder (NVJPEG) 4 video decoders (NVDEC) |
|
GPU memory | 24GB HBM2 | |
GPU memory bandwidth | 933GB/s | |
Interconnect | PCIe Gen4: 64GB/s Third-gen NVLINK: 200GB/s** |
|
Form factor | Dual-slot, full-height, full-length (FHFL) | |
Max thermal design power (TDP) | 165W | |
Multi-Instance GPU (MIG) | 4 GPU instances @ 6GB each 2 GPU instances @ 12GB each 1 GPU instance @ 24GB |
|
Virtual GPU (vGPU) software support | NVIDIA AI Enterprise NVIDIA Virtual Compute Server |
** NVLink Bridge for up to two GPUs
NVIDIA Authorized Distributor