NVIDIA A100 Tensor Core GPU

NVIDIA A100 Tensor Core GPU

Category

Segment :

Data Center GPU

Generation / Form Factor:

Accelerator

Description

A100 80GB PCIe A100 80GB SXM
FP64 9.7 TFLOPS
FP64 Tensor Core 19.5 TFLOPS
FP32 19.5 TFLOPS
Tensor Float 32 (TF32) 156 TFLOPS | 312 TFLOPS*
BFLOAT16 Tensor Core 312 TFLOPS | 624 TFLOPS*
FP16 Tensor Core 312 TFLOPS | 624 TFLOPS*
INT8 Tensor Core 624 TOPS | 1248 TOPS*
GPU Memory 80GB HBM2e 80GB HBM2e
GPU Memory Bandwidth 1,935 GB/s 2,039 GB/s
Max Thermal Design Power (TDP) 300W 400W ***
Multi-Instance GPU Up to 7 MIGs @ 10GB Up to 7 MIGs @ 10GB
Form Factor PCIe
Dual-slot air-cooled or single-slot liquid-cooled
SXM
Interconnect NVIDIA® NVLink® Bridge
for 2 GPUs: 600 GB/s **
PCIe Gen4: 64 GB/s
NVLink: 600 GB/s
PCIe Gen4: 64 GB/s
Server Options Partner and NVIDIA-Certified Systems™ with 1-8 GPUs NVIDIA HGX™ A100-Partner and NVIDIA-Certified Systems with 4, 8, or 16 GPUs; NVIDIA DGX™ A100 with 8 GPUs

Reviews

There are no reviews yet.

Be the first to review “NVIDIA A100 Tensor Core GPU”

Your email address will not be published. Required fields are marked *