LocalAIPrices
← Back to comparison
NVIDIA

NVIDIA A100 80GB PCIe

Ampere · A-Series · ACTIVE

We earn commissions from purchases made through links on this site. This doesn't affect our rankings or recommendations.
Prices updated 5 days ago
New price
Used avg (eBay 30d)
$/GB VRAM (best)
$69/GB
MSRP: $11,000

Specifications

VRAM80 GB
Memory typeHBM2e
Bus width5120-bit
Memory bandwidth2,039 GB/s
CUDA cores6,912
Tensor cores432
FP1677.97 TFLOPS
TDP300W
Power connector8-pin
Card length267 mm
Slot width2 slots
PCIeGen 4 x16
CUDA compute8.0
Max model (Q4)~156B parameters

Inference Benchmarks (Q4_K_M)

Llama 3.3 8B
110.0 tok/s
Qwen 3 32B
48.0 tok/s
Llama 3.3 70B
22.0 tok/s

llama.cpp, batch_size=1, ctx=4096, single GPU.

What Can You Run?

ModelQ4_K_MQ8_0FP16
Llama 3.3 8B8BExcellent~110 tok/sUsableUsable
Llama 3.3 70B70.6BGood~22 tok/sUsableWon't fit
Qwen 3 8B8.2BUsableUsableUsable
Qwen 3 32B32.8BExcellent~48 tok/sUsableUsable
DeepSeek R1 70B70.6BUsableUsableWon't fit
Mistral Nemo 12B12.2BUsableUsableUsable
Phi-4 14B14BUsableUsableUsable
Gemma 3 27B27.4BUsableUsableUsable
Codestral 25B25.3BUsableUsableUsable
Command R 35B35BUsableUsableUsable

Notes

Available used from cloud decommissions. HBM2e bandwidth is excellent for inference.