NVIDIA
AILimited
NVIDIA Authorized partnerDatacenter GPUs

NVIDIA H100 80GB PCIe

SKU H100-PCIE-80GB

Hopper-architecture accelerator for LLM training and inference.

MSRP indication
$30,000
Quote pricing may differ — request below.
View RFQ

Overview

80 GB HBM3 memory at 2 TB/s. Designed for transformer engine workloads and FP8 inference.

Specifications

Architecture
Hopper
Memory
80 GB HBM3 @ ~2 TB/s
Form factor
Dual-slot PCIe Gen5
TDP
350 W
FP8 Tensor
Up to 3,026 TFLOPS
NVIDIA
Sourced from NVIDIA
Authorized Qubit partner
Vendor site
Need it bundled with deployment? Mention in your RFQ — we handle install and support too.

More in Datacenter GPUs

NVIDIA
NVIDIADatacenter GPUs

NVIDIA L40S 48GB

Universal GPU for AI inference, training, and graphics workloads.

MSRP $8,190
Details