AI Hardware
26 GPUs & accelerators · self-host price guide · cloud rates
23 hardware items · Datacenter
- Google TPU v5e
Google · 2023 · 819 GB/s
16 GBVRAM▾$0$1.20/hr/hr cloudFits: Up to 70B in pods (multi-chip)
Price (new): $0Cloud rental: $1.20/hr (GCP) · only available as serviceMemory BW: 819 GB/sGCP-only. Cheap, fast for JAX/TF workloads. Works with vLLM via JetStream.
- NVIDIA T4 (16GB)
NVIDIA · 2018 · 65 TFLOPS · 320 GB/s
16 GBGDDR6▾$900 used$0.35/hr/hr cloudFits: 7-13B 4-bit
Price (used): $900Cloud rental: $0.35/hr (AWS g4dn)FP16 compute: 65 TFLOPSMemory BW: 320 GB/sPower: 70W TDPVery cheap to rent. Too slow for production serving but fine for batch inference.
- NVIDIA L4 (24GB)
NVIDIA · 2023 · 121 TFLOPS · 300 GB/s
24 GBGDDR6▾$2,500$0.44/hr liveFits: 13-14B FP16 / 30B 4-bit
Price (new): $2,500Cloud on-demand: $0.44/hr · RunPod (live)FP16 compute: 121 TFLOPSMemory BW: 300 GB/sPower: 72W TDPModern replacement for T4. Single-slot, low-power — great for density.
- NVIDIA A100 40GBSweet spot
NVIDIA · 2020 · 312 TFLOPS · 1555 GB/s
40 GBHBM2e▾$8,000 used$1.10/hr/hr cloudFits: 30B FP16 / 70B 4-bit
Price (used): $8,000Cloud rental: $1.10/hr (RunPod) / $3.06/hr (AWS)FP16 compute: 312 TFLOPSMemory BW: 1555 GB/sPower: 400W TDPThe de-facto standard for serious AI training + inference. Huge ecosystem. Used price dropping fast.
- NVIDIA A100 80GB
NVIDIA · 2021 · 312 TFLOPS · 2039 GB/s
80 GBHBM2e▾$12,000 used$1.60/hr/hr cloudFits: 70B FP16 / 200B+ 4-bit
Price (used): $12,000Cloud rental: $1.60/hr (RunPod) / $4.10/hr (AWS p4de)FP16 compute: 312 TFLOPSMemory BW: 2039 GB/sPower: 400W TDPMost rented ML GPU. 80GB fits Llama 3 70B at FP16 on a single card.
- AMD MI300X (192GB)Best value
AMD · 2023 · 1307 TFLOPS · 5300 GB/s
192 GBHBM3▾$15,000$0.50/hr liveFits: 405B FP16 / frontier models at 4-bit
Price (new): $15,000Cloud on-demand: $0.50/hr · RunPod (live)FP16 compute: 1307 TFLOPSMemory BW: 5300 GB/sPower: 750W TDPMore VRAM than H100 at half the price. ROCm is decent now — works with vLLM, PyTorch, llama.cpp.
- NVIDIA H100 (80GB)Flagship
NVIDIA · 2022 · 989 TFLOPS · 3350 GB/s
80 GBHBM3▾$30,000$2.50/hr/hr cloudFits: 70B FP16 / 200B+ 4-bit / 405B with 2+ cards
Price (new): $30,000Cloud rental: $2.50/hr (RunPod) / $8.00/hr (AWS p5)FP16 compute: 989 TFLOPSMemory BW: 3350 GB/sPower: 700W TDP3x faster than A100 for modern transformer workloads. FP8 support doubles it again.
- NVIDIA H200 (141GB)
NVIDIA · 2024 · 989 TFLOPS · 4800 GB/s
141 GBHBM3e▾$40,000$3.50/hr/hr cloudFits: Llama 3 405B at 4-bit on ONE card. 70B FP16 with massive batch size.
Price (new): $40,000Cloud rental: $3.50/hr (RunPod)FP16 compute: 989 TFLOPSMemory BW: 4800 GB/sPower: 700W TDPSame compute as H100 but 76% more VRAM + 43% more bandwidth. Huge win for long context.
- NVIDIA B200 (192GB)Flagship
NVIDIA · 2024 · 2250 TFLOPS · 8000 GB/s
192 GBHBM3e▾$45,000$5.00/hr/hr cloudFits: Single card runs Llama 405B FP16 with room to spare
Price (new): $45,000Cloud rental: $5.00/hr (limited availability)FP16 compute: 2250 TFLOPSMemory BW: 8000 GB/sPower: 1000W TDPBlackwell gen. 2.5x H100 perf. FP4 native. Current king for new deployments.
- NVIDIA GB200 (NVL72)
NVIDIA · 2024 · 162000 TFLOPS · 576000 GB/s
13824 GBHBM3e▾$3,000,000Fits: Frontier training — GPT-5-scale models
Price (new): $3,000,000FP16 compute: 162000 TFLOPSMemory BW: 576000 GB/sPower: 120000W TDPFull rack: 72 Blackwell GPUs + 36 Grace CPUs. Only relevant if you're training a frontier model.
- NVIDIA A10G (24GB)
NVIDIA · 2021 · 125 TFLOPS · 600 GB/s
24 GBGDDR6▾rent only$1.00/hr/hr cloudFits: 13-14B FP16 / 30B 4-bit
Cloud rental: $1.00/hr (AWS g5)FP16 compute: 125 TFLOPSMemory BW: 600 GB/sPower: 150W TDPAWS's workhorse inference GPU. 4x A10G matches a single A100 40GB for many workloads.
- Google TPU v5p
Google · 2024 · 2765 GB/s
95 GBVRAM▾rent only$4.20/hr/hr cloudFits: Frontier training (Gemini-scale)
Cloud rental: $4.20/hr (GCP)Memory BW: 2765 GB/sTop-tier GCP training chip. Compares to H100-H200 for transformer workloads.
- A100 PCIe
NVIDIA · 2026
80 GBVRAM▾rent onlyFits: 80GB VRAM — large models
Auto-discovered via RunPod live pricing.
- A100 SXM
NVIDIA · 2026
80 GBVRAM▾rent onlyFits: 80GB VRAM — large models
Auto-discovered via RunPod live pricing.
- B300
NVIDIA · 2026
288 GBVRAM▾rent onlyFits: 288GB VRAM — large models
Auto-discovered via RunPod live pricing.
- H100 SXM
NVIDIA · 2026
80 GBVRAM▾rent onlyFits: 80GB VRAM — large models
Auto-discovered via RunPod live pricing.
- H100 NVL
NVIDIA · 2026
94 GBVRAM▾rent onlyFits: 94GB VRAM — large models
Auto-discovered via RunPod live pricing.
- H100 PCIe
NVIDIA · 2026
80 GBVRAM▾rent onlyFits: 80GB VRAM — large models
Auto-discovered via RunPod live pricing.
- H200 SXM
NVIDIA · 2026
141 GBVRAM▾rent onlyFits: 141GB VRAM — large models
Auto-discovered via RunPod live pricing.
- NVIDIA H200 NVL
NVIDIA · 2026
143 GBVRAM▾rent onlyFits: 143GB VRAM — large models
Auto-discovered via RunPod live pricing.
- RTX PRO 6000 MaxQ
NVIDIA · 2026
96 GBVRAM▾rent onlyFits: 96GB VRAM — large models
Auto-discovered via RunPod live pricing.
- RTX PRO 6000
NVIDIA · 2026
96 GBVRAM▾rent onlyFits: 96GB VRAM — large models
Auto-discovered via RunPod live pricing.
- RTX PRO 6000 WK
NVIDIA · 2026
96 GBVRAM▾rent onlyFits: 96GB VRAM — large models
Auto-discovered via RunPod live pricing.