Skip to main content

GPUs for Running Stable Diffusion

··156 words

Note that this list is aimed at cloud GPUs where more expensive GPUs are comparatively cheap vs buying the whole GPU outright.

You can run stable diffusion on smaller/cheaper GPUs!

GPU VRAM (GB) Speed relative to H100 for SD Speed / $ Lowest cost per hour Cost at Runpod Cost at FluidStack Cost at Lambda Labs
RTX 4090 24 50% 👌 0.72 $0.69 ✅ $0.69 None None
H100 PCIe 80 🏆 100% 0.50 $1.99 None ✅ $1.99 ✅ $1.99
RTX 3090 24 21% 0.49 🪙 $0.44 ✅ $0.44 $0.59 None
RTX 3080 10 21% 0.43 $0.50 None $0.50 None
6000 Ada 48 48% 0.40 $1.19 $1.19 None None
A100 40 43% 0.39 $1.10 None $1.20 $1.10
L40 48 43% 0.36 $1.19 $1.19 None None
V100 16 24% 0.27 $0.87 None $0.87 None
A6000 48 19% 0.24 $0.79 $0.79 $0.80 $0.80
A40 48 19% 0.24 $0.79 $0.79 $1.57 None
A100 80 43% 0.24 $1.79 $1.79 $2.91 None