Estimate inference costs and throughput for your LLM deployment
| GPU Type | Memory | AWS ($/hour) | GCP ($/hour) | Azure ($/hour) | Best For | 
|---|---|---|---|---|---|
| NVIDIA A10G | 24 GB | $1.10 | $1.28 | $1.15 | 7B models, cost-sensitive | 
| NVIDIA A100 40GB | 40 GB | $4.10 | $3.90 | $4.25 | 13B-70B models | 
| NVIDIA A100 80GB | 80 GB | $5.00 | $4.80 | $5.20 | 70B models, high batch | 
| NVIDIA H100 | 80 GB | $8.10 | $7.90 | $8.40 | Maximum performance |