New pricing: More AI power, less cost!
Learn more

Compare GPU Performance on AI Workloads

RTX 6000 Ada 48GB
RTX 6000 Ada 48GB image
Vs.
H100 80GB SXM
H100 80GB SXM image

LLM Benchmarks

Benchmarks were run on RunPod gpus using vllm. For more details on vllm, check out the vllm github repository.
Output Token Throughput (tok/s)
Get started with RunPod 
today.
We handle millions of gpu requests a day. Scale your machine learning workloads while keeping costs low with RunPod.
Get Started