We just raised 20M to revolutionize AI/ML cloud computing
Learn more

Run machine learning
inference 
at scale.

Only pay for what you use — no idle costs, just unparalleled speed and scalability.

1
2
3
4
5
import runpod
def handler(job):
job_input = job["input"]
return "Running on Runpod!"
runpod.serverless.start({"handler":handler})
runpodctl -- zsh


Spend more time training your models.
Let us handle your 
inference.

For your expected load, provision active workers running 24/7 with a 40% discount & flex workers to handle any sudden traffic.
Try it now
Bring Your Container
Network Storage
9 Regions
Streaming
Webhooks
Autoscale in seconds
Respond to user demand in real time with GPU workers that
scale from 0 to 100s in seconds.
Flex 
Workers
Active 
Workers
10 GPUs
6:24AM
100 GPUs
11:34AM
20 GPUs
1:34PM
Active Workers
-40% DISCOUNT
Dedicated GPUs that handle consistent workloads 24/7.
Get them at a lower cost so you don't break the bank for stable usage.
Flex Workers
Flexible GPUs that cost nothing when idle.
Ready to scale up as soon as your launch goes viral.
Zero Cold-Starts with Active Workers
No cold-start time — because the workers are always
running. Get instant execution when speed is all that matters.
<250ms Cold-Start with Flashboot
Flashboot is an optimization layer for our container system to manage deployments and scale up workers in real time.
Handle more consistent workloads like fine-tuning
Scale workers by Queue Delay or Request count
Monitor your endpoint with
real-time analytics
Usage Analytics
Real-time usage analytics for your endpoint with metrics on completed and failed requests. Useful for endpoints that have fluctuating usage profiles throughout the day.
See the console
Active
Requests
Completed:
2,277
Retried:
21
Failed:
9
Execution Time
Total:
1,420s
P70:
8s
P90:
19s
P98:
22s
Execution Time Analytics
Debug your endpoints with detailed metrics on execution time. Useful for hosting models that have varying execution times, like large language models. You can also monitor delay time, cold start time, cold start count, GPU utilization, and more.
See the console
Real-Time Logs
Get descriptive, real-time logs to show you exactly what's happening across your active and flex GPU workers at all times.
See the console
worker logs -- zsh
2024-03-15T19:56:00.8264895Z INFO | Started job db7c79
2024-03-15T19:56:03.2667597Z
0% | | 0/28 [00:00<?, ?it/s]
12% |██ | 4/28 [00:00<00:01, 12.06it/s]
38% |████ | 12/28 [00:00<00:01, 12.14it/s]
77% |████████ | 22/28 [00:01<00:00, 12.14it/s]
100% |██████████| 28/28 [00:02<00:00, 12.13it/s]
2024-03-15T19:56:04.7438407Z INFO | Completed job db7c79 in 2.9s
2024-03-15T19:57:00.8264895Z INFO | Started job ea1r14
2024-03-15T19:57:03.2667597Z
0% | | 0/28 [00:00<?, ?it/s]
15% |██ | 4/28 [00:00<00:01, 12.06it/s]
41% |████ | 12/28 [00:00<00:01, 12.14it/s]
80% |████████ | 22/28 [00:01<00:00, 12.14it/s]
100% |██████████| 28/28 [00:02<00:00, 12.13it/s]
2024-03-15T19:57:04.7438407Z INFO | Completed job ea1r14 in 2.9s
2024-03-15T19:58:00.8264895Z INFO | Started job gn3a25
2024-03-15T19:58:03.2667597Z
0% | | 0/28 [00:00<?, ?it/s]
18% |██ | 4/28 [00:00<00:01, 12.06it/s]
44% |████ | 12/28 [00:00<00:01, 12.14it/s]
83% |████████ | 22/28 [00:01<00:00, 12.14it/s]
100% |██████████| 28/28 [00:02<00:00, 12.13it/s]
2024-03-15T19:58:04.7438407Z INFO | Completed job gn3a25 in 2.9s

Cost effective
for every inference workload

Save 15% over other Serverless cloud providers on flex workers alone.
Create active workers and configure queue delay for even more savings.
80 GB
A100
Extremely performant GPUs, yet still very cost effective for running any machine learning model.
Flex
$0.0013/s
Active
$0.00078/s
80 GB
H100
PRO
Our most powerful GPUs. Most useful when maximizing inference throughput is critical.
Flex
$0.0025/s
Active
$0.0015/s
48 GB
A6000
A cost-effect option for running diffusion models, LoRAs, whisper, and many others. Less effective for large language models.
Flex
$0.00048/s
Active
$0.00029/s
48 GB
L40
PRO
Useful for when having high inference throughput on LLMs like Llama 3 7B and medium sized models like Yi 34B.
Flex
$0.00069/s
Active
$0.00041/s
24 GB
A500
Great for small-to-medium sized models with consistent workloads, lower throughput than 24GB PRO.
Flex
$0.00026/s
Active
$0.00016/s
24 GB
4090
PRO
Extremely high throughput for small to medium sized models.Great for running Llama3 8B and Mistral 7B.
Flex
$0.00044/s
Active
$0.00026/s
16 GB
A4000
The most cost-effective option for running inference on small models like LoRAs, diffusion models, and whisper.
Flex
$0.0002/s
Active
$0.00012/s
Thousands of GPUs across 9 Regions
Update your endpoint's region in two clicks. Scale up to 9 regions at a time. Global automated failover is supported out-of-the-box, so you won't have to worry about GPU errors interrupting your ML inference.
Pending Certifications
Although many of our data center partners have these compliance certifications, RunPod is in the process of getting SOC 2, ISO 27001, and HIPAA. We aim to have all three by early Q3, 2024.
gpu backgroundgpu backgroundgpu background
North America
UR-OR-1
CA-MTL-1
CA-MTL-2
European Union
EUR-IS-1
EUR-IS-2
EUR-NO-1
Europe
EU-NL-1
EU-RO-1
EU-SE-1
Serverless Pricing Calculator
seconds
$ 55 /mo
1
72,000 requests per month
1. Cost estimation includes 50% of the requests using active price & running into 1s cold-start.
We're with you from seed to scale
Book a call with our sales team to learn more.
Gain
Additional Savings
with Reservations
Save more by committing to longer-term usage. Reserve discounted active and flex workers by speaking with our team.
Book a call
Are you an early-stage startup or ML researcher?
Get up to $25K in free compute credits with Runpod. These can be used towards on-demand GPUs and Serverless endpoints.
Apply
Loved
by the
developer
community
RunPod is built by developers, for developers. Our community of 10,000+ developers on Discord is here for support while you get started.
Join our Discord
CTO, LOVO AI
Hara Kang
"There are definitely providers who offer much cheaper pricing than Runpod. But they always have an inferior developer experience. If you're paying 50% less for a GPU elsewhere, that cost is coming out somewhere else, be it developer time or lack of reliability. For the value, Runpod provides competitive prices and we're willing to pay a premium to reduce the headache that normally comes with ML ops."
Read more
CEO, Coframe
Josh Payne
"The setup process was great! Very quick and easy. RunPod had the exact GPUs we needed for AI inference and the pricing was very fair based on what I saw out on the market. The main value proposition for us was the flexibility RunPod offered. We were able to scale up effortlessly to meet the demand at launch."
Read more
CPO, KRNL.ai
Giacomo Locci
"The cost savings on RunPod have been incredible. Since switching, our team has been able to focus on building the product instead of the infrastructure. We often have unpredictable demand from our users which makes it hard to manage our cloud costs. But with RunPod, we've been able to scale up and down quickly and painlessly. Great reliability in multiple regions and great customer support is why we've been with them for over a year now."
Read more
CTO, LOVO AI
Hara Kang
"It really shows that RunPod is made by developers. They know exactly what engineers really want and they ship those features in order of importance."
Hara Kang - CTO, LOVO AI
4,221,728,376
4,221,727,876
requests & 100k+ developers since launch
Get started with RunPod 
today.
We handle millions of serverless requests a day. Scale your machine learning inference while keeping costs low.