Vast.ai Pricing 2026
Complete pricing guide with plans, and cost analysis
Vast.ai costs $0.29/GPU/hour for Interruptible. Plans range from $0.29 to $2.50/GPU/hour.
Vast.ai costs $0.29 to $2.50 per GPU/hour as of April 2026, with 3 plans available. Plans: Interruptible at $0.1/GPU/hour, On-Demand at $0.29/GPU/hour, and Reserved at $0.2/GPU/hour. Pricing depends on your chosen tier, contract length, and negotiated discounts.
Use the interactive pricing calculator to estimate your exact cost based on team size and requirements.
- Free tier: No free tier available
Vast.ai offers 3 pricing tiers: Interruptible, On-Demand, Reserved. Paid plans include Interruptible at $0.1/undefined, On-Demand at $0.29/undefined, Reserved at $0.2/undefined.
Compared to other ai/gpu cloud compute software, Vast.ai is positioned at the budget-friendly price point.
How much does Vast.ai cost?
Vast.ai Pricing Overview
Vast.ai has 3 pricing plans ranging from $0.29 to $2.50/GPU/hour. The Interruptible plan costs $0.10/GPU/hour. The On-Demand plan costs $0.29/GPU/hour. The Reserved plan costs $0.20/GPU/hour.
This pricing was last verified in April 13, 2026 from 1 independent sources.
Vast.ai is a ai gpu cloud solution. Compare pricing plans and costs below.
All Vast.ai Plans & Pricing
| Plan | Monthly | Annual | Best For |
|---|---|---|---|
| Interruptible | $0.1 / | Custom | — |
| On-Demand | $0.29 / | Custom | — |
| Reserved | $0.2 / | Custom | — |
View all features by plan
Interruptible
- 50%+ cheaper than on-demand rates
- Per-second billing, no minimum hours
- 68+ GPU types available
- RTX 4090 from ~$0.15/hr interruptible
- Can be reclaimed by host with short notice
- Suited for fault-tolerant and batch workloads
On-Demand
- Guaranteed uptime, high priority
- Per-second billing, no minimum hours
- RTX 4090 from $0.29/hr
- A100 from $0.29/hr (varies by host)
- H100 from ~$1.47/hr
- Prices set by individual hosts (marketplace)
Reserved
- Up to 50% discount vs on-demand
- 1, 3, or 6-month commitment terms
- Upfront payment required
- Guaranteed capacity for commitment period
How Vast.ai Pricing Compares
| Software | Starting Price | Top Price |
|---|---|---|
| Vast.ai | $0.29/GPU/hour | $2.5/GPU/hour |
| RunPod | $0.34/GPU/hour | $3.49/GPU/hour |
| Lambda | $0.69/GPU/hour | $6.99/GPU/hour |
| CoreWeave | $10/instance/hour | $68.8/instance/hour |
| Hyperbolic | $0.3/GPU/hour | $3.2/GPU/hour |
| Paperspace | $0.56/GPU/hour | $5.95/GPU/hour |
Detailed pricing comparisons:
Vast.ai Pricing FAQ
01 How does Vast.ai pricing work?
Vast.ai operates as a GPU marketplace where independent hosts set their own prices based on supply and demand. Prices are dynamic and vary by GPU type, host reliability, location, and market conditions. You pay per second with no minimum commitment on on-demand instances. Interruptible instances are 50%+ cheaper but can be reclaimed by hosts. A $5 minimum deposit is required to start.
02 How much does an H100 cost on Vast.ai?
H100 GPUs on Vast.ai are available from approximately $1.47/hr on the marketplace, significantly cheaper than traditional cloud providers like AWS or Lambda Labs. Prices fluctuate with supply and demand — interruptible H100 instances can be even lower. For comparison, on-demand H100s on Lambda Labs list at $4.29/hr.
03 How much does an RTX 4090 cost on Vast.ai?
RTX 4090 instances on Vast.ai start from $0.29/hr for on-demand and can be lower on interruptible pricing. Stable on-demand instances typically run $0.39–$0.59/hr. This makes it one of the cheapest places to rent RTX 4090s for inference and fine-tuning workloads.
04 Does Vast.ai charge any platform fees?
Vast.ai does not charge platform fees or markups on top of host-set prices. You pay the host's listed rate directly. There are no hidden fees, though you also pay separately for storage and bandwidth at rates set by individual hosts.
05 Is Vast.ai reliable for production workloads?
Vast.ai's on-demand instances offer guaranteed uptime and are suitable for production inference. However, as a peer-to-peer marketplace, host reliability varies. Interruptible instances are not suitable for latency-sensitive or stateful production workloads. For critical workloads, on-demand instances from highly-rated hosts are recommended.
Is this pricing incorrect? — we'll verify and update it.