Public Cloud

The AI cloud built
for developers.

Deploy AI-ready GPU instances in 60 seconds. Full root access, built-in IDE, real-time monitoring, and per-minute billing. No DevOps required.

60s
Deploy time
99.9%
Uptime SLA
70%
Cheaper than AWS

Developer Experience

Everything you need, nothing you don't.

Deploy in 60 Seconds

Pick a GPU, choose a template or bring your own Docker image, and deploy. No queues, no approvals, no waiting.

Built-in IDE

VS Code Server and Jupyter notebooks running in your browser. Pre-configured with CUDA, PyTorch, and your favorite tools.

Full Root Access

SSH into your instance with full root. Install anything, run anything. Your machine, your rules.

Real-time Monitoring

GPU utilization, memory, temperature, and spend — all live in your dashboard. Stream logs in real time.

Per-minute Billing

Pay only when your instances are running. Stop an instance, stop paying. No minimums, no commitments.

Team Collaboration

Invite your team, set roles and permissions, share projects and billing. Built for teams from day one.

Hardware

Latest NVIDIA
GPUs on tap.

From fine-tuning on a single A100 to distributed training across H200 clusters. Always available, always ready.

View pricing
H10080 GB · HBM3Training & Inference
H200141 GB · HBM3eLarge model training
B200192 GB · HBM3eFrontier training
A10080 GB · HBM2eFine-tuning & inference
L40S48 GB · GDDR6Inference & rendering

Developer Tools

Coming Soon

Build with Runcrate, your way.

We're building the tools to let you integrate Runcrate into every part of your workflow — from your terminal to your IDE to your AI agents.

REST API

Programmatic access

Full control over your instances, billing, and deployments through a RESTful API. Create, manage, and destroy instances programmatically.

POST /v1/instances/create
GET  /v1/instances/:id/status
DELETE /v1/instances/:id

Python & Node SDK

Native libraries

Type-safe SDKs for Python and Node.js. Deploy instances, manage billing, and stream logs — all from your application code.

from runcrate import Client
client = Client(api_key="...")
instance = client.deploy("h100")

CLI

Terminal-first

Deploy, SSH, tail logs, and manage instances from your terminal. Designed for developers who live in the command line.

$ runcrate deploy --gpu h100
$ runcrate ssh instance-a4f2
$ runcrate logs --follow

MCP Server

AI agent integration

Let your AI agents deploy and manage GPU instances autonomously. Model Context Protocol support for Claude, GPT, and any MCP-compatible agent.

"mcpServers": {
  "runcrate": { ... }
}

Ecosystem

Works with
everything.

Pre-configured with the tools and frameworks you already use. No setup, no configuration.

PyTorchPyTorch
TensorFlowTensorFlow
HuggingFaceHuggingFace
CUDACUDA
DockerDocker
JupyterJupyter
VS CodeVS Code
GitGit
PythonPython
Node.jsNode.js
wandbwandb
MLflowMLflow

Even if it doesn't, we'll make it work together.

Why Runcrate

Built different.

No DevOps tax

Every instance comes with CUDA drivers, security, monitoring, and backups pre-configured. You focus on your model, we handle the infrastructure.

No vendor lock-in

Standard Docker images, SSH access, and open APIs. Move your workloads in or out anytime. Your code runs anywhere.

No surprise bills

Per-minute billing with no egress fees, no data transfer costs, no hidden charges. Set budget limits and auto-stop to stay in control.

Start building on Runcrate.

Deploy your first GPU in under 60 seconds. No commitments, no credit card required to explore.

Pay-as-you-go
No upfront commitments
60s deploy
From signup to running GPU
Cancel anytime
No lock-in, no penalties