TOGETHER AI ALTERNATIVE

Like Together AI, but with GPU rentals too.

Together AI ships open-source LLM inference. Runcrate matches that — DeepSeek, Llama, Qwen, Mistral, all OpenAI-compatible — and adds image / video / audio models, plus on-demand GPU rentals (H100, H200, B200) when you need to fine-tune or self-host. One platform for inference and training.

200+
Models
OpenAI-compatible
Format
Per-second
Billing

COMPARISON

Runcrate vs Together AI.

Open-source LLMs
Runcrate: 200+ models
Together AI: 100+ models
Image / video models
Runcrate: FLUX, Sora, Veo, Kling
Together AI: Limited image (FLUX)
Audio models
Runcrate: Whisper, TTS, Voxtral
Together AI: Not offered
GPU rentals
Runcrate: H100, H200, B200, MI300X
Together AI: Not offered direct
Fine-tuning
Runcrate: Self-host via instances + LoRA
Together AI: Together fine-tuning service
OpenAI-compatible
Runcrate: Yes
Together AI: Yes

GPU PRICING

GPU pricing comparison.

deepseek-ai/DeepSeek-V3.2
DeepSeek$0.27 / 1M
Reasoning, code, 128K ctx
anthropic/claude-4-sonnet
Anthropic$3 / 1M in, $15 / 1M out
Top-tier reasoning
meta-llama/Llama-4-Scout
Meta$0.20 / 1M
Open weights, multilingual
Qwen/Qwen3-Max
Alibaba$0.30 / 1M
30+ languages, 128K ctx
openai/whisper-large-v3
OpenAI$0.02 / min
Speech-to-text, 100+ langs
black-forest-labs/FLUX.1-pro
Black Forest Labs$0.04 / image
Photorealistic

WHY SWITCH

Why teams switch to Runcrate.

200+ models, one API key

Chat, code, image, video, audio, embeddings, vision — all under a single OpenAI-compatible endpoint with per-token / per-image / per-second billing.

OpenAI-compatible drop-in

Swap the base URL and your existing OpenAI SDK code keeps working. No custom client library, no rewrite, no lock-in.

Inference + GPU rentals

When the API isn't enough, rent a dedicated H100, H200, or B200 from the same account — same billing, same dashboard, no separate vendor.

Per-second billing, no minimums

Pay only for what you use. No hourly bucketing, no commitment, no idle charges. Prepaid credits never expire.

GET STARTED

Try it now.

import Runcrate from "@runcrate/sdk";

const rc = new Runcrate({ apiKey: "rc_live_YOUR_API_KEY" });

const response = await rc.chat.completions.create({
  model: "deepseek/deepseek-v3.2",
  messages: [{ role: "user", content: "Hello from Runcrate" }],
});

console.log(response.choices[0].message.content);

FAQ

Common questions.

Try the Together AI alternative.