QWEN API
The Qwen3 family covers chat, vision, text-to-speech, and code generation in a single model family. Access all of them through Runcrate's OpenAI-compatible API. Strong multilingual performance, long context windows, and competitive pricing.

QUICK START
from openai import OpenAI
client = OpenAI(
base_url="https://api.runcrate.ai/v1",
api_key="rc_live_YOUR_API_KEY",
)
response = client.chat.completions.create(
model="Qwen/Qwen3-32B",
messages=[
{"role": "user", "content": "Write a REST API in FastAPI with pagination."}
],
)
print(response.choices[0].message.content)AVAILABLE MODELS
| Model | Provider | Price | Detail |
|---|---|---|---|
| Qwen/Qwen3-235B-A22B | Alibaba | Per-token | 235B MoE, flagship model |
| Qwen/Qwen3-32B | Alibaba | Per-token | 32B dense, strong all-round |
| Qwen/Qwen3-14B | Alibaba | Per-token | 14B, cost-effective |
| Qwen/Qwen3-TTS | Alibaba | Per-token | Text-to-speech synthesis |
WHY RUNCRATE
Chat, vision, TTS, and code models all under one model family. Use the same API patterns across modalities.
Native support for Chinese, English, Japanese, Korean, Arabic, and 24 more languages. Strong performance on multilingual benchmarks.
Process long documents, large codebases, or extended conversations without hitting context limits.
Qwen models offer frontier-level quality at lower token costs than many alternatives, especially for Asian-language workloads.
FAQ