Skip to main content

Available GPUs

Private Cloud clusters are available with the latest NVIDIA data center GPUs, optimized for AI training and inference at scale.

GPU Options

NVIDIA H100 SXM

The industry standard for large-scale AI training.
SpecValue
GPU Memory80 GB HBM3
Memory Bandwidth3.35 TB/s
FP16 Performance989 TFLOPS
InterconnectNVLink 4.0 (900 GB/s)
Best ForLLM training, fine-tuning, high-throughput inference

NVIDIA H200 SXM

Enhanced H100 with more memory and bandwidth for memory-bound workloads.
SpecValue
GPU Memory141 GB HBM3e
Memory Bandwidth4.8 TB/s
FP16 Performance989 TFLOPS
InterconnectNVLink 4.0 (900 GB/s)
Best ForLarge model inference, long-context training, models that need more VRAM

NVIDIA B200

Next-generation Blackwell architecture with massive compute gains.
SpecValue
GPU Memory192 GB HBM3e
Memory Bandwidth8 TB/s
FP16 Performance2,250 TFLOPS
FP4 Performance9,000 TFLOPS
InterconnectNVLink 5.0 (1,800 GB/s)
Best ForFrontier model training, next-gen inference, FP4 quantized workloads

NVIDIA B300

The latest Blackwell Ultra with maximum memory and performance.
SpecValue
GPU Memory288 GB HBM3e
Memory Bandwidth12 TB/s
FP16 Performance2,250 TFLOPS
FP4 Performance9,000 TFLOPS
InterconnectNVLink 5.0 (1,800 GB/s)
Best ForLargest-scale training, trillion-parameter models, maximum memory capacity

Choosing a GPU

GPUMemoryBest ForAvailability
H10080 GBGeneral AI training, proven and widely supportedHigh
H200141 GBMemory-hungry models, large batch inferenceModerate
B200192 GBNext-gen training, 2x compute over H100Growing
B300288 GBMaximum scale, highest memory capacityLimited
Not sure which GPU is right for your workload? Contact us at support@runcrate.ai and our team will help you choose the optimal configuration.

Networking

All Private Cloud clusters include high-speed interconnect:
GPUIntra-Node (NVLink)Inter-Node (InfiniBand)
H100900 GB/s400 Gb/s NDR
H200900 GB/s400 Gb/s NDR
B2001,800 GB/s400 Gb/s NDR
B3001,800 GB/s400 Gb/s NDR