Various models to be used in llama.cpp CI workflow.
Do not use it in production.
Run this model on powerful GPU infrastructure. Deploy in 60 seconds.
Deploy on H100, A100, or RTX GPUs. Pay only for what you use. No setup required.