General FAQ
What is Runcrate?
What is Runcrate?
Runcrate is a unified AI platform that combines three core products:
- Models API — Access 200+ AI models through a single, OpenAI-compatible endpoint
- GPU Instances — On-demand GPU-powered Linux environments with SSH access
- Cloud Storage — Persistent volumes with a built-in file explorer
Who is Runcrate for?
Who is Runcrate for?
Runcrate is designed for:
- AI developers building applications with LLMs, image generation, or other AI models
- Researchers who need GPU compute for training, fine-tuning, or experiments
- Indie hackers who want fast, affordable access to AI infrastructure
- Teams who need shared projects with role-based access and billing
What models are available?
What models are available?
Over 200 models across 8 categories: Chat, Reasoning, Code, Vision, Image Generation, Video Generation, Text-to-Speech, and Speech-to-Text. This includes models from leading families like GPT-4o, Claude, DeepSeek, Llama, Qwen, Gemini, FLUX, Sora, and more. See the Model Catalog for the full list.
Do I need to install anything?
Do I need to install anything?
No. The Models API is accessible via HTTP requests from any language or framework. The dashboard is fully web-based. For GPU instances, you only need an SSH client (built into macOS, Linux, and Windows).
Is there a free tier?
Is there a free tier?
There is no free tier. Runcrate uses a prepaid credit system — you add credits and only pay for what you use. You can start with as little as $5.
How fast are deployments?
How fast are deployments?
- Models API — Instant. Make your first API call as soon as you have a key.
- GPU Instances — Typically 1 to 3 minutes from deployment to SSH access.
How can I get support?
How can I get support?
- Discord — Join our Discord community for real-time help
- Email — Contact us at support@runcrate.ai