AI Inferencefaster than you can type
Build real-time AI applications with lightning fast inference (under ~120ms).
No coldstarts. Pay only for what you use.
Model Gallery
Check out some of the ready-to-use models. They have simple API endpoints ready for you to start your own AI-powered applications.
Stable Diffusion with LoRAs
Run Stable Diffusion with customizable LoRA weights.
text-to-image
inference
stylized
Stable Diffusion XL
Run SDXL at the speed of light
text-to-image
inference
Whisper
Whisper is a model for speech transcription and translation.
speech-to-text
inference
speech
Latent Consistency (SDXL & SDv1.5)
Produce high-quality images with minimal inference steps.
text-to-image
inference
Illusion Diffusion
Create illusions conditioned on image.
text-to-image
inference
stylized
Upscale Images
Upscale images by a given factor.
image-to-image
inference
utility
Pricing
Ship custom model endpoints with fine-grained control over idle timeout, max concurrency and autoscaling.
- CPU: 10
- Memory: 64GB
- GPU: A100 (40GB VRAM)
- Total Price: $0.00111/s
Unit Price
CPU
$0.00003/s
Memory
$0.000004/s
GPU A100
$0.001/s
GPU A10G
$0.0002/s
GPU T4
$0.00009/s
Storage
$1/GB/month
Join our community
Join the discussion around our product and help shape the future of AI.