1.0 KiB
1.0 KiB
Replicate Hardware Options
Available hardware SKUs for model creation and deployments:
| SKU | Name | Use Case |
|---|---|---|
cpu |
CPU | Lightweight models, text processing |
gpu-t4 |
Nvidia T4 GPU | Cost-effective inference, smaller models |
gpu-l40s |
Nvidia L40S GPU | Balanced performance and cost |
gpu-l40s-2x |
2x Nvidia L40S GPU | Larger models requiring more VRAM |
gpu-a100-large |
Nvidia A100 (80GB) GPU | High-performance inference, large models |
gpu-a100-large-2x |
2x Nvidia A100 (80GB) GPU | Very large models, maximum VRAM |
gpu-h100 |
Nvidia H100 GPU | Latest generation, highest performance |
Hardware Selection Guidelines
- Image generation models (SDXL, Flux):
gpu-a100-largeorgpu-l40s - Large language models:
gpu-a100-largeor higher - Smaller inference tasks:
gpu-t4for cost efficiency - Fine-tuning/training:
gpu-a100-largeorgpu-h100recommended
To list current hardware options:
replicate hardware list