26 lines
1.0 KiB
Markdown
26 lines
1.0 KiB
Markdown
# Replicate Hardware Options
|
|
|
|
Available hardware SKUs for model creation and deployments:
|
|
|
|
| SKU | Name | Use Case |
|
|
|-----|------|----------|
|
|
| `cpu` | CPU | Lightweight models, text processing |
|
|
| `gpu-t4` | Nvidia T4 GPU | Cost-effective inference, smaller models |
|
|
| `gpu-l40s` | Nvidia L40S GPU | Balanced performance and cost |
|
|
| `gpu-l40s-2x` | 2x Nvidia L40S GPU | Larger models requiring more VRAM |
|
|
| `gpu-a100-large` | Nvidia A100 (80GB) GPU | High-performance inference, large models |
|
|
| `gpu-a100-large-2x` | 2x Nvidia A100 (80GB) GPU | Very large models, maximum VRAM |
|
|
| `gpu-h100` | Nvidia H100 GPU | Latest generation, highest performance |
|
|
|
|
## Hardware Selection Guidelines
|
|
|
|
- **Image generation models (SDXL, Flux)**: `gpu-a100-large` or `gpu-l40s`
|
|
- **Large language models**: `gpu-a100-large` or higher
|
|
- **Smaller inference tasks**: `gpu-t4` for cost efficiency
|
|
- **Fine-tuning/training**: `gpu-a100-large` or `gpu-h100` recommended
|
|
|
|
To list current hardware options:
|
|
```bash
|
|
replicate hardware list
|
|
```
|