Initial commit
This commit is contained in:
25
skills/replicate-cli/references/hardware.md
Normal file
25
skills/replicate-cli/references/hardware.md
Normal file
@@ -0,0 +1,25 @@
|
||||
# Replicate Hardware Options
|
||||
|
||||
Available hardware SKUs for model creation and deployments:
|
||||
|
||||
| SKU | Name | Use Case |
|
||||
|-----|------|----------|
|
||||
| `cpu` | CPU | Lightweight models, text processing |
|
||||
| `gpu-t4` | Nvidia T4 GPU | Cost-effective inference, smaller models |
|
||||
| `gpu-l40s` | Nvidia L40S GPU | Balanced performance and cost |
|
||||
| `gpu-l40s-2x` | 2x Nvidia L40S GPU | Larger models requiring more VRAM |
|
||||
| `gpu-a100-large` | Nvidia A100 (80GB) GPU | High-performance inference, large models |
|
||||
| `gpu-a100-large-2x` | 2x Nvidia A100 (80GB) GPU | Very large models, maximum VRAM |
|
||||
| `gpu-h100` | Nvidia H100 GPU | Latest generation, highest performance |
|
||||
|
||||
## Hardware Selection Guidelines
|
||||
|
||||
- **Image generation models (SDXL, Flux)**: `gpu-a100-large` or `gpu-l40s`
|
||||
- **Large language models**: `gpu-a100-large` or higher
|
||||
- **Smaller inference tasks**: `gpu-t4` for cost efficiency
|
||||
- **Fine-tuning/training**: `gpu-a100-large` or `gpu-h100` recommended
|
||||
|
||||
To list current hardware options:
|
||||
```bash
|
||||
replicate hardware list
|
||||
```
|
||||
Reference in New Issue
Block a user