Initial commit

This commit is contained in:
Zhongwei Li
2025-11-29 17:51:51 +08:00
commit aaa5832fbf
40 changed files with 3362 additions and 0 deletions

View File

@@ -0,0 +1,75 @@
---
description: Configure local AI providers for Warpio
allowed-tools: Write, Read
---
# Local AI Configuration
## Current Configuration
### Primary Provider: LM Studio
- **API URL:** http://192.168.86.20:1234/v1
- **Model:** qwen3-4b-instruct-2507
- **API Key:** lm-studio
- **Status:** ✅ Active
### Supported Providers
- **LM Studio** (Current) - Local model hosting
- **Ollama** - Alternative local model hosting
- **Custom OpenAI-compatible** - Any OpenAI-compatible API
## Configuration Options
### 1. Switch to Ollama
If you prefer to use Ollama instead of LM Studio:
```bash
# Update your .env file
echo "LOCAL_AI_PROVIDER=ollama" >> .env
echo "OLLAMA_API_URL=http://localhost:11434/v1" >> .env
echo "OLLAMA_MODEL=your-model-name" >> .env
```
### 2. Change Model
To use a different model in LM Studio:
```bash
# Update your .env file
echo "LMSTUDIO_MODEL=your-new-model-name" >> .env
```
### 3. Custom Provider
For other OpenAI-compatible APIs:
```bash
# Update your .env file
echo "LOCAL_AI_PROVIDER=custom" >> .env
echo "CUSTOM_API_URL=your-api-url" >> .env
echo "CUSTOM_API_KEY=your-api-key" >> .env
echo "CUSTOM_MODEL=your-model-name" >> .env
```
## Testing Configuration
After making changes, test with:
```bash
/warpio-local-test
```
## Environment Variables
The following variables control local AI behavior:
- `LOCAL_AI_PROVIDER` - Provider type (lmstudio/ollama/custom)
- `LMSTUDIO_API_URL` - LM Studio API endpoint
- `LMSTUDIO_MODEL` - LM Studio model name
- `OLLAMA_API_URL` - Ollama API endpoint
- `OLLAMA_MODEL` - Ollama model name
- `CUSTOM_API_URL` - Custom provider URL
- `CUSTOM_MODEL` - Custom provider model
## Next Steps
1. Update your `.env` file with desired configuration
2. Test the connection with `/warpio-local-test`
3. Check status with `/warpio-local-status`