feat: initial ThinkCentre setup repo — Android, Ollama, Otto migration docs + scripts
This commit is contained in:
@@ -0,0 +1,66 @@
|
||||
# Ollama Models — ThinkCentre 1
|
||||
|
||||
CPU-only inference (Intel UHD 730, no dedicated GPU). 16 GB RAM.
|
||||
|
||||
## Installed
|
||||
|
||||
| Model | Size | RAM | Speed (tok/s) | Best for |
|
||||
|-------|------|-----|---------------|----------|
|
||||
| `qwen2.5:7b` | ~4.7 GB | ~6 GB | ~15-25 | Default — code, German, reasoning |
|
||||
|
||||
## Recommended Candidates
|
||||
|
||||
| Model | Size | RAM needed | Notes |
|
||||
|-------|------|-----------|-------|
|
||||
| `qwen2.5:7b` ✅ | 4.7 GB | 6 GB | Best quality/speed ratio on CPU |
|
||||
| `mistral:7b` | 4.1 GB | 5 GB | Strong English reasoning |
|
||||
| `llama3.2:3b` | 2.0 GB | 3 GB | Fastest, lower quality |
|
||||
| `qwen2.5:14b` | 9.0 GB | 11 GB | Better quality, slower (~8 tok/s) |
|
||||
| `deepseek-r1:7b` | 4.7 GB | 6 GB | Strong at reasoning/math |
|
||||
| `nomic-embed-text` | 0.3 GB | 1 GB | Embeddings (QMD alternative) |
|
||||
|
||||
## API
|
||||
|
||||
```bash
|
||||
# Chat
|
||||
curl http://192.168.0.91:11434/api/generate -d '{
|
||||
"model": "qwen2.5:7b",
|
||||
"prompt": "Your prompt here",
|
||||
"stream": false
|
||||
}'
|
||||
|
||||
# Via OpenAI-compatible endpoint
|
||||
curl http://192.168.0.91:11434/v1/chat/completions \
|
||||
-H "Content-Type: application/json" \
|
||||
-d '{"model":"qwen2.5:7b","messages":[{"role":"user","content":"Hello"}]}'
|
||||
```
|
||||
|
||||
## Management
|
||||
|
||||
```bash
|
||||
ollama list # Installed models
|
||||
ollama pull <model> # Download model
|
||||
ollama rm <model> # Remove model
|
||||
ollama run <model> # Interactive chat
|
||||
sudo systemctl status ollama
|
||||
sudo systemctl restart ollama
|
||||
```
|
||||
|
||||
## OpenClaw Integration (future)
|
||||
|
||||
Add to `openclaw.json` as fallback:
|
||||
```json
|
||||
{
|
||||
"agents": {
|
||||
"defaults": {
|
||||
"model": {
|
||||
"fallbacks": [
|
||||
"openrouter/anthropic/claude-sonnet-4-6",
|
||||
"ollama/qwen2.5:7b@http://192.168.0.91:11434",
|
||||
"google/gemini-2.5-flash"
|
||||
]
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
@@ -0,0 +1,26 @@
|
||||
#!/bin/bash
|
||||
# Ollama Setup — ThinkCentre 1
|
||||
# Usage: bash setup.sh [model]
|
||||
# Default model: qwen2.5:7b
|
||||
|
||||
set -e
|
||||
MODEL="${1:-qwen2.5:7b}"
|
||||
|
||||
echo "=== Ollama Setup ==="
|
||||
|
||||
echo "[1/3] Installing Ollama..."
|
||||
curl -fsSL https://ollama.ai/install.sh | sh
|
||||
|
||||
echo "[2/3] Enabling service..."
|
||||
sudo systemctl enable ollama
|
||||
sudo systemctl start ollama
|
||||
sleep 5
|
||||
|
||||
echo "[3/3] Pulling default model: $MODEL (~5 GB, dauert je nach Netz 5-15 min)..."
|
||||
ollama pull "$MODEL"
|
||||
|
||||
echo ""
|
||||
echo "=== Done ==="
|
||||
ollama list
|
||||
echo ""
|
||||
echo "Test: curl http://localhost:11434/api/generate -d '{\"model\":\"$MODEL\",\"prompt\":\"Hello\",\"stream\":false}'"
|
||||
Reference in New Issue
Block a user