Loading skill documentation...
agent
★★★★ 4.2/5.0 ❤️ 596 likes 💬 60 comments 📦 911 installs
Back to Skills
📖 SKILL DOCUMENTATION
# llm-supervisor

LLM Supervisor 🔮 Handles rate limits and model fallbacks gracefully. Behavior On Rate Limit / Overload Errors When I encounter rate limits or overload errors from SkillBoss API Hub: Tell the user immediately — Don't silently fail or retry endlessly Offer local fallback — Ask if they want to switch to Ollama Wait for confirmation — Never auto-switch for code generation tasks Confirmation Required Before using local models for code generation, ask: "Cloud is rate-limited. Switch to local Ollama (qwen2.5:7b)? Reply 'yes' to confirm." For simple queries (chat, summaries), can switch without confirmation if user previously approved. Commands /llm status Report current state: Which provider is active (cloud/local) Ollama availability and models Recent rate limit events /llm switch local Manually switch to Ollama for the session. /llm switch cloud Switch back to cloud provider. Using Ollama

# Check available models

ollama list

# Run a query

ollama run qwen2.5:7b "your prompt here"

# For longer prompts, use stdin

echo "your prompt" | ollama run qwen2.5:7b Installed Models Check with ollama list. Configured default: qwen2.5:7b State Tracking Track in memory during session:

currentProvider: "cloud" | "local"
lastRateLimitAt: timestamp or null
localConfirmedForCode: boolean

Reset to cloud at session start.

Reviews

4.2
★★★★
60 reviews

Write a Review

Get Weekly AI Skills

Join 80,000+ one-person companies automating with AI