AI SummaryIntelli-router automatically triages user messages by complexity and routes them to the most cost-effective AI model (local Ollama, Claude Sonnet, Codex, or Claude Opus). Developers building multi-model AI systems benefit from reduced costs and optimized latency.
Install
Copy this and paste it into Claude Code, Cursor, or any AI assistant:
I want to install the "intelli-router" skill in my project. Please run this command in my terminal: # Install skill into the correct directory mkdir -p .claude/skills/IntelliRouter && curl --retry 3 --retry-delay 2 --retry-all-errors -o .claude/skills/IntelliRouter/SKILL.md "https://raw.githubusercontent.com/ColeMatthewBienek/IntelliRouter/main/SKILL.md" Then restart Claude Code (or reload the window in Cursor) so the skill is picked up.
Description
Intelligent model router that triages every incoming user message by complexity using a local Ollama model (qwen3:1.7b), then routes to the optimal AI model via session_status. Use on ALL incoming user messages to automatically select between local models (dengcao/Qwen3-32B:Q5_K_M via Ollama) for simple tasks, Claude Sonnet for moderate tasks, Codex for coding tasks, and Claude Opus for complex tasks. Triggers on any user message that is not a slash command. Keywords: model routing, triage, classify, complexity, cost optimization, smart routing.
Purpose
Automatically select the best model for each user message based on task complexity. Saves money by handling simple tasks locally and only escalating to expensive API models when the task requires it.
When to Activate
Run the classification step on EVERY incoming user message BEFORE generating a response. Do NOT classify messages that start with "/" (these are commands, not tasks).
Workflow
• Take the user's full message text. • Run the classifier: ` python3 {baseDir}/scripts/classify.py "USER_MESSAGE_HERE" ` • Parse the JSON output. It returns: `json {"model": "provider/model-name", "tier": "simple|moderate|coding|complex", "reason": "..."} ` • Use the model value from the output to switch the session model: ` session_status(model="<value from model field>") ` • Then proceed to answer the user's message using the newly selected model.
Tier Routing Table
| Tier | Model | Why | |----------|-----------------------------------|----------------------------------------| | simple | ollama/dengcao/Qwen3-32B:Q5_K_M | Free local model, handles basic tasks | | moderate | anthropic/claude-sonnet-4-5 | Balanced cost and capability | | coding | openai-codex/gpt-5.2 | Optimized for code generation/debugging| | complex | anthropic/claude-opus-4-5 | Maximum reasoning power |
Discussion
Health Signals
My Fox Den
Community Rating
Sign in to rate this booster