Skip to content
Skill

intelli-router

by ColeMatthewBienek

AI Summary

Intelli-router automatically triages user messages by complexity and routes them to the most cost-effective AI model (local Ollama, Claude Sonnet, Codex, or Claude Opus). Developers building multi-model AI systems benefit from reduced costs and optimized latency.

Install

Copy this and paste it into Claude Code, Cursor, or any AI assistant:

I want to install the "intelli-router" skill in my project.

Please run this command in my terminal:
# Install skill into the correct directory
mkdir -p .claude/skills/IntelliRouter && curl --retry 3 --retry-delay 2 --retry-all-errors -o .claude/skills/IntelliRouter/SKILL.md "https://raw.githubusercontent.com/ColeMatthewBienek/IntelliRouter/main/SKILL.md"

Then restart Claude Code (or reload the window in Cursor) so the skill is picked up.

Description

Intelligent model router that triages every incoming user message by complexity using a local Ollama model (qwen3:1.7b), then routes to the optimal AI model via session_status. Use on ALL incoming user messages to automatically select between local models (dengcao/Qwen3-32B:Q5_K_M via Ollama) for simple tasks, Claude Sonnet for moderate tasks, Codex for coding tasks, and Claude Opus for complex tasks. Triggers on any user message that is not a slash command. Keywords: model routing, triage, classify, complexity, cost optimization, smart routing.

Purpose

Automatically select the best model for each user message based on task complexity. Saves money by handling simple tasks locally and only escalating to expensive API models when the task requires it.

When to Activate

Run the classification step on EVERY incoming user message BEFORE generating a response. Do NOT classify messages that start with "/" (these are commands, not tasks).

Workflow

• Take the user's full message text. • Run the classifier: ` python3 {baseDir}/scripts/classify.py "USER_MESSAGE_HERE" ` • Parse the JSON output. It returns: `json {"model": "provider/model-name", "tier": "simple|moderate|coding|complex", "reason": "..."} ` • Use the model value from the output to switch the session model: ` session_status(model="<value from model field>") ` • Then proceed to answer the user's message using the newly selected model.

Tier Routing Table

| Tier | Model | Why | |----------|-----------------------------------|----------------------------------------| | simple | ollama/dengcao/Qwen3-32B:Q5_K_M | Free local model, handles basic tasks | | moderate | anthropic/claude-sonnet-4-5 | Balanced cost and capability | | coding | openai-codex/gpt-5.2 | Optimized for code generation/debugging| | complex | anthropic/claude-opus-4-5 | Maximum reasoning power |

Discussion

0/2000
Loading comments...

Health Signals

MaintenanceCommitted 2mo ago
Active
AdoptionUnder 100 stars
3 ★ · Niche
DocsREADME + description
Well-documented

GitHub Signals

Stars3
Issues0
Updated2mo ago
View on GitHub
MIT License

My Fox Den

Community Rating

Sign in to rate this booster

Works With

Claude Code