20 boosters for "ollama" — open source, verified from GitHub, ready to install
A specialized AI agent that automatically detects, classifies, and fixes data anomalies in production pipelines using local SLMs and semantic clustering, with zero data loss guarantee. Data engineers and platform teams benefit most when dealing with broken pipelines that can't afford downtime.
Openmemory JS is a local persistent memory store for LLM applications that enables long-term context retention across Claude Desktop, GitHub Copilot, and other AI platforms. Developers building AI agents and applications benefit from enhanced memory management without external dependencies.
"version": "9.18.0", "description": "v9.18.0 — Multi-LLM orchestration for Claude Code with Double Diamond workflows, provider routing, safety gates, and automation. Use '/octo:auto' or just describe what you need. Run /octo:setup.", "repository": "https://github.com/nyldn/claude-octopus",
An AI-powered VSCode extension that auto-generates standardized Git/SVN commit messages using multiple AI services (OpenAI, ChatGPT, Ollama, Gemini, etc.) with multi-language support and weekly report generation. Developers working with Git/SVN in VSCode benefit from faster, consistent commit workflows.
Heuristic scoring (no AI key configured).
VT.ai provides Copilot-specific coding instructions for a multimodal AI chat application, establishing standards for Python development including naming conventions, style guides, and testing practices. Developers building AI-powered features with language models will benefit from these standardized guidelines.
"id": "ac.tandem/docs-mcp", "description": "Remote MCP server for Tandem docs, install guides, SDKs, workflows, and agent setup help.", "url": "https://github.com/frumu-ai/tandem",
"name": "brainstorm-mcp", "mcpName": "io.github.spranab/brainstorm-mcp", "description": "MCP server for multi-round AI brainstorming debates across multiple models",
You are an expert Python developer working on ngx-renamer, an AI-powered document title generator for Paperless NGX. Your role is to maintain, test, and improve the codebase while following established conventions. All LLM providers inherit from and are registered in the provider registry: Always t
"description": "Beacon — semantic code search for Claude Code", "repository": "https://github.com/sagarmk/Claude-Code-Beacon-Plugin", "homepage": "https://github.com/sagarmk/Claude-Code-Beacon-Plugin",
RagCode MCP is a semantic code navigation tool that integrates RAG-powered code search into Windsurf and other IDEs, enabling developers to intelligently query and understand multi-language codebases using local LLMs. It's ideal for developers working with Laravel, Go, Python, and PHP who need fast, context-aware code exploration without leaving their IDE.
ADHARA is an LLM system prompt that enables early detection of learning friction by analyzing behavioral metrics like mouse hesitation, attention, and task performance against age-appropriate baselines. It helps educators identify students who may need additional support before academic failure occurs.
Cortex is a local-first knowledge graph that automatically watches your project files, extracts entities and relationships using LLMs, and enables natural language queries across multiple projects. It's ideal for developers who need intelligent cross-project search and knowledge management without relying on external services.
A migration guide for transitioning from deprecated hardcoded agents to a dynamic ConfigurableAgent system in a Rust-based agentic chatbot server. Developers maintaining or upgrading agent-based systems benefit from clear configuration patterns and best practices for multi-provider LLM setups.
Automates synchronization of custom Ollama models across distributed Vera AI clusters using YAML configuration and templating. Benefits DevOps engineers and teams managing multi-instance LLM deployments who need centralized model management.
Intelli-router automatically triages user messages by complexity and routes them to the most cost-effective AI model (local Ollama, Claude Sonnet, Codex, or Claude Opus). Developers building multi-model AI systems benefit from reduced costs and optimized latency.
An AI-powered code auditing tool that integrates with Claude Desktop via MCP protocol, using local Ollama models to analyze code for security vulnerabilities, performance issues, and quality concerns without sending code to external services. Developers and security teams benefit from automated, privacy-preserving code reviews with customizable analysis parameters.
Ifcli MCP Server enables developers to chat with multiple AI providers (OpenAI, DeepSeek, OpenRouter, Ollama) directly from the CLI, supporting both standard and reasoning models through a unified interface.
A JavaScript-based system for dynamically creating and modifying AI agents without recompilation, enabling local LLM developers to test and share agent configurations interactively.
A practical integration guide for enhancing OCR text extraction with visual and language LLM capabilities using local Ollama models in Caption Extractor. Developers working with document processing, image analysis, and text correction workflows benefit from this reusable agent framework.