172 boosters for "llm" — AI-graded, open source, ready to install
This MCP server bridges n8n workflow automation with Claude by providing programmatic access to workflows, executions, and credentials through the Model Context Protocol. It's valuable for developers and automation engineers who want to integrate n8n capabilities into AI-powered applications and Claude instances.
Dinox MCP Server enables Claude to understand images through advanced object detection, localization, and captioning by connecting to the DINO-X vision model. Developers building multimodal AI applications benefit from seamless integration of real-world visual perception into LLM workflows.
Heuristic scoring (no AI key configured).
This booster provides best practice guidelines for Docker, npx, uv, and package publishing workflows, intended to standardize development processes in Windsurf environments. It benefits developers seeking consistent build and deployment practices across multiple tool ecosystems.
VT.ai provides Copilot-specific coding instructions for a multimodal AI chat application, establishing standards for Python development including naming conventions, style guides, and testing practices. Developers building AI-powered features with language models will benefit from these standardized guidelines.
Native Claude Agent SDK integration for AgentPool that enables file operations, terminal access, and advanced code editing with lower latency and tighter integration than bridge alternatives. Developers building AI agents on Claude Code or Claude Desktop benefit from direct SDK access and streamlined tooling.
A Cursor IDE rule set that enforces consistent Rust coding standards, project structure, and architectural patterns for the LLML library. Developers working on LLML in Cursor will benefit from automated guidance on module organization, naming conventions, and API design.
AniList MCP Server enables Claude to access AniList's anime and manga database through the Model Context Protocol, allowing developers and anime enthusiasts to integrate rich AniList data into AI-powered applications.
Concurrent Sub-Agents is an agentic terminal application that enables multi-agent workflows by automatically executing independent sub-agent tool calls in parallel, reducing latency without requiring explicit concurrency logic from the LLM. Developers building complex AI agent systems on Claude will benefit from faster, more efficient agent coordination.
An MCP server that enables Claude to automate Firefox DevTools for browser testing, debugging, and web automation tasks. Useful for developers who need AI-assisted browser automation and testing workflows.
Search arXiv and ACL Anthology, retrieve citations and references, and browse web sources to accel…
Widget MCP is an MCP server that enables LLMs to render interactive UI widgets within Claude Desktop and Claude Code environments. It's useful for developers building enhanced AI experiences with custom interfaces.
A comprehensive guide for developers adding new agents to Errata, a writing application. Useful for backend engineers extending LLM-powered features like character optimization and prose analysis.
An MCP Server that integrates Paddle Billing with AI assistants, enabling developers to query and manage billing operations directly through Claude and Cursor. Ideal for SaaS developers and billing teams who need AI-powered access to Paddle's billing platform.
This MCP server enables AI coding agents to analyze test coverage from LCOV files, making them coverage-aware while optimizing token usage. It's valuable for developers using Claude Desktop or Code who want to improve code quality by integrating coverage insights into their AI-assisted workflows.
A Model Context Protocol server for interacting with Nostr, enabling posting notes and more
A structured prompt booster for Windsurf that establishes best practices for agentic workflows, covering instruction handling, task analysis, implementation verification, and a complete tech stack (Next.js, Payload CMS, TypeScript). It helps developers build reliable AI-powered applications with clear guidelines and quality standards.
Enables AI agents to parallelize independent development tasks using git worktrees, splitting work across multiple concurrent tracks while maintaining context continuity. Ideal for teams coordinating multiple Claude Code sessions on non-overlapping features.
spec-gen automates the reverse-engineering of codebases into structured OpenSpec specifications through LLM-powered static analysis, enabling developers to maintain accurate, living documentation that reflects actual code behavior rather than intent.
A Claude Code skill that executes simple tasks directly without agent planning overhead, ideal for quick bug fixes and single-file changes. Developers benefit from faster execution for straightforward coding work.
A comprehensive tutorial chapter teaching developers how to build and deploy intelligent AI agents with tool use and automation capabilities in AnythingLLM. Ideal for engineers looking to add autonomous workflows and function-calling to their self-hosted RAG systems.
RagCode MCP is a semantic code navigation tool that integrates RAG-powered code search into Windsurf and other IDEs, enabling developers to intelligently query and understand multi-language codebases using local LLMs. It's ideal for developers working with Laravel, Go, Python, and PHP who need fast, context-aware code exploration without leaving their IDE.
Teaches the functional options pattern for designing flexible Go APIs with optional configuration, helping developers build clean, extensible constructors that scale as requirements grow.