37 boosters for "llms" — AI-graded, open source, ready to install
A comprehensive guide to creating and managing autonomous agents within the CrewAI framework, enabling developers to build specialized AI agents with defined roles, capabilities, and collaboration features. Ideal for AI engineers and developers building multi-agent systems.
A system prompt that guides LLMs to analyze Factorio game implementations and generate detailed natural language plans for achieving objectives. Useful for developers creating AI-driven game planning systems or educational tools.
A PHP SDK for building AI agents with structured outputs and multi-agent orchestration, enabling developers to decompose complex tasks into specialized subagents with isolated contexts and independent execution.
Dinox MCP Server enables Claude to understand images through advanced object detection, localization, and captioning by connecting to the DINO-X vision model. Developers building multimodal AI applications benefit from seamless integration of real-world visual perception into LLM workflows.
VT.ai provides Copilot-specific coding instructions for a multimodal AI chat application, establishing standards for Python development including naming conventions, style guides, and testing practices. Developers building AI-powered features with language models will benefit from these standardized guidelines.
Medical MCP Server provides AI assistants with access to authoritative medical data from FDA, WHO, PubMed, and RxNorm APIs, enabling developers to build healthcare applications with reliable clinical information.
Widget MCP is an MCP server that enables LLMs to render interactive UI widgets within Claude Desktop and Claude Code environments. It's useful for developers building enhanced AI experiences with custom interfaces.
RagCode MCP is a semantic code navigation tool that integrates RAG-powered code search into Windsurf and other IDEs, enabling developers to intelligently query and understand multi-language codebases using local LLMs. It's ideal for developers working with Laravel, Go, Python, and PHP who need fast, context-aware code exploration without leaving their IDE.
This MCP server enables Claude and other LLMs to authenticate with AWS via SSO and securely execute AWS CLI commands using temporary credentials. It's valuable for developers and DevOps teams who want AI-assisted AWS resource management without embedding long-lived credentials.
This MCP server dynamically exposes Swagger/OpenAPI endpoints as callable tools for Claude and other LLMs, enabling AI agents to seamlessly integrate with REST APIs without manual tool definition. Developers building LLM applications that need to interact with documented APIs benefit most from this adapter.
Heuristic scoring (no AI key configured).
THOR Skills is a router skill for LLMs that guides users through THOR security scanner operations—from running scans and analyzing logs to troubleshooting and writing plugins. It's valuable for security professionals, forensic analysts, and DevOps teams working with THOR malware scanning on endpoints.
TaskFlow Platform cursor rules enable multi-tenant human-agent task orchestration with support for Claude, Gemini, and other LLMs through structured prompt engineering. Developers building agent-based systems, multi-tenant applications, and complex task workflows benefit from pre-configured rules that standardize agent behavior across platforms.
Start/stop Docker test databases
A TypeScript MCP server booster for integrating the ReviewWebsite API with Windsurf, providing CLI tools, validation schemas, and structured patterns for AI-assisted development. Ideal for developers building AI-powered review applications or extending Windsurf with ReviewWebsite capabilities.
ArmBench-LLM is a system prompt for benchmarking large language models using Armenian character-to-numeric matching tasks. It's designed for developers evaluating LLM performance across multiple coding platforms.
ArmBench-LLM is a system prompt framework for evaluating large language models on Armenian language tasks through structured multiple-choice questions. It's designed for developers and AI researchers who need standardized benchmarking tools across popular coding assistants and chat platforms.
Dreamfactory MCP Server enables Claude to access governed APIs across any enterprise data source through DreamFactory's self-hosted platform, benefiting enterprise developers and LLM integrators needing secure, standardized data access.
DermaSwarm is a multi-agent AI system that assists dermatologists in diagnosing and treating skin conditions by cross-referencing peer-reviewed research and generating structured treatment plans. It's designed for healthcare professionals seeking evidence-based dermatological decision support.
MCP Server to convert Figma Design to pseudo-code like structure - simple yet efficient for LLMs to understand design context.
The MCP TypeScript SDK enables developers to build and integrate Model Context Protocol servers and clients, allowing applications to provide standardized context to LLMs. It's essential for developers building AI-powered applications that need to expose resources, tools, and prompts in a protocol-compliant way.
Node Code Sandbox MCP enables AI agents and LLMs to safely execute JavaScript code, install npm packages, and test implementations in real-time. Developers building coding assistants, automated testing tools, and interactive AI applications benefit from this secure, MCP-compliant execution environment.
A system prompt that guides LLMs to act as rigorous sustainability and risk analysts for Polkadot governance proposals, evaluating treasury allocations through investment frameworks, fiscal precedent, and accountability mechanisms.
gllm is a system prompt that transforms LLMs into code-first orchestrators for efficient task processing, enabling developers to handle large files and complex workflows through programmatic verification rather than context-heavy text generation.