116 boosters for "lm" — AI-graded, open source, ready to install
A debate judge agent that objectively evaluates arguments using zero-sum scoring across Toulmin structure, evidence strength, and logical rigor. Ideal for researchers, educators, and developers building computational debate systems.
ElixirScope is a specialized Cursor rules configuration that enhances Elixir/BEAM debugging by providing execution analysis guidance, testing protocols, and code mapping standards. It benefits Elixir developers using Cursor who need structured debugging workflows and LLM-assisted runtime visualization.
This MCP Server converts Figma designs into pseudo-code structures that LLMs can easily understand, enabling seamless design-to-code workflows for developers and designers building prototypes and applications.
Automates synchronization of custom Ollama models across distributed Vera AI clusters using YAML configuration and templating. Benefits DevOps engineers and teams managing multi-instance LLM deployments who need centralized model management.
A skill that enables Claude Code to query other LLMs (ChatGPT, Gemini, DeepSeek, Grok) through MCP servers by routing requests based on the first word of a query. Useful for developers who need multi-LLM comparison or fallback capabilities within Claude Code environments.
A Cursor IDE rules configuration providing comprehensive development standards and best practices for LLM application development, including project structure guidelines, coding principles, and documentation conventions.
The MCP TypeScript SDK enables developers to build and integrate Model Context Protocol servers and clients, allowing applications to provide standardized context to LLMs. It's essential for developers building AI-powered applications that need to expose resources, tools, and prompts in a protocol-compliant way.
The MCP TypeScript SDK is a comprehensive toolkit for building and integrating Model Context Protocol servers and clients in TypeScript applications. Developers building AI-powered tools, Claude integrations, or context-aware LLM applications benefit from its standardized approach to managing resources, tools, and prompts.
OpenCap Coding Standards v1.0 is a prompt-based guide for developers working on the OpenCap Stack, emphasizing test-driven development, consistency, and LLM-assisted coding for Python/FastAPI backends on Windsurf. It benefits backend developers and AI coding assistant users seeking standardized, production-ready code practices.
An MCP server that provides structured access to adversarial tactics and cyber attack techniques for security research, penetration testing, and AI safety evaluation. Useful for security professionals, red teamers, and AI safety researchers studying attack vectors.
A system prompt that guides LLMs to act as rigorous sustainability and risk analysts for Polkadot governance proposals, evaluating treasury allocations through investment frameworks, fiscal precedent, and accountability mechanisms.
PrismBench enables developers to create specialized LLM agents through YAML configuration for systematic evaluation of model capabilities using Monte Carlo Tree Search. Useful for ML engineers, researchers, and teams building production LLM systems who need comprehensive benchmarking and evaluation frameworks.
Node Code Sandbox MCP enables AI agents and LLMs to safely execute JavaScript code, install npm packages, and test implementations in real-time. Developers building coding assistants, automated testing tools, and interactive AI applications benefit from this secure, MCP-compliant execution environment.
Geist provides Cursor-integrated coding rules and best practices for Python, SQLAlchemy, logging, and server development using FastAPI and LLM models. It helps developers maintain consistent code standards and proper patterns when building AI-powered applications.
A system prompt that enables LLM-based news headline tagging by categorizing stories into People, Topic, and Geography tags with structured JSON output. Useful for developers building news processing pipelines or content classification systems.
gllm is a system prompt that transforms LLMs into code-first orchestrators for efficient task processing, enabling developers to handle large files and complex workflows through programmatic verification rather than context-heavy text generation.
Windsurf Rules for OpenCap Stack provides coding standards and LLM-aligned prompts to guide developers on test-driven, production-ready Python/FastAPI backend development. Developers working on OpenCap Stack projects and AI coding assistant users benefit from these standardized guidelines.
This booster automates reconnaissance of LLM API endpoints to identify models, authentication methods, and configuration details for security testing. Red team operators and security researchers benefit from structured enumeration workflows.
A Windsurf-specific security framework for detecting and testing OWASP LLM Top 10 vulnerabilities in LLM applications, with AWS integration and CI/CD automation. Ideal for security engineers and LLM developers building production-grade applications.
A comprehensive system prompt and structural guide for building production-ready GenAI applications with clear rules, templates, and best practices across multiple AI platforms. Developers building AI agents and structured LLM applications benefit from its emphasis on proper configuration, testing, and code standards.
A comprehensive guide to creating and configuring AI agents in the AgenticAI Core SDK, enabling developers to build autonomous agents with specific roles, LLM decision-making, and reusable tools. Essential for developers building multi-agent applications.
Intent Kit provides Cursor rules for building hierarchical intent-driven Python workflows with LLMs, enabling developers to create structured, context-aware automation systems. It's useful for developers building complex AI-driven applications that require intent classification and conditional execution.
HoloLoom is a neural decision-making system prompt integrating prompt management, analytics, and Thompson Sampling optimization for LLM applications. Developers working with multi-agent systems and prompt optimization would benefit from its structured approach to prompt selection and performance tracking.
mybin is a personal collection of 65+ production-ready CLI utility scripts (Python, Bash, Ruby) for AI/LLM integration, system utilities, and development tasks that Claude Code can proactively leverage. Developers with Unix-style workflows and frequent CLI automation needs benefit from instant access to tested, reusable tools.