106 boosters for "lua" — AI-graded, open source, ready to install
ArmBench-LLM is a system prompt framework for evaluating large language models on Armenian language tasks through structured multiple-choice questions. It's designed for developers and AI researchers who need standardized benchmarking tools across popular coding assistants and chat platforms.
The plan-generator transforms high-level product requirements into executable project blueprints (genesis.xml files) with structured task DAGs and agent assignments. It's invaluable for cofounders and product teams who need to bridge strategic vision with concrete execution plans.
A debate judge agent that objectively evaluates arguments using zero-sum scoring across Toulmin structure, evidence strength, and logical rigor. Ideal for researchers, educators, and developers building computational debate systems.
TDD Command enforces test-driven development workflows by guiding developers through the RED-GREEN-REFACTOR cycle with 80%+ coverage targets. Developers prioritizing code quality and test coverage will find this particularly valuable.
Tool-evaluator is an agent that rapidly assesses development tools, frameworks, and services through structured benchmarking and comparative analysis to support informed technology adoption decisions. It benefits engineering teams and tech leads evaluating new solutions aligned with studio goals.
This booster helps developers maintain accurate, consistent documentation across README, ARCHITECTURE, and other files by identifying omissions and inconsistencies against the actual implementation. It's valuable for open-source maintainers and teams that prioritize documentation quality.
A system prompt that instructs Claude to evaluate employee self-reviews on six writing-quality dimensions (outcomes, specificity, clarity, conciseness, ownership, collaboration) and return structured JSON scores with rationales and suggestions. Useful for HR and engineering leaders who need to assess self-review quality at scale.
A system prompt that configures an AI agent to evaluate text by delegating to a tool called `evaluate_review_text`, then summarize results and update graph state. Best suited for developers building evaluation workflows in Claude-based IDEs.
skill-auditor is an expert reviewer that evaluates SKILL.md files against Claude Code Skills best practices, helping developers ensure their skills meet structural and effectiveness standards. It's essential for skill creators and maintainers who want to validate compliance before publishing.
This Cursor rule enforces Bun as the default runtime and package manager, automatically guiding developers to use Bun commands instead of Node.js, npm, or other tools. Developers actively using or migrating to Bun will find this invaluable for consistent toolchain decisions.
PrismBench enables developers to create specialized LLM agents through YAML configuration for comprehensive benchmarking and evaluation of language model capabilities. Teams building AI evaluation systems and ML testing pipelines benefit from its systematic Monte Carlo Tree Search approach and containerized deployment.
PrismBench enables developers to create specialized LLM agents through YAML configuration for systematic evaluation of model capabilities using Monte Carlo Tree Search. Useful for ML engineers, researchers, and teams building production LLM systems who need comprehensive benchmarking and evaluation frameworks.
A system prompt that guides LLMs to act as rigorous sustainability and risk analysts for Polkadot governance proposals, evaluating treasury allocations through investment frameworks, fiscal precedent, and accountability mechanisms.
An MCP server that provides structured access to adversarial tactics and cyber attack techniques for security research, penetration testing, and AI safety evaluation. Useful for security professionals, red teamers, and AI safety researchers studying attack vectors.
A complete Coda API client with MCP integration that enables Claude to interact with Coda documents and workspaces, automatically handling large content through chunking. It's valuable for developers and teams using Coda who want to automate workflows and integrate AI capabilities.
SkillGuard is a security reviewer for Claude/Cursor Skills that detects prompt injection, tool injection, data exfiltration, and unsafe automation risks. It's essential for developers and organizations installing or developing AI skills to ensure safe, policy-compliant code execution.
This MCP server enables Claude to generate interactive ECharts visualizations by converting data into chart configurations. It's valuable for developers and data analysts who want to create dynamic, publication-quality charts directly within Claude conversations.
A Chief Technology Officer agent that guides enterprise technology strategy decisions, including investment evaluation, technical vision setting, and architectural planning. Ideal for organizations needing structured CTO-level guidance on technology roadmaps and innovation initiatives.
This skill enables developers to create cryptographically signed, immutable constitutions for AI tool-use governance in OpenClaw, with Ed25519 signing, GitTruth attestation, and policy evaluation artifacts. It's designed for teams implementing constitutional governance frameworks for AI agents.
AgentMail provides AI agents with dedicated email inbox functionality similar to Gmail, enabling agents to send, receive, and manage emails programmatically. This is valuable for developers building autonomous AI systems that need to communicate via email or integrate with email-based workflows.
A specialized Copilot prompt that configures an AI agent as an expert assistant for building and deploying ASP.NET Core web APIs and Blazor WebAssembly apps to Google Cloud, with integrated validation and iterative problem-solving capabilities.
This MCP Server enables Claude to manage Argo Workflows directly through the Argo Server API, allowing DevOps engineers and platform teams to orchestrate and monitor Kubernetes workflows via AI agents. It bridges AI capabilities with enterprise workflow automation, making it valuable for teams running Argo on Kubernetes.
SDET is a skill that enables AI assistants to design and build comprehensive test automation infrastructure, including end-to-end tests, coverage analysis, and testing strategy. It benefits developers who need robust automated testing frameworks and QA engineers seeking to identify and close testing gaps.
A risk assessment agent that evaluates the global risk of code modifications by analyzing security, quality, and impact metrics from AgentDB before merge decisions. Developers use it to quantify change risk and make informed merge decisions.