53 boosters for "lua" — open source, verified from GitHub, ready to install
Test Results Analyzer is an AI agent that transforms raw test data into actionable quality insights through comprehensive metrics analysis and strategic reporting. QA engineers, test managers, and development teams use it to accelerate test result evaluation and drive continuous improvement.
Tool Evaluator is an AI agent that systematically assesses and recommends business tools and software platforms to help teams optimize productivity and technology ROI. Teams evaluating new tools, CTOs building tech stacks, and productivity managers benefit from its comparative analysis and adoption guidance.
Corporate Training Designer is an AI agent that helps enterprises design and optimize training programs through needs analysis, instructional design, and effectiveness evaluation. HR leaders, L&D professionals, and training managers use it to create behavior-change-focused curricula and leadership development initiatives.
A systematic paid media auditor agent that evaluates Google Ads, Microsoft Ads, and Meta accounts across 200+ checkpoints to identify inefficiencies, optimization gaps, and cost-saving opportunities. Ideal for marketing teams, agencies, and businesses seeking data-driven insights into ad spend performance.
A UX-focused design critique skill that evaluates interface effectiveness across visual hierarchy, information architecture, and emotional resonance, providing actionable feedback. Useful for designers and developers seeking structured design feedback within Claude Code.
Train object detection, image classification, and SAM/SAM2 segmentation models on managed cloud GPUs. No local GPU setup required—results are automatically saved to the Hugging Face Hub. Use this skill when users want to: Helper scripts use PEP 723 inline dependencies. Run them with :
This skill is for running evaluations against models on the Hugging Face Hub on local hardware. It does not cover: If the user wants to run the same eval remotely on Hugging Face Jobs, hand off to the skill and pass it one of the local scripts in this skill.
"name": "huggingface-skills", "description": "Agent Skills for AI/ML tasks including dataset creation, model training, evaluation, and research paper publishing on Hugging Face Hub", "name": "Hugging Face"
This skill automates the process of adding, extracting, and managing evaluation results in Hugging Face model cards, supporting multiple data sources including Artificial Analysis API and custom evaluations with vLLM/lighteval. It's valuable for ML practitioners and model maintainers who need to track and display model performance metrics.
Comprehensive quality audit system for Claude Code agents, skills, and commands. Provides quantitative scoring, comparative analysis, and production readiness grading based on industry best practices. The 16-criteria framework is derived from: 1. Claude Code Best Practices (Ultimate Guide line 4921:
A system prompt that guides LLMs to analyze Factorio game implementations and generate detailed natural language plans for achieving objectives. Useful for developers creating AI-driven game planning systems or educational tools.
Mudlet is a cross-platform MUD client built with Qt6 and C++20, providing scripting capabilities in Lua 5.1. The project emphasizes "powerful simplicity" - clean interface with deep customization options. All files should end with a newline character at the end of the file. In general: write modern
Use this agent to review existing code, audit plans, evaluate product requirements, or get architectural guidance that balances pragmatism, user experience, and security. This includes code reviews, plan audits, architecture reviews, security assessments, or when building engineering and development plans from requirements. Use proactively after significant code changes or before merging.
Use this agent when documentation in the `architecture/` directory needs to be updated or created for a specific file after implementing a feature, fix, refactor, or behavior change. Launch one instance of this agent per file that needs updating. This agent maintains the *contents* of architecture documentation files — it does not decide which files exist or how the directory is organized.\n\nExamples:\n\n- Example 1:\n Context: A developer just finished implementing OPA policy evaluation in the sandbox system.\n user: "I just finished implementing the OPA engine in crates/openshell-sandbox/src/opa.rs. Update architecture/sandbox.md to reflect the new policy evaluation flow."\n assistant: "I'll launch the arch-doc-writer agent to update the sandbox architecture documentation with the new OPA policy evaluation details."\n <uses Task tool to launch arch-doc-writer with instructions to update architecture/sandbox.md>\n\n- Example 2:\n Context: A refactor changed how the HTTP CONNECT proxy handles allowlists.\n user: "The proxy allowlist logic was refactored. Please update architecture/proxy.md."\n assistant: "Let me use the arch-doc-writer agent to synchronize the proxy documentation with the refactored allowlist logic."\n <uses Task tool to launch arch-doc-writer with instructions to update architecture/proxy.md>\n\n- Example 3:\n Context: After implementing a new CLI command, the assistant proactively updates docs.\n user: "Add a --rego-policy flag to the CLI."\n assistant: "Here is the implementation of the --rego-policy flag."\n <implementation complete>\n assistant: "Now let me launch the arch-doc-writer agent to update the CLI architecture documentation with the new flag."\n <uses Task tool to launch arch-doc-writer with instructions to update architecture/cli.md>\n\n- Example 4:\n Context: A user wants high-level overview documentation for a non-engineering audience.\n user: "Update architecture/overview.md with a non-engineer-friendly explanation of the sandbox system."\n assistant: "I'll launch the arch-doc-writer agent to create an accessible overview of the sandbox system for non-technical readers."\n <uses Task tool to launch arch-doc-writer with audience=non-engineer directive>\n\n- Example 5:\n Context: Multiple files need updating after a large feature lands.\n user: "I just landed the network namespace isolation feature. Update architecture/sandbox.md and architecture/networking.md."\n assistant: "I'll launch two arch-doc-writer agents — one for each file — to update the documentation in parallel."\n <uses Task tool to launch arch-doc-writer for architecture/sandbox.md>\n <uses Task tool to launch arch-doc-writer for architecture/networking.md>
"name": "research-companion", "description": "Strategic research thinking agents — idea evaluation, project triage, and structured brainstorming inspired by Carlini's research methodology", "name": "Andre Huang",
This skill provides a philosophical framework and analytical methods for evaluating whether end users can "know" what value they can achieve through a product. It guides analysis from a value discovery perspective, rather than providing checklists. End users adopt products when they know what value
Đóng vai Skill Architect — phỏng vấn thông minh để trích xuất quy trình từ đầu người dùng, sinh AI Skill hoàn chỉnh, rồi test và cải thiện liên tục cho đến khi đạt chất lượng production. Người dùng KHÔNG CẦN biết skill là gì.
"name": "orchestrator-supaconductor", "description": "Conductor v3 — Multi-agent orchestration with Evaluate-Loop, parallel execution, Board of Directors, and bundled SupaConductor skills for Claude Code", "orchestrator-supaconductor",
A systematic debugging skill for web applications that leverages Chrome DevTools and Playwright MCPs to troubleshoot APIs, authentication flows, and UI issues. Developers working with web apps will benefit from its structured debugging approach and intelligent context discovery.
"name": "prism-mcp-server", "mcpName": "io.github.dcostenco/prism-mcp", "description": "The Mind Palace for AI Agents — persistent memory (SQLite/Supabase), behavioral learning & IDE rules sync, multimodal VLM image captioning, pluggable LLM providers (OpenAI/Anthropic/Gemini/Ollama), OpenTelemetry
"name": "double-shot-latte", "description": "Automatically evaluates whether Claude should continue working instead of stopping prematurely using Claude-judged decision making", "url": "https://github.com/anthropics"
Use AskUserQuestion to ask the buyer: Tell the user the version was updated, then re-read the EVALUATION.md file from the updated directory and proceed with the skill. After the preamble, read the full evaluation methodology:
"name": "digital-marketing-pro", "description": "Plan, execute, and measure digital marketing across all channels. 25 specialist agents handle strategy, SEO, paid ads, content, email, social, PR, analytics, CRO, and agency operations — with brand voice enforcement, quality evaluation, multilingual s
"description": "Smart command safety filter for Claude Code — parses shell pipelines and evaluates per-command safety rules to auto-approve safe commands and block dangerous ones",