53 boosters for "lua" — open source, verified from GitHub, ready to install
An agent designed to evaluate other agents and tasks, with library-first constraints and multi-tool integration across Claude platforms. Useful for teams building quality assurance workflows into their Claude-based systems.
AgentAsJudge is an agentic evaluation framework that enables AI to systematically assess and compare the quality of multiple-choice questions across educational value, clarity, and answerability. It benefits educators, content creators, and assessment teams looking to automate quality control of exam and quiz questions.
AgentAsJudge is an agentic evaluation framework that enables AI systems to critically review educational introductions by validating them against specified quality metrics and providing constructive feedback. It benefits educators, instructional designers, and developers building AI-assisted learning platforms who need reliable, fair assessment of educational content.
"version": "5.10.0", "description": "Memory → Evaluation → Credential → Access Control for AI agents. Persistent memory with W3C Verifiable Credentials, capability-based access control, drift detection, and FSRS-6 spaced repetition.", "name": "kobie3717",
"name": "open-academic-paper-machine", "description": "Open Academic Paper Machine — Autonomous academic paper production system with idea evaluation gate and paper-vs-code audit. NEW in v6.4: /audit-paper command and audit-engine skill — static audit of a paper's empirical claims (datasets, models,
A development guide for extending PatientHub with new patient simulation agents, enabling researchers and developers to implement custom behavioral models for healthcare simulations.
"name": "cre-skills", "description": "112 institutional-grade CRE skills covering ~97% of commercial real estate workflow steps. Deal screening, underwriting, structuring, due diligence, capital markets, market research, asset management, leasing, investor relations, development, disposition, sourci
Alexi is an expropriation appraisal specialist that provides expert-level valuation analysis using before/after methods, comparable sales, and severance damage assessment for legal and real estate professionals. Real estate appraisers, lawyers, and property valuators benefit from delegating complex expropriation cases to this specialized agent.
Provides a standardized workflow for creating and maintaining unit tests for Jass modules in War3Lib using Zinc, covering file structure, test generation, assertion patterns, and validation. Essential for developers building or refactoring Jass libraries who need reproducible, maintainable test suites.
The architect agent automates system design for new projects and major refactoring efforts, helping teams create scalable architectures with documented trade-offs. Ideal for engineering teams starting greenfield projects or evaluating architectural changes.
ArmBench-LLM is a system prompt framework for evaluating large language models on Armenian language tasks through structured multiple-choice questions. It's designed for developers and AI researchers who need standardized benchmarking tools across popular coding assistants and chat platforms.
ArmBench-LLM is a system prompt for benchmarking large language models using Armenian character-to-numeric matching tasks. It's designed for developers evaluating LLM performance across multiple coding platforms.
Tool-evaluator is an agent that rapidly assesses development tools, frameworks, and services through structured benchmarking and comparative analysis to support informed technology adoption decisions. It benefits engineering teams and tech leads evaluating new solutions aligned with studio goals.
A debate judge agent that objectively evaluates arguments using zero-sum scoring across Toulmin structure, evidence strength, and logical rigor. Ideal for researchers, educators, and developers building computational debate systems.
skill-auditor is an expert reviewer that evaluates SKILL.md files against Claude Code Skills best practices, helping developers ensure their skills meet structural and effectiveness standards. It's essential for skill creators and maintainers who want to validate compliance before publishing.
PrismBench enables developers to create specialized LLM agents through YAML configuration for systematic evaluation of model capabilities using Monte Carlo Tree Search. Useful for ML engineers, researchers, and teams building production LLM systems who need comprehensive benchmarking and evaluation frameworks.
PrismBench enables developers to create specialized LLM agents through YAML configuration for comprehensive benchmarking and evaluation of language model capabilities. Teams building AI evaluation systems and ML testing pipelines benefit from its systematic Monte Carlo Tree Search approach and containerized deployment.
A Chief Technology Officer agent that guides enterprise technology strategy decisions, including investment evaluation, technical vision setting, and architectural planning. Ideal for organizations needing structured CTO-level guidance on technology roadmaps and innovation initiatives.
This skill enables developers to create cryptographically signed, immutable constitutions for AI tool-use governance in OpenClaw, with Ed25519 signing, GitTruth attestation, and policy evaluation artifacts. It's designed for teams implementing constitutional governance frameworks for AI agents.
SkillGuard is a security reviewer for Claude/Cursor Skills that detects prompt injection, tool injection, data exfiltration, and unsafe automation risks. It's essential for developers and organizations installing or developing AI skills to ensure safe, policy-compliant code execution.
SDET is a skill that enables AI assistants to design and build comprehensive test automation infrastructure, including end-to-end tests, coverage analysis, and testing strategy. It benefits developers who need robust automated testing frameworks and QA engineers seeking to identify and close testing gaps.
A Cursor-specific ruleset that enforces Python development standards using uv for package management and Pydantic v2, designed to ensure consistent tooling practices across AI-assisted coding workflows.
A specialized Copilot prompt that configures an AI agent as an expert assistant for building and deploying ASP.NET Core web APIs and Blazor WebAssembly apps to Google Cloud, with integrated validation and iterative problem-solving capabilities.
Spec-Judge evaluates and selects the best versions of requirement, design, and task specification documents based on comprehensive criteria like completeness, clarity, feasibility, and innovation. It helps development teams streamline spec development workflows by providing systematic evaluation and comparison of document versions.