36 boosters for "engineering" — AI-graded, open source, ready to install
Heuristic scoring (no AI key configured).
Get Shit Done (GSD) is a meta-prompting and context engineering system that helps developers coordinate AI-assisted workflows across multiple platforms while preventing quality degradation from context window bloat. Developers building with Claude, OpenCode, Gemini, and Codex benefit from its structured Plan→Execute→Verify→Complete pipeline.
A system prompt for Claude Code that enforces defensive security practices and provides CLI guidance, designed to help developers safely use Claude for software engineering tasks while preventing misuse.
A structured debugging methodology that guides developers through systematic root-cause analysis before implementing fixes, compatible with Claude Code and Cursor. Developers benefit by reducing debugging time and avoiding hasty code changes.
Packmind is an MCP server that captures and enforces your organization's technical decisions and best practices across AI coding tools like Claude Code and Cursor. It helps engineering teams maintain consistency and governance when using AI-powered development assistants.
PromptPal is a prompt management system booster designed to guide AI pair programming for backend engineering projects, specifically for Go/Golang development with REST and GraphQL APIs. Backend developers working on the PromptPal project itself will benefit from structured guidance on architecture, ORM patterns, and API design.
spec-gen automates the reverse-engineering of codebases into structured OpenSpec specifications through LLM-powered static analysis, enabling developers to maintain accurate, living documentation that reflects actual code behavior rather than intent.
Cocode is a Windsurf rules booster that provides guidelines for writing and editing Pipelex pipelines, covering file naming, structure, and best practices for code repository analysis workflows.
Cocode provides GitHub Copilot users with comprehensive coding standards and best practices guidance, covering type hints, Pydantic models, documentation, and error handling to enforce consistent code quality across teams.
Autarch Agent System is a multi-agent architecture that orchestrates AI workflows across scoping, research, planning, execution, and review stages, enabling developers to leverage AI for structured, traceable code generation and codebase analysis at scale.
Claude Kit Engineering helps developers add reference files and scripts to Claude skills, streamlining skill extension and documentation management. It benefits developers building modular AI-assisted coding workflows.
Opendidac Cursor Rules is a specialized prompt that transforms Cursor into a senior fullstack developer assistant optimized for building an educational platform with diverse question types, code execution environments, and real-time evaluation tracking. It benefits educators and developers building sophisticated assessment and training systems.
claudekit-engineering intelligently routes code issues to specialized fix commands based on issue type (type errors, UI/UX, CI/CD, tests), helping developers quickly resolve problems using the right tool. Developers working in Claude Code benefit from automated triage and targeted fixes.
TaskFlow Platform cursor rules enable multi-tenant human-agent task orchestration with support for Claude, Gemini, and other LLMs through structured prompt engineering. Developers building agent-based systems, multi-tenant applications, and complex task workflows benefit from pre-configured rules that standardize agent behavior across platforms.
A specialized agent for Node.js backend development with pure JavaScript, focusing on modern ES2024 patterns, async optimization, and runtime performance. Ideal for developers building high-performance APIs and services without TypeScript.
Reverse engineering workflow
A system prompt that instructs Claude to evaluate employee self-reviews on six writing-quality dimensions (outcomes, specificity, clarity, conciseness, ownership, collaboration) and return structured JSON scores with rationales and suggestions. Useful for HR and engineering leaders who need to assess self-review quality at scale.
A comprehensive Copilot prompt that establishes backend engineering standards and tech stack guidance (Python 3.12, MongoDB, PostgreSQL, Falcon, pytest) for consistent team development. Best suited for teams adopting this specific stack who want to enforce coding patterns and best practices.
A governance skill that automates code review and standards enforcement through delegated agent execution, delivering 70-90% token savings for engineering teams managing code quality at scale.
Gearhead provides Laravel-specific copilot instructions for a Software Engineering 1 & Application Development project, defining ecosystem package versions, coding conventions, and architectural guidelines to maintain consistency across a Laravel + Inertia + React stack.
An ML engineer agent that handles end-to-end production ML workflows including model serving, feature engineering, A/B testing, and monitoring for TensorFlow/PyTorch deployments. Ideal for teams building scalable ML systems who need guidance on MLOps best practices and production readiness.
A versatile AI agent that provides expert guidance across data engineering, machine learning, MLOps, and generative AI workflows—ideal for developers building production ML systems, data pipelines, and AI applications.
A coding agent designed for enterprise teams managing long-running projects, enabling incremental feature development with state management across multiple sessions and context windows. Ideal for DevOps engineers, engineering managers, and development teams working within compliance-heavy environments (HIPAA, SOC2).