440 boosters for "mod" — open source, verified from GitHub, ready to install
Heuristic scoring (no AI key configured).
An MCP server that enables Claude to execute MySQL database operations directly, allowing AI assistants to query, insert, and manage database records. Useful for developers integrating Claude with MySQL-backed applications and data workflows.
A system integration booster that merges and validates outputs from multiple AI modes into production-ready systems with consistency checks and modular architecture. Useful for developers building complex, multi-component applications.
Query Google's AI Search mode to retrieve comprehensive, source-grounded answers from across the web. Trigger this skill when the user: 1. Include Current Year (2026) for up-to-date results
An MCP server that integrates Claude with the Hevy fitness tracking app, enabling AI-powered workout analysis, progress tracking, and fitness insights. Ideal for fitness enthusiasts and developers building AI-assisted health applications.
Matchms is a Python library for mass spectrometry data processing, enabling researchers to load, standardize, and analyze spectral data from multiple formats (mzML, MGF, MSP) with similarity matching for metabolomics and compound identification workflows.
<summary >🌐 Language</summary> <div align="center"> <a href="https://openaitx.github.io/view.html?user=wshobson&project=agents&lang=en">English</a>
This MCP server bridges AI assistants with Linear project management via GraphQL, enabling Claude and Cursor users to query and manage Linear projects directly through natural language. It's essential for teams using Linear who want AI-assisted workflow automation and issue management.
"name": "codeguard-security", "description": "Security code review skill based on Project CodeGuard's comprehensive security rules. Helps AI coding agents write secure code and prevent common vulnerabilities.", "name": "Project CodeGuard",
This MCP server bridges n8n workflow automation with Claude by providing programmatic access to workflows, executions, and credentials through the Model Context Protocol. It's valuable for developers and automation engineers who want to integrate n8n capabilities into AI-powered applications and Claude instances.
Leta is a semantic code navigation booster that replaces manual file searching and ripgrep with fast LSP-powered commands for exploring code structure, finding symbols, and understanding dependencies. It benefits developers working with unfamiliar codebases, refactoring projects, and fixing type errors by dramatically speeding up code comprehension tasks.
A practical guide for deploying serverless Python applications on Modal, enabling developers to run GPU-accelerated AI/ML workloads, web APIs, and batch jobs with minimal infrastructure configuration.
Provides structured guidance for working with CUE schema files and parsing flows in invkfile, invkmod, and config schemas. Developers maintaining or extending CUE-based configuration validation will benefit from this reference.
"name": "matryoshka-rlm", "version": "0.2.17", "description": "Recursive Language Model - Process documents larger than LLM context windows",
A practical guide to Supervised Fine-Tuning using SFTTrainer and Unsloth optimizations, enabling developers to efficiently adapt pre-trained LLMs for instruction-following with 2x faster training. Ideal for ML engineers building custom instruction-tuned models and reasoning systems.
Pi Integration enables developers to route coding tasks to Pi (an alternative AI agent) for comparison testing and benchmarking against Claude Code. It's useful for teams evaluating multiple AI coding assistants and want side-by-side performance comparison.
"name": "minimax-mcp-js", "version": "0.0.17", "description": "Official MiniMax Model Context Protocol (MCP) JavaScript implementation that provides seamless integration with MiniMax's powerful AI capabilities including image generation, video generation, text-to-speech, and voice cloning APIs.",
This skill should be used when users want to train or fine-tune language models using TRL (Transformer Reinforcement Learning) on Hugging Face Jobs infrastructure. Covers SFT, DPO, GRPO and reward modeling training methods, plus GGUF conversion for local deployment. Includes guidance on the TRL Jobs package, UV scripts with PEP 723 format, dataset preparation and validation, hardware selection, cost estimation, Trackio monitoring, Hub authentication, and model persistence. Should be invoked for tasks involving cloud GPU training, GGUF conversion, or when users mention training on Hugging Face Jobs without local GPU setup.
A Prisma/PostgreSQL database management skill that helps developers modify Motivia's schema, add tables/fields/relations, and execute migrations. Essential for anyone working on the Motivia project's data layer.
Dinox MCP Server enables Claude to understand images through advanced object detection, localization, and captioning by connecting to the DINO-X vision model. Developers building multimodal AI applications benefit from seamless integration of real-world visual perception into LLM workflows.
Automates GitHub issue triage and fixing by fetching issues, spawning AI sub-agents to implement solutions, opening PRs, and handling review feedback—ideal for maintainers and teams managing high-volume bug backlogs.
"name": "prism-mcp-server", "mcpName": "io.github.dcostenco/prism-mcp", "description": "The Mind Palace for AI Agents — persistent memory (SQLite/Supabase), behavioral learning & IDE rules sync, multimodal VLM image captioning, pluggable LLM providers (OpenAI/Anthropic/Gemini/Ollama), OpenTelemetry
"name": "webflow-mcp-server", "bin": "dist/index.js", "start": "concurrently \"npm run dev:local\" \"npm run inspector:local\"",