108 boosters for "local" — open source, verified from GitHub, ready to install
A Playwright-based toolkit for testing local web applications, enabling developers to verify frontend functionality, debug UI behavior, and capture browser diagnostics. Essential for AI coding assistants building and validating web applications.
A specialized AI agent that automatically detects, classifies, and fixes data anomalies in production pipelines using local SLMs and semantic clustering, with zero data loss guarantee. Data engineers and platform teams benefit most when dealing with broken pipelines that can't afford downtime.
"email": "tobi@lutke.com" "description": "Search and retrieve documents from local markdown files.", "email": "tobi@lutke.com"
Sora enables Claude Code users to generate, remix, and manage AI videos directly through OpenAI's video API. Developers building products with video content (demos, marketing, UI mocks) benefit from integrated video generation workflows.
Create a real CLI that future Codex threads can run by command name from any working directory. This skill is for durable tools, not one-off scripts. If a short script in the current repo solves the task, write the script there instead. Name the target tool, its source, and the first real jobs it sh
Train language models using TRL (Transformer Reinforcement Learning) on fully managed Hugging Face infrastructure. No local GPU setup required—models train on cloud GPUs and results are automatically saved to the Hugging Face Hub. Use this skill when users want to: Use Unsloth () instead of standard
Run any workload on fully managed Hugging Face infrastructure. No local setup required—jobs run on cloud CPUs, GPUs, or TPUs and can persist results to the Hugging Face Hub. Use this skill when users want to: When assisting with jobs:
Provides the Hugging Face Hub CLI (`hf`) tool for downloading, uploading, and managing models, datasets, and Spaces directly from Claude Code. Essential for developers integrating Hugging Face resources into AI workflows.
This skill is for running evaluations against models on the Hugging Face Hub on local hardware. It does not cover: If the user wants to run the same eval remotely on Hugging Face Jobs, hand off to the skill and pass it one of the local scripts in this skill.
This skill enables users to run Python workloads, Docker jobs, and GPU-intensive tasks on Hugging Face's managed infrastructure without local setup. It's valuable for ML engineers, data scientists, and developers needing cloud compute for training, inference, and batch processing.
A skill for fine-tuning and training language models on Hugging Face's cloud GPU infrastructure using TRL, supporting SFT, DPO, GRPO methods and GGUF conversion for local deployment. Developers and ML engineers working with cloud-based model training benefit from this comprehensive guidance.
Enables developers to interact with Hugging Face Hub directly from Claude Code using the `hf` CLI—downloading models/datasets, uploading files, creating repositories, and managing cache without leaving the coding environment.
"description": "Develop Apps Script projects locally using clasp", "args": ["-y", "@google/clasp", "mcp"]
Automates GitHub pull request analysis by gathering diffs, comments, related issues, and local code context to provide comprehensive reviews. Developers and code reviewers benefit from faster, more thorough PR evaluations.
Do NOT check or review pull requests. Do NOT call commands. Run CodeRabbit locally against the working repository only. From the output, extract for each finding:
Local GraphRAG knowledge base backed by SQLite + MNN embeddings. Fully compatible with Android OfflineAI RAG database format. On first use, (~400 MB) is auto-downloaded into .
Openmemory JS is a local persistent memory store for LLM applications that enables long-term context retention across Claude Desktop, GitHub Copilot, and other AI platforms. Developers building AI agents and applications benefit from enhanced memory management without external dependencies.
A hands-on course teaching developers how to build AI-powered agents using Foundry Local with modern SDK patterns, function calling, and hybrid cloud designs. Benefits beginners and intermediate developers wanting to prototype agentic applications quickly on edge devices.
MCP Remote is a proxy server that enables local-only AI clients to connect to remote Model Context Protocol servers using OAuth authentication. It's essential for developers building Claude integrations that need to bridge local and remote MCP server connections securely.
A web search MCP server powered by Brave Search API, enabling Claude and other AI tools to search the web, find local businesses, and retrieve images in real-time. Useful for developers building AI applications that need current information beyond training data.
Quoroom is an experimental open-source framework for building local AI agent systems with a Queen-Workers-Quorum architecture, designed for researchers and developers exploring multi-agent AI patterns with Claude and other LLMs.
This skill provides comprehensive guidance for all aspects of Playwright test development, from writing new tests to debugging and maintaining existing test suites. Consult these references based on what you're doing:
Smart Coding MCP enhances developer productivity with AI-powered semantic code search, intelligent indexing, and hybrid search capabilities integrated into Cursor, Claude Desktop, and VS Code. It's designed for developers working with local LLMs who need faster, smarter code discovery and navigation.
Generate XML files and export to PNG/SVG/PDF/JPG locally using the native draw.io desktop app CLI. PNG, SVG, and PDF exports support () — the exported file contains the full diagram XML, so opening it in draw.io recovers the editable diagram. Use double extensions () to signal embedded XML. The dr