4,043 boosters for "r" — open source, verified from GitHub, ready to install
This skill is for running evaluations against models on the Hugging Face Hub on local hardware. It does not cover: If the user wants to run the same eval remotely on Hugging Face Jobs, hand off to the skill and pass it one of the local scripts in this skill.
"name": "huggingface-skills", "description": "Agent Skills for AI/ML tasks including dataset creation, model training, evaluation, and research paper publishing on Hugging Face Hub", "name": "Hugging Face"
Run any workload on fully managed Hugging Face infrastructure. No local setup required—jobs run on cloud CPUs, GPUs, or TPUs and can persist results to the Hugging Face Hub. Use this skill when users want to: When assisting with jobs:
These instructions apply to the entire repository.
A system prompt for Claude Code that enforces defensive security practices and provides CLI guidance, designed to help developers safely use Claude for software engineering tasks while preventing misuse.
This skill automates the process of adding, extracting, and managing evaluation results in Hugging Face model cards, supporting multiple data sources including Artificial Analysis API and custom evaluations with vLLM/lighteval. It's valuable for ML practitioners and model maintainers who need to track and display model performance metrics.
A skill for querying and exploring Hugging Face datasets through the Dataset Viewer API, enabling developers to fetch metadata, paginate rows, search, filter, and download parquet files. Useful for data scientists and engineers working with public datasets in their AI/ML workflows.
Trackio is an ML experiment tracking library that integrates with Hugging Face to log metrics, visualize training progress, and trigger alerts during model development. It's useful for ML engineers and researchers who need real-time monitoring and experiment management.
A skill that generates reusable command-line scripts for automating Hugging Face API interactions, useful for developers who need to repeatedly fetch, process, or chain API calls.
This skill enables users to run Python workloads, Docker jobs, and GPU-intensive tasks on Hugging Face's managed infrastructure without local setup. It's valuable for ML engineers, data scientists, and developers needing cloud compute for training, inference, and batch processing.
This skill enables AI assistants to create, configure, and manage datasets on Hugging Face Hub with SQL-based querying and transformation capabilities. It's valuable for developers building data workflows and ML projects that require programmatic dataset management.
A skill booster for building interactive web UIs and ML demos using Gradio in Python. Developers creating data applications, ML interfaces, and chatbots benefit from guided assistance with Gradio components and patterns.
A skill for fine-tuning and training language models on Hugging Face's cloud GPU infrastructure using TRL, supporting SFT, DPO, GRPO methods and GGUF conversion for local deployment. Developers and ML engineers working with cloud-based model training benefit from this comprehensive guidance.
A skill that enables researchers and AI engineers to publish, manage, and link research papers on Hugging Face Hub with arXiv integration and professional markdown generation. Useful for academics and ML practitioners looking to streamline paper publication workflows.
A multi-agent orchestration system for Claude Code that enables complex agentic workflows through a CLI and MCP server interface. Useful for developers building sophisticated AI-powered automation and coordinated agent systems within Claude's ecosystem.
Enables developers to interact with Hugging Face Hub directly from Claude Code using the `hf` CLI—downloading models/datasets, uploading files, creating repositories, and managing cache without leaving the coding environment.
A comprehensive tutorial for creating, registering, and deploying custom third-party agents that extend UFO² automation capabilities beyond Windows GUI automation. Developers building domain-specific automation agents (hardware control, Linux CLI, web automation, IoT) will find this invaluable for understanding the complete agent architecture and implementation workflow.
Generates per-category catalog JSON files from the npm package's public API. Each catalog lists all UseCase and EventHandler abstractions in a category with their resolved source file paths. LLMs read source files on demand for exact, up-to-date types — no enrichment phase needed. 1. Discovers all
"name": "understand-anything", "description": "AI-powered codebase understanding — analyze, visualize, and explain any project", "homepage": "https://github.com/Lum1104/Understand-Anything",
"description": "Developer experience essentials: GitHub Actions debugging, conversation cloning/half-cloning, context handoffs, and Reddit research via Gemini CLI", "version": "0.14.12", "email": "yoyoyosss@wearehackerone.com"
Ripple is a TypeScript UI framework that combines the best parts of React, Solid, and Svelte. Created by Dominic Gannaway (@trueadm), Ripple is designed to be JS/TS-first with its own file extension that
"name": "context-mode", "version": "1.0.75", "description": "MCP server that saves 98% of your context window with session continuity. Sandboxed code execution in 11 languages, FTS5 knowledge base with BM25 ranking, and automatic state restore across compactions.",
Heuristic scoring (no AI key configured).
Context Mode is an MCP server that dramatically reduces Claude's context window consumption through sandboxed code execution and intelligent full-text search, enabling developers to work with larger codebases and knowledge bases without hitting token limits.