7 boosters for "transformer" — open source, verified from GitHub, ready to install
Train object detection, image classification, and SAM/SAM2 segmentation models on managed cloud GPUs. No local GPU setup required—results are automatically saved to the Hugging Face Hub. Use this skill when users want to: Helper scripts use PEP 723 inline dependencies. Run them with :
Train language models using TRL (Transformer Reinforcement Learning) on fully managed Hugging Face infrastructure. No local GPU setup required—models train on cloud GPUs and results are automatically saved to the Hugging Face Hub. Use this skill when users want to: Use Unsloth () instead of standard
Transformers.js enables running state-of-the-art machine learning models directly in JavaScript, both in browsers and Node.js environments, with no server required. Use this skill when you need to: The pipeline API is the easiest way to use models. It groups together preprocessing, model inference,
This skill is for running evaluations against models on the Hugging Face Hub on local hardware. It does not cover: If the user wants to run the same eval remotely on Hugging Face Jobs, hand off to the skill and pass it one of the local scripts in this skill.
A skill for fine-tuning and training language models on Hugging Face's cloud GPU infrastructure using TRL, supporting SFT, DPO, GRPO methods and GGUF conversion for local deployment. Developers and ML engineers working with cloud-based model training benefit from this comprehensive guidance.
A skill for adapting and optimizing Hugging Face or custom LLM models to run efficiently on vLLM with Ascend NPU support, enabling developers to validate and deploy models with deterministic testing and single-commit delivery.
This skill should be used when users want to train or fine-tune language models using TRL (Transformer Reinforcement Learning) on Hugging Face Jobs infrastructure. Covers SFT, DPO, GRPO and reward modeling training methods, plus GGUF conversion for local deployment. Includes guidance on the TRL Jobs package, UV scripts with PEP 723 format, dataset preparation and validation, hardware selection, cost estimation, Trackio monitoring, Hub authentication, and model persistence. Should be invoked for tasks involving cloud GPU training, GGUF conversion, or when users mention training on Hugging Face Jobs without local GPU setup.