4 boosters for "fine-tuning" — open source, verified from GitHub, ready to install
Train object detection, image classification, and SAM/SAM2 segmentation models on managed cloud GPUs. No local GPU setup required—results are automatically saved to the Hugging Face Hub. Use this skill when users want to: Helper scripts use PEP 723 inline dependencies. Run them with :
A Cursor-integrated coding standards and linting configuration toolkit for NeMo Curator projects, helping teams enforce consistent data processing code quality with Ruff-based rules and exceptions.
A practical guide to Supervised Fine-Tuning using SFTTrainer and Unsloth optimizations, enabling developers to efficiently adapt pre-trained LLMs for instruction-following with 2x faster training. Ideal for ML engineers building custom instruction-tuned models and reasoning systems.
"name": "Chris von Csefalvay", "email": "chris@chrisvoncsefalvay.com" "description": "Claude Code skills marketplace for fine-tuning language models with Unsloth",