hugging-face-model-trainer by huggingface

This skill should be used when users want to train or fine-tune language models using TRL (Transformer Reinforcement Learning) on Hugging Face Jobs infrastructure. Covers SFT, DPO, GRPO and reward modeling training methods, plus GGUF conversion for local deployment. Includes guidance on the TRL Jobs package, UV scripts with PEP 723 format, dataset preparation and validation, hardware selection, cost estimation, Trackio monitoring, Hub authentication, and model persistence. Should be invoked for tasks involving cloud GPU training, GGUF conversion, or when users mention training on Hugging Face Jobs without local GPU setup.

Content & Writing
7.4K Stars
433 Forks
Updated Jan 30, 2026, 09:18 PM

Why Use This

This skill provides specialized capabilities for huggingface's codebase.

Use Cases

  • Developing new features in the huggingface repository
  • Refactoring existing code to follow huggingface standards
  • Understanding and working with huggingface's codebase structure

Install Guide

2 steps
  1. 1

    Download Ananke

    Skip this step if Ananke is already installed.

  2. 2

    Install inside Ananke

    Click Install Skill, paste the link below, then press Install.

    https://github.com/huggingface/skills/tree/main/skills/hugging-face-model-trainer

Skill Snapshot

Auto scan of skill assets. Informational only.

Valid SKILL.md

Checks against SKILL.md specification

Source & Community

Repository skills
Skill Version
main
Community
7.4K 433
Updated At Jan 30, 2026, 09:18 PM

Skill Stats

SKILL.md 719 Lines
Total Files 3
Total Size 27.3 KB
License Complete terms in LICENSE.txt