Unsloth Studio bundles local inference, fine-tuning, and model export into a single no-code web UI. One curl command installs it; then you can run GGUF or safetensor models on Mac, Windows, or Linux without writing any code.

The training side is the main draw: 2x faster fine-tuning with 70% less VRAM across 500+ model families (text, vision, TTS, embeddings). LoRA, FP8, and full fine-tuning all work on NVIDIA hardware, with multi-GPU support already in.

A few things that stand out beyond basic inference:

  • Data Recipes: converts raw PDFs, CSVs, or DOCX files into training datasets through a graph-node workflow powered by NVIDIA NeMo Data Designer.
  • Model Arena: side-by-side chat with two models, useful for comparing a base model against a fine-tuned version.
  • Self-healing tool calling: the UI detects malformed tool outputs and retries automatically, with a dedicated Tool Outputs panel.
curl -fsSL https://unsloth.ai/install.sh | sh
unsloth studio -H 0.0.0.0 -p 8888

Unsloth Studio docs