mirror of
https://github.com/unslothai/unsloth.git
synced 2026-04-28 03:19:57 +00:00
230 lines
15 KiB
Markdown
230 lines
15 KiB
Markdown
<h1 align="center" style="margin:0;">
|
||
<a href="https://unsloth.ai/docs"><picture>
|
||
<source media="(prefers-color-scheme: dark)" srcset="https://raw.githubusercontent.com/unslothai/unsloth/main/images/STUDIO%20WHITE%20LOGO.png">
|
||
<source media="(prefers-color-scheme: light)" srcset="https://raw.githubusercontent.com/unslothai/unsloth/main/images/STUDIO%20BLACK%20LOGO.png">
|
||
<img alt="Unsloth logo" src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/STUDIO%20BLACK%20LOGO.png" height="60" style="max-width:100%;">
|
||
</picture></a>
|
||
</h1>
|
||
<h3 align="center" style="margin: 0; margin-top: 0;">
|
||
Run and train AI models with a unified local interface.
|
||
</h3>
|
||
|
||
<p align="center">
|
||
<a href="#-features">Features</a> •
|
||
<a href="#-quickstart">Quickstart</a> •
|
||
<a href="#-free-notebooks">Notebooks</a> •
|
||
<a href="https://unsloth.ai/docs">Documentation</a> •
|
||
<a href="https://discord.com/invite/unsloth">Discord</a>
|
||
</p>
|
||
<a href="https://unsloth.ai/docs/new/studio">
|
||
<img alt="unsloth studio ui homepage" src="https://raw.githubusercontent.com/unslothai/unsloth/main/studio/frontend/public/studio%20github%20landscape%20colab%20display.png" style="max-width: 100%; margin-bottom: 0;"></a>
|
||
|
||
Unsloth Studio (Beta) lets you run and train text, [audio](https://unsloth.ai/docs/basics/text-to-speech-tts-fine-tuning), [embedding](https://unsloth.ai/docs/new/embedding-finetuning), [vision](https://unsloth.ai/docs/basics/vision-fine-tuning) models on Windows, Linux and macOS.
|
||
|
||
## ⭐ Features
|
||
Unsloth provides several key features for both inference and training:
|
||
### Inference
|
||
* **Search + download + run models** including GGUF, LoRA adapters, safetensors
|
||
* **Export models**: [Save or export](https://unsloth.ai/docs/new/studio/export) models to GGUF, 16-bit safetensors and other formats.
|
||
* **Tool calling**: Support for [self-healing tool calling](https://unsloth.ai/docs/new/studio/chat#auto-healing-tool-calling) and web search
|
||
* **[Code execution](https://unsloth.ai/docs/new/studio/chat#code-execution)**: lets LLMs test code in Claude artifacts and sandbox environments
|
||
* [Auto-tune inference parameters](https://unsloth.ai/docs/new/studio/chat#auto-parameter-tuning) and customize chat templates.
|
||
* Upload images, audio, PDFs, code, DOCX and more file types to chat with.
|
||
### Training
|
||
* Train **500+ models** up to **2x faster** with up to **70% less VRAM**, with no accuracy loss.
|
||
* Supports full fine-tuning, pretraining, 4-bit, 16-bit and, FP8 training.
|
||
* **Observability**: Monitor training live, track loss and GPU usage and customize graphs.
|
||
* **Data Recipes**: [Auto-create datasets](https://unsloth.ai/docs/new/studio/data-recipe) from **PDF, CSV, DOCX** etc. Edit data in a visual-node workflow.
|
||
* **Reinforcement Learning**: The most efficient [RL](https://unsloth.ai/docs/get-started/reinforcement-learning-rl-guide) library, using **80% less VRAM** for GRPO, [FP8](https://unsloth.ai/docs/get-started/reinforcement-learning-rl-guide/fp8-reinforcement-learning) etc.
|
||
* [Multi-GPU](https://unsloth.ai/docs/basics/multi-gpu-training-with-unsloth) training is supported, with major improvements coming soon.
|
||
|
||
## ⚡ Quickstart
|
||
Unsloth can be used in two ways: through **[Unsloth Studio](https://unsloth.ai/docs/new/studio/)**, the web UI, or through **Unsloth Core**, the code-based version. Each has different requirements.
|
||
|
||
### Unsloth Studio (web UI)
|
||
Unsloth Studio (Beta) works on **Windows, Linux, WSL** and **macOS**.
|
||
|
||
* **CPU:** Supported for Chat and Data Recipes currently
|
||
* **NVIDIA:** Training works on RTX 30/40/50, Blackwell, DGX Spark, Station and more
|
||
* **macOS:** Currently supports chat and Data Recipes. **MLX training** is coming very soon
|
||
* **AMD:** Chat works. Train with [Unsloth Core](#unsloth-core-code-based). Studio support is coming soon.
|
||
* **Coming soon:** Training support for Apple MLX, AMD, and Intel.
|
||
* **Multi-GPU:** Available now, with a major upgrade on the way
|
||
|
||
#### macOS, Linux, WSL:
|
||
```bash
|
||
curl -fsSL https://unsloth.ai/install.sh | sh
|
||
```
|
||
If you don't have `curl`, use `wget`. Launch after setup via:
|
||
```bash
|
||
source unsloth_studio/bin/activate
|
||
unsloth studio -H 0.0.0.0 -p 8888
|
||
```
|
||
|
||
#### Windows:
|
||
```powershell
|
||
irm https://unsloth.ai/install.ps1 | iex
|
||
```
|
||
Launch after setup via:
|
||
```powershell
|
||
& .\unsloth_studio\Scripts\unsloth.exe studio -H 0.0.0.0 -p 8888
|
||
```
|
||
|
||
#### Docker
|
||
Use our [Docker image](https://hub.docker.com/r/unsloth/unsloth) ```unsloth/unsloth``` container. Run:
|
||
```bash
|
||
docker run -d -e JUPYTER_PASSWORD="mypassword" \
|
||
-p 8888:8888 -p 8000:8000 -p 2222:22 \
|
||
-v $(pwd)/work:/workspace/work \
|
||
--gpus all \
|
||
unsloth/unsloth
|
||
```
|
||
|
||
#### macOS, Linux, WSL developer installs:
|
||
```bash
|
||
curl -LsSf https://astral.sh/uv/install.sh | sh
|
||
uv venv unsloth_studio --python 3.13
|
||
source unsloth_studio/bin/activate
|
||
uv pip install unsloth --torch-backend=auto
|
||
unsloth studio setup
|
||
unsloth studio -H 0.0.0.0 -p 8888
|
||
```
|
||
|
||
#### Windows PowerShell developer installs:
|
||
```powershell
|
||
winget install -e --id Python.Python.3.13
|
||
winget install --id=astral-sh.uv -e
|
||
uv venv unsloth_studio --python 3.13
|
||
.\unsloth_studio\Scripts\activate
|
||
uv pip install unsloth --torch-backend=auto
|
||
unsloth studio setup
|
||
unsloth studio -H 0.0.0.0 -p 8888
|
||
```
|
||
|
||
#### Nightly - MacOS, Linux, WSL:
|
||
```bash
|
||
curl -LsSf https://astral.sh/uv/install.sh | sh
|
||
git clone --filter=blob:none https://github.com/unslothai/unsloth.git unsloth_studio
|
||
cd unsloth_studio
|
||
uv venv --python 3.13
|
||
source .venv/bin/activate
|
||
uv pip install -e . --torch-backend=auto
|
||
unsloth studio setup
|
||
unsloth studio -H 0.0.0.0 -p 8888
|
||
```
|
||
Then to launch every time:
|
||
```bash
|
||
cd unsloth_studio
|
||
source .venv/bin/activate
|
||
unsloth studio -H 0.0.0.0 -p 8888
|
||
```
|
||
|
||
#### Nightly - Windows:
|
||
Run in Windows Powershell:
|
||
```bash
|
||
winget install -e --id Python.Python.3.13
|
||
winget install --id=astral-sh.uv -e
|
||
git clone --filter=blob:none https://github.com/unslothai/unsloth.git unsloth_studio
|
||
cd unsloth_studio
|
||
uv venv --python 3.13
|
||
.\.venv\Scripts\activate
|
||
uv pip install -e . --torch-backend=auto
|
||
unsloth studio setup
|
||
unsloth studio -H 0.0.0.0 -p 8888
|
||
```
|
||
Then to launch every time:
|
||
```bash
|
||
cd unsloth_studio
|
||
.\.venv\Scripts\activate
|
||
unsloth studio -H 0.0.0.0 -p 8888
|
||
```
|
||
|
||
### Unsloth Core (code-based)
|
||
#### Linux, WSL:
|
||
```bash
|
||
curl -LsSf https://astral.sh/uv/install.sh | sh
|
||
uv venv unsloth_env --python 3.13
|
||
source unsloth_env/bin/activate
|
||
uv pip install unsloth --torch-backend=auto
|
||
```
|
||
#### Windows:
|
||
```powershell
|
||
winget install -e --id Python.Python.3.13
|
||
winget install --id=astral-sh.uv -e
|
||
uv venv unsloth_env --python 3.13
|
||
.\unsloth_env\Scripts\activate
|
||
uv pip install unsloth --torch-backend=auto
|
||
```
|
||
For Windows, `pip install unsloth` works only if you have PyTorch installed. Read our [Windows Guide](https://unsloth.ai/docs/get-started/install/windows-installation).
|
||
You can use the same Docker image as Unsloth Studio.
|
||
|
||
#### AMD, Intel:
|
||
For RTX 50x, B200, 6000 GPUs: `uv pip install unsloth --torch-backend=auto`. Read our guides for: [Blackwell](https://unsloth.ai/docs/blog/fine-tuning-llms-with-blackwell-rtx-50-series-and-unsloth) and [DGX Spark](https://unsloth.ai/docs/blog/fine-tuning-llms-with-nvidia-dgx-spark-and-unsloth). <br>
|
||
To install Unsloth on **AMD** and **Intel** GPUs, follow our [AMD Guide](https://unsloth.ai/docs/get-started/install/amd) and [Intel Guide](https://unsloth.ai/docs/get-started/install/intel).
|
||
|
||
## ✨ Free Notebooks
|
||
|
||
Train for free with our notebooks. Read our [guide](https://unsloth.ai/docs/get-started/fine-tuning-llms-guide). Add dataset, run, then deploy your trained model.
|
||
|
||
| Model | Free Notebooks | Performance | Memory use |
|
||
|-----------|---------|--------|----------|
|
||
| **Qwen3.5 (4B)** | [▶️ Start for free](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Qwen3_5_(4B)_Vision.ipynb) | 1.5x faster | 60% less |
|
||
| **gpt-oss (20B)** | [▶️ Start for free](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/gpt-oss-(20B)-Fine-tuning.ipynb) | 2x faster | 70% less |
|
||
| **gpt-oss (20B): GRPO** | [▶️ Start for free](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/gpt-oss-(20B)-GRPO.ipynb) | 2x faster | 80% less |
|
||
| **Qwen3: Advanced GRPO** | [▶️ Start for free](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Qwen3_(4B)-GRPO.ipynb) | 2x faster | 50% less |
|
||
| **Gemma 3 (4B) Vision** | [▶️ Start for free](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Gemma3_(4B)-Vision.ipynb) | 1.7x faster | 60% less |
|
||
| **embeddinggemma (300M)** | [▶️ Start for free](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/EmbeddingGemma_(300M).ipynb) | 2x faster | 20% less |
|
||
| **Mistral Ministral 3 (3B)** | [▶️ Start for free](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Ministral_3_VL_(3B)_Vision.ipynb) | 1.5x faster | 60% less |
|
||
| **Llama 3.1 (8B) Alpaca** | [▶️ Start for free](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Llama3.1_(8B)-Alpaca.ipynb) | 2x faster | 70% less |
|
||
| **Llama 3.2 Conversational** | [▶️ Start for free](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Llama3.2_(1B_and_3B)-Conversational.ipynb) | 2x faster | 70% less |
|
||
| **Orpheus-TTS (3B)** | [▶️ Start for free](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Orpheus_(3B)-TTS.ipynb) | 1.5x faster | 50% less |
|
||
|
||
- See all our notebooks for: [Kaggle](https://github.com/unslothai/notebooks?tab=readme-ov-file#-kaggle-notebooks), [GRPO](https://unsloth.ai/docs/get-started/unsloth-notebooks#grpo-reasoning-rl-notebooks), [TTS](https://unsloth.ai/docs/get-started/unsloth-notebooks#text-to-speech-tts-notebooks), [embedding](https://unsloth.ai/docs/new/embedding-finetuning) & [Vision](https://unsloth.ai/docs/get-started/unsloth-notebooks#vision-multimodal-notebooks)
|
||
- See [all our models](https://unsloth.ai/docs/get-started/unsloth-model-catalog) and [all our notebooks](https://unsloth.ai/docs/get-started/unsloth-notebooks)
|
||
- See detailed documentation for Unsloth [here](https://unsloth.ai/docs)
|
||
|
||
## 🦥 Unsloth News
|
||
- **Introducing Unsloth Studio**: our new web UI for running and training LLMs. [Blog](https://unsloth.ai/docs/new/studio)
|
||
- **Qwen3.5** - 0.8B, 2B, 4B, 9B, 27B, 35-A3B, 112B-A10B are now supported. [Guide + notebooks](https://unsloth.ai/docs/models/qwen3.5/fine-tune)
|
||
- Train **MoE LLMs 12x faster** with 35% less VRAM - DeepSeek, GLM, Qwen and gpt-oss. [Blog](https://unsloth.ai/docs/new/faster-moe)
|
||
- **Embedding models**: Unsloth now supports ~1.8-3.3x faster embedding fine-tuning. [Blog](https://unsloth.ai/docs/new/embedding-finetuning) • [Notebooks](https://unsloth.ai/docs/get-started/unsloth-notebooks#embedding-models)
|
||
- New **7x longer context RL** vs. all other setups, via our new batching algorithms. [Blog](https://unsloth.ai/docs/new/grpo-long-context)
|
||
- New RoPE & MLP **Triton Kernels** & **Padding Free + Packing**: 3x faster training & 30% less VRAM. [Blog](https://unsloth.ai/docs/new/3x-faster-training-packing)
|
||
- **500K Context**: Training a 20B model with >500K context is now possible on an 80GB GPU. [Blog](https://unsloth.ai/docs/blog/500k-context-length-fine-tuning)
|
||
- **FP8 & Vision RL**: You can now do FP8 & VLM GRPO on consumer GPUs. [FP8 Blog](https://unsloth.ai/docs/get-started/reinforcement-learning-rl-guide/fp8-reinforcement-learning) • [Vision RL](https://unsloth.ai/docs/get-started/reinforcement-learning-rl-guide/vision-reinforcement-learning-vlm-rl)
|
||
- **gpt-oss** by OpenAI: Read our [RL blog](https://unsloth.ai/docs/models/gpt-oss-how-to-run-and-fine-tune/gpt-oss-reinforcement-learning), [Flex Attention](https://unsloth.ai/docs/models/gpt-oss-how-to-run-and-fine-tune/long-context-gpt-oss-training) blog and [Guide](https://unsloth.ai/docs/models/gpt-oss-how-to-run-and-fine-tune).
|
||
|
||
## 🔗 Links and Resources
|
||
| Type | Links |
|
||
| ----------------------------------------------------------------------------------------------------------------------------------------- | ------------------------------------------------------------------------------ |
|
||
| <img width="15" src="https://redditinc.com/hs-fs/hubfs/Reddit%20Inc/Brand/Reddit_Logo.png" /> **r/unsloth Reddit** | [Join Reddit community](https://reddit.com/r/unsloth) |
|
||
| 📚 **Documentation & Wiki** | [Read Our Docs](https://unsloth.ai/docs) |
|
||
| <img width="13" src="https://upload.wikimedia.org/wikipedia/commons/0/09/X_(formerly_Twitter)_logo_late_2025.svg" /> **Twitter (aka X)** | [Follow us on X](https://twitter.com/unslothai) |
|
||
| 💾 **Installation** | [Pip & Docker Install](https://unsloth.ai/docs/get-started/install) |
|
||
| 🔮 **Our Models** | [Unsloth Catalog](https://unsloth.ai/docs/get-started/unsloth-model-catalog) |
|
||
| ✍️ **Blog** | [Read our Blogs](https://unsloth.ai/blog) |
|
||
|
||
### Citation
|
||
|
||
You can cite the Unsloth repo as follows:
|
||
```bibtex
|
||
@software{unsloth,
|
||
author = {Daniel Han, Michael Han and Unsloth team},
|
||
title = {Unsloth},
|
||
url = {https://github.com/unslothai/unsloth},
|
||
year = {2023}
|
||
}
|
||
```
|
||
If you trained a model with 🦥Unsloth, you can use this cool sticker! <img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/made with unsloth.png" width="200" align="center" />
|
||
|
||
### License
|
||
Unsloth uses a dual-licensing model of Apache 2.0 and AGPL-3.0. The core Unsloth package remains licensed under **[Apache 2.0](https://github.com/unslothai/unsloth?tab=Apache-2.0-1-ov-file)**, while certain optional components, such as the Unsloth Studio UI are licensed under the open-source license **[AGPL-3.0](https://github.com/unslothai/unsloth?tab=AGPL-3.0-2-ov-file)**.
|
||
|
||
This structure helps support ongoing Unsloth development while keeping the project open source and enabling the broader ecosystem to continue growing.
|
||
|
||
### Thank You to
|
||
- The [llama.cpp library](https://github.com/ggml-org/llama.cpp) that lets users run and save models with Unsloth
|
||
- The Hugging Face team and their libraries: [transformers](https://github.com/huggingface/transformers) and [TRL](https://github.com/huggingface/trl)
|
||
- The Pytorch and [Torch AO](https://github.com/unslothai/unsloth/pull/3391) team for their contributions
|
||
- NVIDIA for their [NeMo DataDesigner](https://github.com/NVIDIA-NeMo/DataDesigner) library and their contributions
|
||
- And of course for every single person who has contributed or has used Unsloth!
|