Commit graph

222 commits

Author SHA1 Message Date
Daniel Han
880b59a301 Update README.md 2026-03-17 08:03:32 -07:00
Michael Han
deb76dfa1d
Update README.md 2026-03-17 07:57:46 -07:00
Michael Han
c60636695c
Unsloth Studio.md 2026-03-17 07:53:50 -07:00
Daniel Han
30a18786bf Revert "Improve documentation on how to export model from Colab"
This reverts commit 703c235a7d.
2026-03-13 22:38:41 -07:00
Vishwanath Martur
022a5d566a Improve documentation on how to export model from Colab
Related to #1615

Add documentation and function for exporting models from Colab to local machines.

* **README.md**: Add a new section titled "Exporting Models from Colab to Local Machine" under " Finetune for Free" with detailed steps for exporting models from Colab to local machines.
* **CONTRIBUTING.md**: Add a note about the new documentation section for exporting models from Colab.
* **unsloth/save.py**: Add a new function `export_model_to_local` to handle exporting models from Colab to local machines.

(cherry picked from commit 0361bd658f)
2026-03-13 22:38:41 -07:00
Michael Han
59f7a9006a Qwen3.5 Update.md
Updated with Qwen3.5 Small models
2026-03-02 23:33:22 -08:00
Daniel Han
7c68ec439f Update README.md (#4119) 2026-02-26 09:18:29 -08:00
Daniel Han
618ac74ae0 Update README.md (#4118) 2026-02-26 08:06:21 -08:00
Michael Han
e8ae589e84 Qwen3.5 update.md 2026-02-25 23:56:48 -08:00
Michael Han
ac70db5556 Update README Install.md
Updating to include new installation links
2026-02-17 07:23:31 -08:00
Michael Han
2a7d098203 Update README with faster MoE.md
Adding MoE
2026-02-13 19:38:23 -08:00
Daniel Han
3b1e8d0ae6 Update README.md 2026-02-09 04:50:54 -08:00
Daniel Han
60dd7269a5 Fix broken documentation links, typos, and formatting in README (#4003)
- Fix 14 broken documentation links (all returning 404) caused by docs
  site restructuring (install-and-update -> install, pages moved to
  /docs/blog/ and /docs/models/tutorials/)
- Fix "Qwen2.3-VL" -> "Qwen3-VL" (model does not exist)
- Fix incorrect "GSPO" label on gpt-oss GRPO notebook
- Fix "4b-bit" typo -> "4-bit"
- Fix "sodoku" typo -> "sudoku"
- Fix double dash formatting on FP8 GRPO notebook list item
- Fix citation URL from http:// to https://
- Update "MultiGPU coming soon" to "is now supported"
- Fix Windows installation step numbering (1,3,5,6,7 -> 1,2,3,4,5)
- Fix Advanced/Troubleshooting step numbering (5,6,5 -> 4,5,6)

Co-authored-by: Daniel Hanchen <danielhanchen@users.noreply.github.com>
2026-02-09 04:46:46 -08:00
Daniel Han
ff3f78b6b9 Add PyTorch 2.10 and xformers 0.0.34 support (#3985)
- Add cu126/cu128/cu130 xformers 0.0.34 wheel dependencies for torch 2.10
- Add cu126-torch2100, cu128-torch2100, cu130-torch2100 meta-dependencies
- Add cu126-ampere-torch2100, cu128-ampere-torch2100, cu130-ampere-torch2100 variants
- Update _auto_install.py version detection for torch 2.10.x
- Add CUDA check for torch 2.10 (requires CUDA 12.6, 12.8, or 13.0)
- Update README.md with torch 2.10 installation instructions

Co-authored-by: Daniel Hanchen <danielhanchen@users.noreply.github.com>
2026-02-05 05:56:26 -08:00
Michael Han
f3efb70823 Embedding model fine-tuning support 2026-01-22 21:35:46 -08:00
Michael Han
08e07e7865 Embedding model support 2026-01-22 14:22:03 -08:00
Michael Han
fda54f2634 Update README.md 2026-01-15 08:01:01 -08:00
Michael Han
9b5571fb69 Refresh of Unsloth README.md with https://unsloth.ai/docs 2025-12-30 15:14:27 -08:00
Michael Han
96de7a817d Update README for new unsloth.ai/docs.md 2025-12-27 00:49:19 -08:00
Michael Han
086ccd377f Update README.md 2025-12-13 16:44:44 -08:00
Michael Han
401de54fba Padding free packing update 2025-12-10 21:12:13 -08:00
Michael Han
bff336c7a3 Adding new padding free packing support 2025-12-10 21:10:19 -08:00
Michael Han
3d4f236155 Update README.md 2025-12-04 08:21:20 -08:00
Daniel Han
845e61d351 Update README.md 2025-12-02 04:08:54 -08:00
Daniel Han
14e8e3137d Update README.md 2025-12-02 03:52:50 -08:00
gitpullpull
77d47ecee5 Fix broken link for Advanced pip install instructions (#3652) 2025-11-29 15:33:48 -08:00
Michael Han
bfdf73fa66 Update README.md 2025-11-29 08:01:00 -08:00
Michael Han
f668897b3c Update README.md 2025-11-27 20:52:27 -08:00
Michael Han
ef30739fd7 Update README.md 2025-11-27 20:49:47 -08:00
Daniel Han
86f708097d Float8 GRPO, RL (#3640)
* Enable FP8 + RL training for bf16 models (#3440)

* Enable FP8 + RL training for bf16 models

**Summary:** Enable FP8 + RL training using TorchAO for 1.33x faster training and 42% less model memory usage:
- We quantize the frozen LoRA weights into fp8 and keep the LoRA adapters in bf16
- We leverage TorchAO's `Float8Tensor`, which calls into fbgemm's fp8 x fp8 rowwise matmul kernel
- For now, we need to do an offline quantization first, because vllm doesn't support on-the-fly quantization for torchao yet  (this is in progress: https://github.com/vllm-project/vllm/pull/26327)

**Example usage:**
```
model, tokenizer = FastLanguageModel.from_pretrained(
    model_name = "unsloth/Qwen3-8B-Base",
    max_seq_length = 2048,
    load_in_4bit = False,
    fast_inference = True,
    max_lora_rank = 32,
    load_in_fp8 = True,  # set this to True
)

\# the rest is the same as before
model = FastLanguageModel.get_peft_model(...)
```

**Initial results:**
```
\# fp8
{'train_runtime': 1725.4337, 'train_samples_per_second': 0.232, 'train_steps_per_second': 0.058, 'train_loss': 0.00015715716748673002, 'epoch': 0.01}

\# bf16
{'train_runtime': 2297.8145, 'train_samples_per_second': 0.174, 'train_steps_per_second': 0.044, 'train_loss': 0.00016081033063528594, 'epoch': 0.01}
```

<img width="1199" height="448" alt="Screenshot 2025-11-11 at 4 10 50 PM" src="https://github.com/user-attachments/assets/b6304afd-89e9-42b1-8064-775807e17b23" />

Test script: https://gist.github.com/andrewor14/5b85119fae46845d07b608d420907423

**Requires:**
- https://github.com/pytorch/ao/pull/3158 (torchao nightly or 0.15.0+)
- https://github.com/unslothai/unsloth-zoo/pull/351

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* Update utils.py

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* _get_inference_mode_context_manager

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* Update utils.py

* Update utils.py

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Co-authored-by: Daniel Han <danielhanchen@gmail.com>

* Update __init__.py

* Fix/save torchao model loading logic (#3621)

* make loading gpt-oss-BF16 faster. Linked to unsloth-zoo PR #314

* fix model loading and clean merged model directory

* revert default quant

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* revert mapper.py

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>

* Update loader_utils.py

* Update loader_utils.py

* Add 128x128 PerBlock FP8 + RL (#3629)

* Add 128x128 PerBlock FP8 + RL

**Summary:** Following https://github.com/unslothai/unsloth/pull/3440,
this PR extends torchao FP8 + RL support to also handle 128x128
PerBlock granularity (in addition to PerRow).

**Example usage:**

```
model, tokenizer = FastLanguageModel.from_pretrained(
    model_name = "unsloth/Qwen3-8B-Base",
    max_seq_length = 2048,
    load_in_4bit = False,
    fast_inference = True,
    max_lora_rank = 32,
    load_in_fp8 = "block",  # or "row" or True
)
```

**Initial results:** TBD

**Note:**
- Requires https://github.com/pytorch/ao/pull/3370

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>

* Version

* Update vision.py

* Update rl.py

* Add torch 2.9.1

* Fix auto installer

* Update fp8.py

* Float8

* Update fp8.py

* Update mapper.py

* Update mapper.py

* Update loader_utils.py

* Update loader.py

* Update fp8.py

* Versioning

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

---------

Co-authored-by: andrewor14 <andrewor14@gmail.com>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Co-authored-by: Roland Tannous <115670425+rolandtannous@users.noreply.github.com>
2025-11-25 07:23:26 -08:00
mk0walsk
8efbd5ac9c Fix broken links and typo in README (#3611)
* README Link Fixes

* Update README.md

Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>

---------

Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
2025-11-18 20:04:14 -08:00
Michael Han
c8421a939b Update README.md 2025-11-04 22:00:06 -08:00
Michael Han
381e181e99 Update README.md 2025-10-25 19:26:05 -07:00
Michael Han
e1a9c130e5 Update README.md
Qwen3-VL + DGX
2025-10-14 20:23:32 -07:00
Michael Han
a049fcd460 Update README.md 2025-10-12 05:32:42 -07:00
Michael Han
b235ec7f7f Update README.md 2025-10-04 16:12:02 -07:00
Michael Han
afe9d39981 Update README.md 2025-10-03 04:18:21 -07:00
Michael Han
aeb2829ec9 Update README.md 2025-10-03 04:01:17 -07:00
Michael Han
5745677718 Adding Docker support 2025-10-01 17:04:46 -07:00
Michael Han
a6dfb2894d Update README.md 2025-09-26 17:31:46 -07:00
laz-001
373c3188e1 correct python support statement (#3374) 2025-09-26 04:52:23 -07:00
Michael Han
9c9f85b28a Update README.md
Fresh upate
2025-09-26 02:50:02 -07:00
Michael Han
1b3fdd5565 Update README.md 2025-09-16 10:07:02 -07:00
Daniel Han
6f5b6e90fd Update README.md 2025-09-15 01:46:07 -07:00
Daniel Han
846a5dcbc4 Update README.md 2025-09-15 01:43:11 -07:00
Daniel Han
29ed805a13 Update README.md 2025-09-15 01:42:59 -07:00
Daniel Han
db4f3cde14 Update README.md 2025-09-15 01:40:28 -07:00
Daniel Han
2f8baabd7a Update README.md 2025-09-15 01:40:06 -07:00
Daniel Han
92f972bb7c Update README.md 2025-09-15 01:39:39 -07:00
Daniel Han
46e7370878 Blackwell support 2025-09-15 01:39:03 -07:00