fix glm5 docs (#1845)
Some checks failed
Book-CI / test (push) Has been cancelled
Book-CI / test-1 (push) Has been cancelled
Book-CI / test-2 (push) Has been cancelled
Deploy / deploy (macos-latest) (push) Has been cancelled
Deploy / deploy (ubuntu-latest) (push) Has been cancelled
Deploy / deploy (windows-latest) (push) Has been cancelled

This commit is contained in:
Oql 2026-02-12 02:33:37 +08:00 committed by GitHub
parent 7f7aeaeff6
commit 1c72b3f5bd
No known key found for this signature in database
GPG key ID: B5690EEEBB952194

View file

@ -31,12 +31,10 @@ Before starting, ensure you have:
2. **KT-Kernel installed**
Please follow [kt-kernel](https://github.com/kvcache-ai/ktransformers/blob/main/kt-kernel/README.md)
After installation, verify the CLI is working:
```bash
kt version
git clone https://github.com/kvcache-ai/ktransformers.git
git submodule update --init --recursive
cd kt-kernel && ./install.sh
```
3. **transformers reinstalled**
@ -141,6 +139,8 @@ Layerwise prefill requires one extra MoE layer's worth of VRAM.
If you encounter OOM, adjust `--kt-num-gpu-experts`, `--chunked-prefill-size`, `--mem-fraction-static` and `--max-total-tokens` when launching the server.
If you encounter other issues, try `kt doctor` to diagnose your setup.
See [KT-Kernel Parameters](https://github.com/kvcache-ai/ktransformers/tree/main/kt-kernel#kt-kernel-parameters) for detailed parameter tuning guidelines.
## Step 3: Send Inference Requests