mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2026-04-28 11:49:51 +00:00
|
Some checks are pending
Book-CI / test (push) Waiting to run
Book-CI / test-1 (push) Waiting to run
Book-CI / test-2 (push) Waiting to run
Deploy / deploy (macos-latest) (push) Waiting to run
Deploy / deploy (ubuntu-latest) (push) Waiting to run
Deploy / deploy (windows-latest) (push) Waiting to run
* feat: support avx2 bf16 fp8 inference * feat: support avx2 gptq int4 inference * fix: numeric issues in fp8 dequant * Tutorial avx2 (#1900) * fix: prevent injecting -DLLAMA_AVX512=ON on AVX2-only machines * docs: add AVX2 tutorial for running KTransformers on AVX2-only CPUs * Tutorial avx2 (#1901) * fix: prevent injecting -DLLAMA_AVX512=ON on AVX2-only machines * docs: add AVX2 tutorial for running KTransformers on AVX2-only CPUs * docs: update README.md --------- Co-authored-by: Benjamin F <159887351+yyj6666667@users.noreply.github.com> |
||
|---|---|---|
| .. | ||
| amd_blis.md | ||
| AVX2-Tutorial.md | ||
| deepseek-v3.2-sglang-tutorial.md | ||
| experts-sched-Tutorial.md | ||
| GLM-5-Tutorial.md | ||
| Kimi-K2-Thinking-Native.md | ||
| kt-cli.md | ||
| kt-kernel_intro.md | ||
| MiniMax-M2.1-Tutorial.md | ||
| Native-Precision-Tutorial.md | ||
| Qwen3-Coder-Next-Tutorial.md | ||
| README.md | ||