mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2026-04-28 03:39:48 +00:00
|
Some checks are pending
Book-CI / test (push) Waiting to run
Book-CI / test-1 (push) Waiting to run
Book-CI / test-2 (push) Waiting to run
Deploy / deploy (macos-latest) (push) Waiting to run
Deploy / deploy (ubuntu-latest) (push) Waiting to run
Deploy / deploy (windows-latest) (push) Waiting to run
* feat: support avx2 bf16 fp8 inference * feat: support avx2 gptq int4 inference * fix: numeric issues in fp8 dequant * Tutorial avx2 (#1900) * fix: prevent injecting -DLLAMA_AVX512=ON on AVX2-only machines * docs: add AVX2 tutorial for running KTransformers on AVX2-only CPUs * Tutorial avx2 (#1901) * fix: prevent injecting -DLLAMA_AVX512=ON on AVX2-only machines * docs: add AVX2 tutorial for running KTransformers on AVX2-only CPUs * docs: update README.md --------- Co-authored-by: Benjamin F <159887351+yyj6666667@users.noreply.github.com> |
||
|---|---|---|
| .. | ||
| api/server | ||
| kt-kernel | ||
| operators | ||
| SFT | ||
| AMX.md | ||
| balance-serve.md | ||
| benchmark.md | ||
| deepseek-v2-injection.md | ||
| DeepseekR1_V3_tutorial.md | ||
| Docker.md | ||
| Docker_xpu.md | ||
| FAQ.md | ||
| fp8_kernel.md | ||
| install.md | ||
| Kimi-K2-Thinking.md | ||
| Kimi-K2.5.md | ||
| Kimi-K2.md | ||
| Kllama_tutorial_DeepSeekV2Lite.ipynb | ||
| KTransformers Full Introduction for Motivation and Practice.pdf | ||
| KTransformers-FT_PPT_share.pdf | ||
| llama4.md | ||
| long_context_introduction.md | ||
| long_context_tutorial.md | ||
| makefile_usage.md | ||
| MiniMax-M2.5.md | ||
| multi-gpu-tutorial.md | ||
| prefix_cache.md | ||
| Qwen3-Next.md | ||
| Qwen3.5.md | ||
| ROCm.md | ||
| SFT_Installation_Guide_KimiK2.5.md | ||
| SFT_Installation_Guide_KimiK2.md | ||
| SmallThinker_and_Glm4moe.md | ||
| V3-success.md | ||
| xpu.md | ||