mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2026-04-28 03:39:48 +00:00
|
Some checks are pending
Book-CI / test (push) Waiting to run
Book-CI / test-1 (push) Waiting to run
Book-CI / test-2 (push) Waiting to run
Deploy / deploy (macos-latest) (push) Waiting to run
Deploy / deploy (ubuntu-latest) (push) Waiting to run
Deploy / deploy (windows-latest) (push) Waiting to run
* feat: support avx2 bf16 fp8 inference * feat: support avx2 gptq int4 inference * fix: numeric issues in fp8 dequant * Tutorial avx2 (#1900) * fix: prevent injecting -DLLAMA_AVX512=ON on AVX2-only machines * docs: add AVX2 tutorial for running KTransformers on AVX2-only CPUs * Tutorial avx2 (#1901) * fix: prevent injecting -DLLAMA_AVX512=ON on AVX2-only machines * docs: add AVX2 tutorial for running KTransformers on AVX2-only CPUs * docs: update README.md --------- Co-authored-by: Benjamin F <159887351+yyj6666667@users.noreply.github.com> |
||
|---|---|---|
| .. | ||
| api/server | ||
| AVX2-Tutorial_zh.md | ||
| clawdbot_integration_guide.md | ||
| DeepseekR1_V3_tutorial_zh.md | ||
| DeepseekR1_V3_tutorial_zh_for_Ascend_NPU.md | ||
| KTransformers-Fine-Tuning_Developer-Technical-Notes_zh.md | ||
| KTransformers-Fine-Tuning_User-Guide_zh.md | ||
| Low-cost_Cloud_Training_and_Inference_KTransformers+AutoDL+LlamaFactory_An_On-demand_Cost-efficient_Integrated_Pipeline_for_Ultra-large_Model_Fine-tuning_and_Inference.pdf | ||
| Qwen3-MoE_tutorial_zh_for_Ascend_NPU.md | ||
| 【云端低价训推】 KTransformers+AutoDL+LlamaFactory:随用随租的低成本超大模型「微调+推理」一体化流程.pdf | ||