kvcache-ai-ktransformers/ktransformers/optimize/optimize_rules
2025-02-25 13:43:26 +00:00
..
DeepSeek-V2-Chat-multi-gpu-4.yaml optimize gguf dequant, save mem, support Q2_K 2025-02-22 06:13:01 +00:00
DeepSeek-V2-Chat-multi-gpu.yaml optimize gguf dequant, save mem, support Q2_K 2025-02-22 06:13:01 +00:00
DeepSeek-V2-Chat.yaml optimize gguf dequant, save mem, support Q2_K 2025-02-22 06:13:01 +00:00
DeepSeek-V2-Lite-Chat-multi-gpu.yaml optimize gguf dequant, save mem, support Q2_K 2025-02-22 06:13:01 +00:00
DeepSeek-V2-Lite-Chat.yaml optimize gguf dequant, save mem, support Q2_K 2025-02-22 06:13:01 +00:00
DeepSeek-V3-Chat-fp8-linear-ggml-experts.yaml Add data loader to read special weights for fp8; Add special weight process script 2025-02-24 11:34:17 +00:00
DeepSeek-V3-Chat-multi-gpu-4.yaml fix-update-flashinfer_wrapper_local_chat 2025-02-25 12:47:31 +00:00
DeepSeek-V3-Chat-multi-gpu-8.yaml optimize gguf dequant, save mem, support Q2_K 2025-02-22 06:13:01 +00:00
DeepSeek-V3-Chat-multi-gpu-fp8-linear-ggml-experts.yaml add fp8 multi gpu yaml example 2025-02-25 13:32:09 +00:00
DeepSeek-V3-Chat-multi-gpu-marlin.yaml optimize gguf dequant, save mem, support Q2_K 2025-02-22 06:13:01 +00:00
DeepSeek-V3-Chat-multi-gpu.yaml optimize gguf dequant, save mem, support Q2_K 2025-02-22 06:13:01 +00:00
DeepSeek-V3-Chat.yaml support absorb for prefill long context 2025-02-25 08:52:02 +00:00
Internlm2_5-7b-Chat-1m.yaml [feature] release 0.1.3 2024-08-28 16:11:43 +00:00
Mixtral.yaml optimize gguf dequant, save mem, support Q2_K 2025-02-22 06:13:01 +00:00
Moonlight-16B-A3B.yaml fix KExpertsMarlin on GPU with out CUDA Graph 2025-02-24 09:30:54 +00:00
Qwen2-57B-A14B-Instruct-multi-gpu.yaml optimize gguf dequant, save mem, support Q2_K 2025-02-22 06:13:01 +00:00
Qwen2-57B-A14B-Instruct.yaml optimize gguf dequant, save mem, support Q2_K 2025-02-22 06:13:01 +00:00