kvcache-ai-ktransformers/kt-kernel/python
mrhaoxx dd1da65d90
feat(sft): add Qwen3.5 MoE support + fused checkpoint loading
- arch.py: add Qwen3_5Moe arch match, read config from text_config,
  _get_layers_prefix returns model.language_model.layers for Qwen3.5,
  _get_model_container_and_layers searches language_model attr
- weights.py: load_experts_from_checkpoint_files detects fused format
  (gate_up_proj in weight_map) and splits into gate/up/down
- wrapper.py: hidden_size fallback to text_config

Verified: Qwen3.5-35B-A3B (256 experts, fused format) E2E pass.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-04-20 17:19:15 +08:00
..
cli [fix] improve Sglang kt-kernel detect time duration (#1887) 2026-03-18 23:07:40 +08:00
sft feat(sft): add Qwen3.5 MoE support + fused checkpoint loading 2026-04-20 17:19:15 +08:00
utils merge: integrate origin/main into sft branch 2026-04-08 23:19:28 +08:00
__init__.py merge: integrate origin/main into sft branch 2026-04-08 23:19:28 +08:00
_cpu_detect.py [feat](kt-kernel): Fix CPU instruction set variants for build & install (#1746) 2025-12-24 18:57:45 +08:00
experts.py merge: integrate origin/main into sft branch 2026-04-08 23:19:28 +08:00
experts_base.py merge: integrate origin/main into sft branch 2026-04-08 23:19:28 +08:00