mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2025-09-10 15:29:39 +00:00
fix local_chat.py chunk_size not effect experts
This commit is contained in:
parent
adc0906967
commit
71a5fc5770
2 changed files with 3 additions and 2 deletions
|
@ -213,7 +213,7 @@ class KExpertsCPU(KExpertsBase):
|
|||
self.config.num_experts_per_tok,
|
||||
self.config.hidden_size,
|
||||
self.config.moe_intermediate_size,
|
||||
max(cuda_graphs),
|
||||
max(cuda_graphs) if isinstance(cuda_graphs, list) else Config().chunk_size,
|
||||
gate_ptr,
|
||||
up_ptr,
|
||||
down_ptr,
|
||||
|
@ -231,7 +231,7 @@ class KExpertsCPU(KExpertsBase):
|
|||
self.config.num_experts_per_tok,
|
||||
self.config.hidden_size,
|
||||
self.config.moe_intermediate_size,
|
||||
max(cuda_graphs),
|
||||
max(cuda_graphs) if isinstance(cuda_graphs, list) else Config().chunk_size,
|
||||
gate_ptr,
|
||||
up_ptr,
|
||||
down_ptr,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue