mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2025-09-13 00:29:59 +00:00
fix KExpertsMarlin on GPU with out CUDA Graph
This commit is contained in:
parent
f5f6c6b95d
commit
f327695079
2 changed files with 13 additions and 0 deletions
|
@ -310,6 +310,8 @@ class GGUFLoader:
|
|||
values = GGML_DEQUANTIZE[ggml_name](data)
|
||||
values = torch.from_numpy(values.copy())
|
||||
|
||||
if ggml_name == "BF16":
|
||||
values = values.view(torch.bfloat16)
|
||||
values = values.view(shape[-2::-1])
|
||||
|
||||
return values
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue