[fix] f16 dequantize device ignored

This commit is contained in:
molamooo 2024-08-22 15:10:06 +08:00 committed by GitHub
parent cbc47d0b68
commit 29f4151ebc
No known key found for this signature in database
GPG key ID: B5690EEEBB952194

View file

@ -681,7 +681,7 @@ def dequantize_f16_gpu(data, device):
res = torch.from_numpy(data)
res_gpu = torch.empty_like(res, device=device)
res_gpu.copy_(res)
return res
return res_gpu
GGML_DEQUANTIZE = {
"F32": dequantize_f32,