revert using FP16

This commit is contained in:
rnwang04 2025-07-01 14:24:27 +08:00
parent 495ae37478
commit 5b5deda420

View file

@ -79,8 +79,9 @@ def local_chat(
if mode == 'long_context': if mode == 'long_context':
assert config.architectures[0] == "LlamaForCausalLM", "only LlamaForCausalLM support long_context mode" assert config.architectures[0] == "LlamaForCausalLM", "only LlamaForCausalLM support long_context mode"
torch.set_default_dtype(torch.float16) torch.set_default_dtype(torch.float16)
elif xpu_fp16_model(config): # elif xpu_fp16_model(config):
torch.set_default_dtype(torch.float16) # # using FP16 may cause accuracy issues, triggering core dumped during runtime
# torch.set_default_dtype(torch.float16)
else: else:
torch.set_default_dtype(config.torch_dtype) torch.set_default_dtype(config.torch_dtype)