mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2025-09-06 04:30:03 +00:00
modify moeinfer param
This commit is contained in:
parent
ee24a27001
commit
027b11266c
2 changed files with 3 additions and 1 deletions
|
@ -81,8 +81,10 @@ def local_chat():
|
||||||
content = "hi"
|
content = "hi"
|
||||||
else:
|
else:
|
||||||
content = open(config.prompt_file, "r").read()
|
content = open(config.prompt_file, "r").read()
|
||||||
|
print("User: ", content)
|
||||||
elif os.path.isfile(content):
|
elif os.path.isfile(content):
|
||||||
content = open(content, "r").read()
|
content = open(content, "r").read()
|
||||||
|
print("User: ", content)
|
||||||
messages = his_content + [{"role": "user", "content": content}]
|
messages = his_content + [{"role": "user", "content": content}]
|
||||||
|
|
||||||
async def async_inference(messages):
|
async def async_inference(messages):
|
||||||
|
|
|
@ -163,7 +163,7 @@ class KExpertsCPU(KExpertsBase):
|
||||||
self.config.hidden_size,
|
self.config.hidden_size,
|
||||||
self.config.moe_intermediate_size,
|
self.config.moe_intermediate_size,
|
||||||
64,
|
64,
|
||||||
10,
|
1024,
|
||||||
1024,
|
1024,
|
||||||
gate_ptr,
|
gate_ptr,
|
||||||
up_ptr,
|
up_ptr,
|
||||||
|
|
Loading…
Add table
Reference in a new issue