mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2025-09-10 06:14:58 +00:00
Initial commit
This commit is contained in:
commit
18c42e67df
247 changed files with 53775 additions and 0 deletions
33
ktransformers/server/api/openai/legacy/completions.py
Normal file
33
ktransformers/server/api/openai/legacy/completions.py
Normal file
|
@ -0,0 +1,33 @@
|
|||
import json
|
||||
from time import time
|
||||
from uuid import uuid4
|
||||
from fastapi import APIRouter
|
||||
from fastapi.requests import Request
|
||||
from ktransformers.server.utils.create_interface import get_interface
|
||||
from ktransformers.server.schemas.assistants.streaming import stream_response
|
||||
from ktransformers.server.schemas.legacy.completions import CompletionCreate,CompletionObject
|
||||
|
||||
router = APIRouter()
|
||||
|
||||
@router.post("/completions",tags=['openai'])
|
||||
async def create_completion(request:Request,create:CompletionCreate):
|
||||
id = str(uuid4())
|
||||
|
||||
interface = get_interface()
|
||||
print(f'COMPLETION INPUT:----\n{create.prompt}\n----')
|
||||
|
||||
|
||||
|
||||
if create.stream:
|
||||
async def inner():
|
||||
async for token in interface.inference(create.prompt,id):
|
||||
d = {'choices':[{'delta':{'content':token}}]}
|
||||
yield f"data:{json.dumps(d)}\n\n"
|
||||
d = {'choices':[{'delta':{'content':''},'finish_reason':''}]}
|
||||
yield f"data:{json.dumps(d)}\n\n"
|
||||
return stream_response(request,inner())
|
||||
else:
|
||||
comp = CompletionObject(id=id,object='text_completion',created=int(time()))
|
||||
async for token in interface.inference(create.prompt,id):
|
||||
comp.append_token(token)
|
||||
return comp
|
Loading…
Add table
Add a link
Reference in a new issue