diff --git a/ktransformers/server/api/openai/endpoints/chat.py b/ktransformers/server/api/openai/endpoints/chat.py index 4008aa7..84e3ee2 100644 --- a/ktransformers/server/api/openai/endpoints/chat.py +++ b/ktransformers/server/api/openai/endpoints/chat.py @@ -143,16 +143,6 @@ async def chat_completion(request: Request, create: ChatCompletionCreate): # Process messages with tool functionality if needed enhanced_messages = list(create.messages) - if create.model != Config().model_name: - return JSONResponse( - status_code=400, - content={ - "error": { - "message": "Model not found", - "code": 404, - "type": "NotFound" - } - }) if create.max_tokens<0 or create.max_completion_tokens<0: return JSONResponse( status_code=400, diff --git a/ktransformers/server/api/openai/legacy/completions.py b/ktransformers/server/api/openai/legacy/completions.py index 63535e2..2f0475e 100644 --- a/ktransformers/server/api/openai/legacy/completions.py +++ b/ktransformers/server/api/openai/legacy/completions.py @@ -14,16 +14,6 @@ router = APIRouter() @router.post("/completions",tags=['openai']) async def create_completion(request:Request, create:CompletionCreate): id = str(uuid4()) - if create.model != Config().model_name: - return JSONResponse( - status_code=400, - content={ - "error": { - "message": "Model not found", - "code": 404, - "type": "NotFound" - } - }) if create.max_tokens<0: return JSONResponse( status_code=400,