diff --git a/backend/app/model/chat.py b/backend/app/model/chat.py index 883cc522d..496ef70c3 100644 --- a/backend/app/model/chat.py +++ b/backend/app/model/chat.py @@ -36,6 +36,7 @@ McpServers = dict[Literal["mcpServers"], dict[str, dict]] PLATFORM_MAPPING = { "Z.ai": "openai-compatible-model", + "ModelArk": "openai-compatible-model", } class Chat(BaseModel): diff --git a/src/lib/llm.ts b/src/lib/llm.ts index fac1cb2a5..7d6739939 100644 --- a/src/lib/llm.ts +++ b/src/lib/llm.ts @@ -73,6 +73,15 @@ export const INIT_PROVODERS: Provider[] = [ is_valid: false, model_type: "" }, + { + id: 'ModelArk', + name: 'ModelArk', + apiKey: '', + apiHost: 'https://ark.ap-southeast.bytepluses.com/api/v3', + description: "ModelArk model configuration.", + is_valid: false, + model_type: "" + }, { id: 'aws-bedrock', name: 'AWS Bedrock',