mirror of
https://github.com/Alishahryar1/free-claude-code.git
synced 2026-04-28 11:30:03 +00:00
Per claude model mapping (#66)
This commit is contained in:
parent
763c8b62b7
commit
0b324e0421
15 changed files with 454 additions and 81 deletions
19
.env.example
19
.env.example
|
|
@ -10,20 +10,23 @@ OPENROUTER_API_KEY=""
|
|||
LM_STUDIO_BASE_URL="http://localhost:1234/v1"
|
||||
|
||||
|
||||
# All Claude model requests are mapped to this model
|
||||
# All Claude model requests are mapped to these models, plain model is fallback
|
||||
# Format: provider_type/model/name
|
||||
# Valid providers: "nvidia_nim" | "open_router" | "lmstudio"
|
||||
MODEL="nvidia_nim/stepfun-ai/step-3.5-flash"
|
||||
MODEL_OPUS="nvidia_nim/z-ai/glm4.7"
|
||||
MODEL_SONNET="arcee-ai/trinity-large-preview:free"
|
||||
MODEL_HAIKU="stepfun/step-3.5-flash:free"
|
||||
MODEL="nvidia_nim/z-ai/glm4.7"
|
||||
|
||||
|
||||
# Provider Config
|
||||
PROVIDER_RATE_LIMIT=40
|
||||
PROVIDER_RATE_WINDOW=60
|
||||
PROVIDER_MAX_CONCURRENCY=5
|
||||
# Provider config
|
||||
PROVIDER_RATE_LIMIT=1
|
||||
PROVIDER_RATE_WINDOW=3
|
||||
PROVIDER_MAX_CONCURRENCY=50
|
||||
|
||||
|
||||
# HTTP client timeouts (seconds) for provider API requests
|
||||
HTTP_READ_TIMEOUT=300
|
||||
HTTP_READ_TIMEOUT=60
|
||||
HTTP_WRITE_TIMEOUT=10
|
||||
HTTP_CONNECT_TIMEOUT=2
|
||||
|
||||
|
|
@ -44,7 +47,7 @@ WHISPER_DEVICE="nvidia_nim"
|
|||
# - For cpu/cuda: Hugging Face ID or short name (tiny, base, small, medium, large-v2, large-v3, large-v3-turbo)
|
||||
# - For nvidia_nim: NVIDIA NIM model (e.g., "nvidia/parakeet-ctc-1.1b-asr", "openai/whisper-large-v3")
|
||||
# - For nvidia_nim, default to "openai/whisper-large-v3" for best performance
|
||||
WHISPER_MODEL="openai/whisper-large-v3"
|
||||
WHISPER_MODEL="openai/whisper-large-v3"
|
||||
HF_TOKEN=""
|
||||
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue