"""Centralized configuration using Pydantic Settings.""" import os from collections.abc import Mapping from functools import lru_cache from pathlib import Path from typing import Any from dotenv import dotenv_values from pydantic import Field, field_validator, model_validator from pydantic_settings import BaseSettings, SettingsConfigDict from .constants import HTTP_CONNECT_TIMEOUT_DEFAULT from .nim import NimSettings from .provider_ids import SUPPORTED_PROVIDER_IDS def _env_files() -> tuple[Path, ...]: """Return env file paths in priority order (later overrides earlier).""" files: list[Path] = [ Path.home() / ".config" / "free-claude-code" / ".env", Path(".env"), ] if explicit := os.environ.get("FCC_ENV_FILE"): files.append(Path(explicit)) return tuple(files) def _configured_env_files(model_config: Mapping[str, Any]) -> tuple[Path, ...]: """Return the currently configured env files for Settings.""" configured = model_config.get("env_file") if configured is None: return () if isinstance(configured, (str, Path)): return (Path(configured),) return tuple(Path(item) for item in configured) def _env_file_contains_key(path: Path, key: str) -> bool: """Check whether a dotenv-style file defines the given key.""" return _env_file_value(path, key) is not None def _env_file_value(path: Path, key: str) -> str | None: """Return a dotenv value when the file explicitly defines the key.""" if not path.is_file(): return None try: values = dotenv_values(path) except OSError: return None if key not in values: return None value = values[key] return "" if value is None else value def _env_file_override(model_config: Mapping[str, Any], key: str) -> str | None: """Return the last configured dotenv value that explicitly defines a key.""" configured_value: str | None = None for env_file in _configured_env_files(model_config): value = _env_file_value(env_file, key) if value is not None: configured_value = value return configured_value def _removed_env_var_message(model_config: Mapping[str, Any]) -> str | None: """Return a migration error for removed env vars, if present.""" removed_keys = ("NIM_ENABLE_THINKING", "ENABLE_THINKING") replacement = ( "ENABLE_MODEL_THINKING, ENABLE_OPUS_THINKING, " "ENABLE_SONNET_THINKING, or ENABLE_HAIKU_THINKING" ) for removed_key in removed_keys: if removed_key in os.environ: return ( f"{removed_key} has been removed in this release. " f"Rename it to {replacement}." ) for env_file in _configured_env_files(model_config): if _env_file_contains_key(env_file, removed_key): return ( f"{removed_key} has been removed in this release. " f"Rename it to {replacement}. Found in {env_file}." ) return None class Settings(BaseSettings): """Application settings loaded from environment variables.""" # ==================== OpenRouter Config ==================== open_router_api_key: str = Field(default="", validation_alias="OPENROUTER_API_KEY") # ==================== DeepSeek Config ==================== deepseek_api_key: str = Field(default="", validation_alias="DEEPSEEK_API_KEY") # ==================== Messaging Platform Selection ==================== # Valid: "telegram" | "discord" | "none" messaging_platform: str = Field( default="discord", validation_alias="MESSAGING_PLATFORM" ) messaging_rate_limit: int = Field( default=1, validation_alias="MESSAGING_RATE_LIMIT" ) messaging_rate_window: float = Field( default=1.0, validation_alias="MESSAGING_RATE_WINDOW" ) # ==================== NVIDIA NIM Config ==================== nvidia_nim_api_key: str = "" # ==================== LM Studio Config ==================== lm_studio_base_url: str = Field( default="http://localhost:1234/v1", validation_alias="LM_STUDIO_BASE_URL", ) # ==================== Llama.cpp Config ==================== llamacpp_base_url: str = Field( default="http://localhost:8080/v1", validation_alias="LLAMACPP_BASE_URL", ) # ==================== Ollama Config ==================== ollama_base_url: str = Field( default="http://localhost:11434", validation_alias="OLLAMA_BASE_URL", ) # ==================== Model ==================== # All Claude model requests are mapped to this single model (fallback) # Format: provider_type/model/name model: str = "nvidia_nim/z-ai/glm4.7" # Per-model overrides (optional, falls back to MODEL) # Each can use a different provider model_opus: str | None = Field(default=None, validation_alias="MODEL_OPUS") model_sonnet: str | None = Field(default=None, validation_alias="MODEL_SONNET") model_haiku: str | None = Field(default=None, validation_alias="MODEL_HAIKU") # ==================== Per-Provider Proxy ==================== nvidia_nim_proxy: str = Field(default="", validation_alias="NVIDIA_NIM_PROXY") open_router_proxy: str = Field(default="", validation_alias="OPENROUTER_PROXY") lmstudio_proxy: str = Field(default="", validation_alias="LMSTUDIO_PROXY") llamacpp_proxy: str = Field(default="", validation_alias="LLAMACPP_PROXY") # ==================== Provider Rate Limiting ==================== provider_rate_limit: int = Field(default=40, validation_alias="PROVIDER_RATE_LIMIT") provider_rate_window: int = Field( default=60, validation_alias="PROVIDER_RATE_WINDOW" ) provider_max_concurrency: int = Field( default=5, validation_alias="PROVIDER_MAX_CONCURRENCY" ) enable_model_thinking: bool = Field( default=True, validation_alias="ENABLE_MODEL_THINKING" ) enable_opus_thinking: bool | None = Field( default=None, validation_alias="ENABLE_OPUS_THINKING" ) enable_sonnet_thinking: bool | None = Field( default=None, validation_alias="ENABLE_SONNET_THINKING" ) enable_haiku_thinking: bool | None = Field( default=None, validation_alias="ENABLE_HAIKU_THINKING" ) # ==================== HTTP Client Timeouts ==================== http_read_timeout: float = Field( default=120.0, validation_alias="HTTP_READ_TIMEOUT" ) http_write_timeout: float = Field( default=10.0, validation_alias="HTTP_WRITE_TIMEOUT" ) http_connect_timeout: float = Field( default=HTTP_CONNECT_TIMEOUT_DEFAULT, validation_alias="HTTP_CONNECT_TIMEOUT", ) # ==================== Fast Prefix Detection ==================== fast_prefix_detection: bool = True # ==================== Optimizations ==================== enable_network_probe_mock: bool = True enable_title_generation_skip: bool = True enable_suggestion_mode_skip: bool = True enable_filepath_extraction_mock: bool = True # ==================== Local web server tools (web_search / web_fetch) ==================== # Off by default: these tools perform outbound HTTP from the proxy (SSRF risk). enable_web_server_tools: bool = Field( default=False, validation_alias="ENABLE_WEB_SERVER_TOOLS" ) # Comma-separated URL schemes allowed for web_fetch (default: http,https). web_fetch_allowed_schemes: str = Field( default="http,https", validation_alias="WEB_FETCH_ALLOWED_SCHEMES" ) # When true, skip private/loopback/link-local IP blocking for web_fetch (lab only). web_fetch_allow_private_networks: bool = Field( default=False, validation_alias="WEB_FETCH_ALLOW_PRIVATE_NETWORKS" ) # ==================== Debug / diagnostic logging (avoid sensitive content) ==================== # When false (default), API and SSE helpers log only metadata (counts, lengths, ids). log_raw_api_payloads: bool = Field( default=False, validation_alias="LOG_RAW_API_PAYLOADS" ) log_raw_sse_events: bool = Field( default=False, validation_alias="LOG_RAW_SSE_EVENTS" ) # When false (default), unhandled exceptions log only type + route metadata (no message/traceback). log_api_error_tracebacks: bool = Field( default=False, validation_alias="LOG_API_ERROR_TRACEBACKS" ) # When false (default), messaging logs omit text/transcription previews (metadata only). log_raw_messaging_content: bool = Field( default=False, validation_alias="LOG_RAW_MESSAGING_CONTENT" ) # When true, log full Claude CLI stderr, non-JSON lines, and parser error text. log_raw_cli_diagnostics: bool = Field( default=False, validation_alias="LOG_RAW_CLI_DIAGNOSTICS" ) # When true, log exception text / CLI error strings in messaging (may leak user content). log_messaging_error_details: bool = Field( default=False, validation_alias="LOG_MESSAGING_ERROR_DETAILS" ) debug_platform_edits: bool = Field( default=False, validation_alias="DEBUG_PLATFORM_EDITS" ) debug_subagent_stack: bool = Field( default=False, validation_alias="DEBUG_SUBAGENT_STACK" ) # ==================== NIM Settings ==================== nim: NimSettings = Field(default_factory=NimSettings) # ==================== Voice Note Transcription ==================== voice_note_enabled: bool = Field( default=True, validation_alias="VOICE_NOTE_ENABLED" ) # Device: "cpu" | "cuda" | "nvidia_nim" # - "cpu"/"cuda": local Whisper (requires voice_local extra: uv sync --extra voice_local) # - "nvidia_nim": NVIDIA NIM Whisper API (requires voice extra: uv sync --extra voice) whisper_device: str = Field(default="cpu", validation_alias="WHISPER_DEVICE") # Whisper model ID or short name (for local Whisper) or NVIDIA NIM model (for nvidia_nim) # Local Whisper: "tiny", "base", "small", "medium", "large-v2", "large-v3", "large-v3-turbo" # NVIDIA NIM: "nvidia/parakeet-ctc-1.1b-asr", "openai/whisper-large-v3", etc. whisper_model: str = Field(default="base", validation_alias="WHISPER_MODEL") # Hugging Face token for faster model downloads (optional, for local Whisper) hf_token: str = Field(default="", validation_alias="HF_TOKEN") # ==================== Bot Wrapper Config ==================== telegram_bot_token: str | None = None allowed_telegram_user_id: str | None = None discord_bot_token: str | None = Field( default=None, validation_alias="DISCORD_BOT_TOKEN" ) allowed_discord_channels: str | None = Field( default=None, validation_alias="ALLOWED_DISCORD_CHANNELS" ) claude_workspace: str = "./agent_workspace" allowed_dir: str = "" claude_cli_bin: str = Field(default="claude", validation_alias="CLAUDE_CLI_BIN") max_message_log_entries_per_chat: int | None = Field( default=None, validation_alias="MAX_MESSAGE_LOG_ENTRIES_PER_CHAT" ) # ==================== Server ==================== host: str = "0.0.0.0" port: int = 8082 log_file: str = "server.log" # Optional server API key to protect endpoints (Anthropic-style) # Set via env `ANTHROPIC_AUTH_TOKEN`. When empty, no auth is required. anthropic_auth_token: str = Field( default="", validation_alias="ANTHROPIC_AUTH_TOKEN" ) @model_validator(mode="before") @classmethod def reject_removed_env_vars(cls, data: Any) -> Any: """Fail fast when removed environment variables are still configured.""" if message := _removed_env_var_message(cls.model_config): raise ValueError(message) return data # Handle empty strings for optional string fields @field_validator( "telegram_bot_token", "allowed_telegram_user_id", "discord_bot_token", "allowed_discord_channels", "model_opus", "model_sonnet", "model_haiku", "enable_opus_thinking", "enable_sonnet_thinking", "enable_haiku_thinking", mode="before", ) @classmethod def parse_optional_str(cls, v: Any) -> Any: if v == "": return None return v @field_validator("max_message_log_entries_per_chat", mode="before") @classmethod def parse_optional_log_cap(cls, v: Any) -> Any: if v == "" or v is None: return None return v @field_validator("whisper_device") @classmethod def validate_whisper_device(cls, v: str) -> str: if v not in ("cpu", "cuda", "nvidia_nim"): raise ValueError( f"whisper_device must be 'cpu', 'cuda', or 'nvidia_nim', got {v!r}" ) return v @field_validator("messaging_platform") @classmethod def validate_messaging_platform(cls, v: str) -> str: if v not in ("telegram", "discord", "none"): raise ValueError( f"messaging_platform must be 'telegram', 'discord', or 'none', got {v!r}" ) return v @field_validator("messaging_rate_limit") @classmethod def validate_messaging_rate_limit(cls, v: int) -> int: if v <= 0: raise ValueError("messaging_rate_limit must be > 0") return v @field_validator("messaging_rate_window") @classmethod def validate_messaging_rate_window(cls, v: float) -> float: if v <= 0: raise ValueError("messaging_rate_window must be > 0") return float(v) @field_validator("web_fetch_allowed_schemes") @classmethod def validate_web_fetch_allowed_schemes(cls, v: str) -> str: schemes = [part.strip().lower() for part in v.split(",") if part.strip()] if not schemes: raise ValueError("web_fetch_allowed_schemes must list at least one scheme") for scheme in schemes: if not scheme.isascii() or not scheme.isalpha(): raise ValueError( f"Invalid URL scheme in web_fetch_allowed_schemes: {scheme!r}" ) return ",".join(schemes) @field_validator("ollama_base_url") @classmethod def validate_ollama_base_url(cls, v: str) -> str: if v.rstrip("/").endswith("/v1"): raise ValueError( "OLLAMA_BASE_URL must be the Ollama root URL for native Anthropic " "messages, e.g. http://localhost:11434 (without /v1)." ) return v @field_validator("model", "model_opus", "model_sonnet", "model_haiku") @classmethod def validate_model_format(cls, v: str | None) -> str | None: if v is None: return None if "/" not in v: raise ValueError( f"Model must be prefixed with provider type. " f"Valid providers: {', '.join(SUPPORTED_PROVIDER_IDS)}. " f"Format: provider_type/model/name" ) provider = v.split("/", 1)[0] if provider not in SUPPORTED_PROVIDER_IDS: supported = ", ".join(f"'{p}'" for p in SUPPORTED_PROVIDER_IDS) raise ValueError(f"Invalid provider: '{provider}'. Supported: {supported}") return v @model_validator(mode="after") def check_nvidia_nim_api_key(self) -> Settings: if ( self.voice_note_enabled and self.whisper_device == "nvidia_nim" and not self.nvidia_nim_api_key.strip() ): raise ValueError( "NVIDIA_NIM_API_KEY is required when WHISPER_DEVICE is 'nvidia_nim'. " "Set it in your .env file." ) return self @model_validator(mode="after") def prefer_dotenv_anthropic_auth_token(self) -> Settings: """Let explicit .env auth config override stale shell/client tokens.""" dotenv_value = _env_file_override(self.model_config, "ANTHROPIC_AUTH_TOKEN") if dotenv_value is not None: self.anthropic_auth_token = dotenv_value return self def uses_process_anthropic_auth_token(self) -> bool: """Return whether proxy auth came from process env, not dotenv config.""" if _env_file_override(self.model_config, "ANTHROPIC_AUTH_TOKEN") is not None: return False return bool(os.environ.get("ANTHROPIC_AUTH_TOKEN")) @property def provider_type(self) -> str: """Extract provider type from the default model string.""" return Settings.parse_provider_type(self.model) @property def model_name(self) -> str: """Extract the actual model name from the default model string.""" return Settings.parse_model_name(self.model) def resolve_model(self, claude_model_name: str) -> str: """Resolve a Claude model name to the configured provider/model string. Classifies the incoming Claude model (opus/sonnet/haiku) and returns the model-specific override if configured, otherwise the fallback MODEL. """ name_lower = claude_model_name.lower() if "opus" in name_lower and self.model_opus is not None: return self.model_opus if "haiku" in name_lower and self.model_haiku is not None: return self.model_haiku if "sonnet" in name_lower and self.model_sonnet is not None: return self.model_sonnet return self.model def resolve_thinking(self, claude_model_name: str) -> bool: """Resolve whether thinking is enabled for an incoming Claude model name.""" name_lower = claude_model_name.lower() if "opus" in name_lower and self.enable_opus_thinking is not None: return self.enable_opus_thinking if "haiku" in name_lower and self.enable_haiku_thinking is not None: return self.enable_haiku_thinking if "sonnet" in name_lower and self.enable_sonnet_thinking is not None: return self.enable_sonnet_thinking return self.enable_model_thinking def web_fetch_allowed_scheme_set(self) -> frozenset[str]: """Return normalized schemes allowed for web_fetch.""" return frozenset( part.strip().lower() for part in self.web_fetch_allowed_schemes.split(",") if part.strip() ) @staticmethod def parse_provider_type(model_string: str) -> str: """Extract provider type from any 'provider/model' string.""" return model_string.split("/", 1)[0] @staticmethod def parse_model_name(model_string: str) -> str: """Extract model name from any 'provider/model' string.""" return model_string.split("/", 1)[1] model_config = SettingsConfigDict( env_file=_env_files(), env_file_encoding="utf-8", extra="ignore", ) @lru_cache def get_settings() -> Settings: """Get cached settings instance.""" return Settings()