open-notebook/open_notebook/ai/models.py
Luis Novo 47c513edfd
Some checks are pending
Development Build / extract-version (push) Waiting to run
Development Build / test-build (push) Blocked by required conditions
Development Build / summary (push) Blocked by required conditions
fix: improve error logging for chat model configuration issues (#458)
* docs: update CHANGELOG for v1.6.0 release

* fix: improve error logging for chat model configuration issues (#358)

- Add detailed error logging in provision.py when model lookup fails
- Add warning logging in models.py when default model is not configured
- Add traceback logging in chat router exception handler
- Update Ollama docs with model name configuration guidance
- Update troubleshooting docs with "Failed to send message" solutions
- Bump version to 1.6.1

* chore: uvlock
2026-01-23 16:45:13 -03:00

207 lines
7.3 KiB
Python

from typing import ClassVar, Dict, Optional, Union
from esperanto import (
AIFactory,
EmbeddingModel,
LanguageModel,
SpeechToTextModel,
TextToSpeechModel,
)
from loguru import logger
from open_notebook.database.repository import ensure_record_id, repo_query
from open_notebook.domain.base import ObjectModel, RecordModel
ModelType = Union[LanguageModel, EmbeddingModel, SpeechToTextModel, TextToSpeechModel]
class Model(ObjectModel):
table_name: ClassVar[str] = "model"
name: str
provider: str
type: str
@classmethod
async def get_models_by_type(cls, model_type):
models = await repo_query(
"SELECT * FROM model WHERE type=$model_type;", {"model_type": model_type}
)
return [Model(**model) for model in models]
class DefaultModels(RecordModel):
record_id: ClassVar[str] = "open_notebook:default_models"
default_chat_model: Optional[str] = None
default_transformation_model: Optional[str] = None
large_context_model: Optional[str] = None
default_text_to_speech_model: Optional[str] = None
default_speech_to_text_model: Optional[str] = None
# default_vision_model: Optional[str]
default_embedding_model: Optional[str] = None
default_tools_model: Optional[str] = None
@classmethod
async def get_instance(cls) -> "DefaultModels":
"""Always fetch fresh defaults from database (override parent caching behavior)"""
result = await repo_query(
"SELECT * FROM ONLY $record_id",
{"record_id": ensure_record_id(cls.record_id)},
)
if result:
if isinstance(result, list) and len(result) > 0:
data = result[0]
elif isinstance(result, dict):
data = result
else:
data = {}
else:
data = {}
# Create new instance with fresh data (bypass singleton cache)
instance = object.__new__(cls)
object.__setattr__(instance, "__dict__", {})
super(RecordModel, instance).__init__(**data)
return instance
class ModelManager:
def __init__(self):
pass # No caching needed
async def get_model(self, model_id: str, **kwargs) -> Optional[ModelType]:
"""Get a model by ID. Esperanto will cache the actual model instance."""
if not model_id:
return None
try:
model: Model = await Model.get(model_id)
except Exception:
raise ValueError(f"Model with ID {model_id} not found")
if not model.type or model.type not in [
"language",
"embedding",
"speech_to_text",
"text_to_speech",
]:
raise ValueError(f"Invalid model type: {model.type}")
# Create model based on type (Esperanto will cache the instance)
if model.type == "language":
return AIFactory.create_language(
model_name=model.name,
provider=model.provider,
config=kwargs,
)
elif model.type == "embedding":
return AIFactory.create_embedding(
model_name=model.name,
provider=model.provider,
config=kwargs,
)
elif model.type == "speech_to_text":
return AIFactory.create_speech_to_text(
model_name=model.name,
provider=model.provider,
config=kwargs,
)
elif model.type == "text_to_speech":
return AIFactory.create_text_to_speech(
model_name=model.name,
provider=model.provider,
config=kwargs,
)
else:
raise ValueError(f"Invalid model type: {model.type}")
async def get_defaults(self) -> DefaultModels:
"""Get the default models configuration from database"""
defaults = await DefaultModels.get_instance()
if not defaults:
raise RuntimeError("Failed to load default models configuration")
return defaults
async def get_speech_to_text(self, **kwargs) -> Optional[SpeechToTextModel]:
"""Get the default speech-to-text model"""
defaults = await self.get_defaults()
model_id = defaults.default_speech_to_text_model
if not model_id:
return None
model = await self.get_model(model_id, **kwargs)
assert model is None or isinstance(model, SpeechToTextModel), (
f"Expected SpeechToTextModel but got {type(model)}"
)
return model
async def get_text_to_speech(self, **kwargs) -> Optional[TextToSpeechModel]:
"""Get the default text-to-speech model"""
defaults = await self.get_defaults()
model_id = defaults.default_text_to_speech_model
if not model_id:
return None
model = await self.get_model(model_id, **kwargs)
assert model is None or isinstance(model, TextToSpeechModel), (
f"Expected TextToSpeechModel but got {type(model)}"
)
return model
async def get_embedding_model(self, **kwargs) -> Optional[EmbeddingModel]:
"""Get the default embedding model"""
defaults = await self.get_defaults()
model_id = defaults.default_embedding_model
if not model_id:
return None
model = await self.get_model(model_id, **kwargs)
assert model is None or isinstance(model, EmbeddingModel), (
f"Expected EmbeddingModel but got {type(model)}"
)
return model
async def get_default_model(self, model_type: str, **kwargs) -> Optional[ModelType]:
"""
Get the default model for a specific type.
Args:
model_type: The type of model to retrieve (e.g., 'chat', 'embedding', etc.)
**kwargs: Additional arguments to pass to the model constructor
"""
defaults = await self.get_defaults()
model_id = None
if model_type == "chat":
model_id = defaults.default_chat_model
elif model_type == "transformation":
model_id = (
defaults.default_transformation_model or defaults.default_chat_model
)
elif model_type == "tools":
model_id = defaults.default_tools_model or defaults.default_chat_model
elif model_type == "embedding":
model_id = defaults.default_embedding_model
elif model_type == "text_to_speech":
model_id = defaults.default_text_to_speech_model
elif model_type == "speech_to_text":
model_id = defaults.default_speech_to_text_model
elif model_type == "large_context":
model_id = defaults.large_context_model
if not model_id:
logger.warning(
f"No default model configured for type '{model_type}'. "
f"Please go to Settings → Models and set a default model."
)
return None
try:
return await self.get_model(model_id, **kwargs)
except ValueError as e:
logger.error(
f"Failed to load default model for type '{model_type}': {e}. "
f"The configured model_id '{model_id}' may have been deleted or misconfigured. "
f"Please go to Settings → Models and reconfigure the default model."
)
return None
model_manager = ModelManager()