Release 1.2 (#242)

* chore: improve podcast transcripts

* fix: remove date from insight - fixes #241

* fix: improve scrolling on source and insights - fixes #237

* chore: update esperanto to fix: #234

* chore: update esperanto to fix #226

* fix: process vectorization as subcommands to handle larger documents more gracefully - fix: #229

* feat: enable background job retry capabilities

* feat: reenable content types that were disabled during alpha version

* fix: remove unnecessary model caching causing many issues.

* feat: support multiple azure endpoints and keys just like openai compatible. Fixes #215

* docs: update azure variables

* chore: bump and update dependencies
This commit is contained in:
Luis Novo 2025-11-01 14:40:00 -03:00 committed by GitHub
parent bc35a95117
commit f79a9040ae
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
20 changed files with 1077 additions and 435 deletions

View file

@ -10,6 +10,7 @@ from loguru import logger
from typing_extensions import Annotated, TypedDict
from open_notebook.domain.content_settings import ContentSettings
from open_notebook.domain.models import Model, ModelManager
from open_notebook.domain.notebook import Asset, Source
from open_notebook.domain.transformation import Transformation
from open_notebook.graphs.transformation import graph as transform_graph
@ -48,6 +49,20 @@ async def content_process(state: SourceState) -> dict:
)
content_state["output_format"] = "markdown"
# Add speech-to-text model configuration from Default Models
try:
model_manager = ModelManager()
defaults = await model_manager.get_defaults()
if defaults.default_speech_to_text_model:
stt_model = await Model.get(defaults.default_speech_to_text_model)
if stt_model:
content_state["audio_provider"] = stt_model.provider
content_state["audio_model"] = stt_model.name
logger.debug(f"Using speech-to-text model: {stt_model.provider}/{stt_model.name}")
except Exception as e:
logger.warning(f"Failed to retrieve speech-to-text model configuration: {e}")
# Continue without custom audio model (content-core will use its default)
processed_state = await extract_content(content_state)
return {"content_state": processed_state}