mirror of
https://github.com/lfnovo/open-notebook.git
synced 2026-04-30 20:39:55 +00:00
* docs: update CHANGELOG for v1.6.0 release * fix: connection error with llama.cpp and OpenAI-compatible providers Bump Esperanto to 2.17.2 which fixes LangChain connection errors caused by garbage collection closing shared HTTP clients. Closes #465
5.5 KiB
5.5 KiB
Changelog
All notable changes to this project will be documented in this file.
The format is based on Keep a Changelog, and this project adheres to Semantic Versioning.
[1.6.2] - 2026-01-24
Fixed
- Connection error with llama.cpp and OpenAI-compatible providers (#465)
- Bump Esperanto to 2.17.2 which fixes LangChain connection errors caused by garbage collection
[1.6.1] - 2026-01-22
Fixed
- "Failed to send message" error with unhelpful logs when chat model is not configured (#358)
- Added detailed error logging with model selection context and full traceback
- Improved error messages to guide users to Settings → Models
- Added warnings when default models are not configured
Docs
- Ollama troubleshooting: Added "Model Name Configuration" section emphasizing exact model names from
ollama list - Added troubleshooting entry for "Failed to send message" error with step-by-step solutions
- Updated AI Chat Issues documentation with model configuration guidance
[1.6.0] - 2026-01-21
Added
- Content-type aware text chunking with automatic HTML, Markdown, and plain text detection (#350, #142)
- Unified embedding generation with mean pooling for large content that exceeds model context limits
- Dedicated embedding commands:
embed_note,embed_insight,embed_source - New utility modules:
chunking.pyandembedding.pyinopen_notebook/utils/ - Japanese (ja-JP) language support (#450)
Changed
- Embedding is now fire-and-forget: domain models submit embedding commands asynchronously after save
rebuild_embeddings_commandnow delegates to individual embed_* commands instead of inline processing- Chunk size reduced to 1500 characters for better compatibility with Ollama embedding models
- Bump Esperanto to 2.16 for increased Ollama context window support
Removed
- Legacy embedding commands:
embed_single_item_command,embed_chunk_command,vectorize_source_command needs_embedding()andget_embedding_content()methods from domain modelssplit_text()function from text_utils (replaced bychunk_text()in chunking module)
Fixed
- Embedding failures when content exceeds model context limits (#350, #142)
- Empty note titles when saving from chat (clean thinking tags from prompt graph output)
- Orphaned embedding/insight records when deleting sources (cascade delete)
- Search results crash with null parent_id (defensive frontend check)
- Database migration 10 cleans up existing orphaned records
[1.5.2] - 2026-01-15
Performance
- Improved source listing speed by 20-30x (#436, closes #351)
- Added database indexes on
sourcefield forsource_insightandsource_embeddingtables - Use SurrealDB
FETCHclause for command status instead of N async calls
- Added database indexes on
[1.5.1] - 2026-01-15
Fixed
- Podcast dialog infinite loop error caused by excessive translation Proxy accesses in loops
- Podcast dialog UI freezing when typing episode name or additional instructions
- Removed incorrect translation keys for user-defined episode profiles (user content should not be translated)
[1.5.0] - 2026-01-15
Added
- Internationalization (i18n) support with Chinese (Simplified and Traditional) translations (#371, closes #344, #349, #360)
- Frontend test infrastructure with Vitest (#371)
- Language toggle component for switching UI language (#371)
- Date localization using date-fns locales (#371)
- Error message translation system (#371)
Fixed
- Accessibility improvements: added missing
id,name, andautoCompleteattributes to form inputs (#371) - Added
DialogDescriptionto dialogs for Radix UI accessibility compliance (#371) - Fixed "Collapsible is changing from uncontrolled to controlled" warning in SettingsForm (#371)
- Fixed lint command for Next.js 16 compatibility (
eslintinstead ofnext lint)
Changed
- Dockerfile optimizations: better layer caching,
--no-install-recommendsfor smaller images (#371) - Dockerfile.single refactored into 3 separate build stages for better caching (#371)
[1.4.0] - 2026-01-14
Added
- CTA button to empty state notebook list for better onboarding (#408)
- Offline deployment support for Docker containers (#414)
Fixed
- Large file uploads (>10MB) by upgrading to Next.js 16 (#423)
- Orphaned uploaded files when sources are removed (#421)
- Broken documentation links to ai-providers.md (#419)
- ZIP support indication removed from UI (#418)
- Duplicate Claude Code workflow runs on PRs (#417)
- Claude Code review workflow now runs on PRs from forks (#416)
Changed
- Upgraded Next.js from 15.4.10 to 16.1.1 (#423)
- Upgraded React from 19.1.0 to 19.2.3 (#423)
- Renamed
middleware.tstoproxy.tsfor Next.js 16 compatibility (#423)
Dependencies
- next: 15.4.10 → 16.1.1
- react: 19.1.0 → 19.2.3
- react-dom: 19.1.0 → 19.2.3
[1.2.4] - 2025-12-14
Added
- Infinite scroll for notebook sources - no more 50 source limit (#325)
- Markdown table rendering in chat responses, search results, and insights (#325)
Fixed
- Timeout errors with Ollama and local LLMs - increased to 10 minutes (#325)
- "Unable to Connect to API Server" on Docker startup - frontend now waits for API health check (#325, #315)
- SSL issues with langchain (#274)
- Query key consistency for source mutations to properly refresh infinite scroll (#325)
- Docker compose start-all flow (#323)
Changed
- Timeout configuration now uses granular httpx.Timeout (short connect, long read) (#325)
Dependencies
- Updated next.js to 15.4.10
- Updated httpx to >=0.27.0 for SSL fix