Concedo
b6f6338bba
Merge branch 'upstream' into concedo_experimental
...
# Conflicts:
# .github/workflows/build-linux-cross.yml
# .github/workflows/build.yml
# CODEOWNERS
# ggml/CMakeLists.txt
# ggml/src/ggml-cuda/fattn.cu
# ggml/src/ggml-webgpu/CMakeLists.txt
# ggml/src/ggml-webgpu/ggml-webgpu.cpp
# ggml/src/ggml-webgpu/wgsl-shaders/mul_mat.tmpl.wgsl
# tests/test-backend-ops.cpp
# tests/test-chat-template.cpp
# tools/llama-bench/llama-bench.cpp
# tools/rpc/README.md
# tools/server/README.md
2025-10-09 01:33:27 +08:00
issixx
d2ee056e1d
server : fix cancel pending task ( #16467 )
...
Check Pre-Tokenizer Hashes / pre-tokenizer-hashes (push) Waiting to run
Python check requirements.txt / check-requirements (push) Waiting to run
Python Type-Check / pyright type-check (push) Waiting to run
Co-authored-by: DevAI <DevAI@gmail.com>
2025-10-08 11:20:18 +03:00
Georgi Gerganov
7fdd16b432
server : improve context checkpoint logic ( #16440 )
2025-10-08 10:57:29 +03:00
Georgi Gerganov
df1b612e29
server : add /v1/health endpoint ( #16461 )
...
* server : add /v1/health endpoint
* cont : update readme
2025-10-07 15:57:14 +03:00
Sascha Rogmann
4e0388aa8a
webui : added download action ( #13552 ) ( #16282 )
...
* webui : added download action (#13552 )
* webui : import and export (for all conversations)
* webui : fixed download-format, import of one conversation
* webui : add ExportedConversations type for chat import/export
* feat: Update naming & order
* chore: Linting
* webui : Updated static build output
---------
Co-authored-by: Aleksander Grygier <aleksander.grygier@gmail.com>
2025-10-07 11:11:08 +02:00
Oleksandr Kuvshynov
c5fef0fcea
server: update readme to mention n_past_max metric ( #16436 )
...
https://github.com/ggml-org/llama.cpp/pull/15361 added new metric
exported, but I've missed this doc.
2025-10-06 10:53:31 +03:00
Concedo
1d728bbc89
Merge commit ' 128d522c04' into concedo_experimental
...
# Conflicts:
# .github/workflows/build.yml
# .github/workflows/release.yml
# ggml/src/ggml-vulkan/ggml-vulkan.cpp
# tests/test-alloc.cpp
# tests/test-chat.cpp
2025-10-04 23:51:22 +08:00
ddh0
f6dcda3900
server : context checkpointing for hybrid and recurrent models ( #16382 )
...
* initial commit for branch 3
* generalize `swa_checkpoint` to `ctx_checkpoint`
this extends `llama-server`'s SWA checkpointing logic to include
hybrid/recurrent models such as Jamba, Granite
* oops
* disable debug prints
* keep backwards compat with `--swa-checkpoints`
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
* update prompt re-processing message
* fix off-by-one error per GG
* keep `seq_rm` log per GG
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
* server : fix checkpoint logic to support recurrent caches
* server : cleanup and fixes
---------
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
2025-10-03 21:34:51 +03:00
Aleksander Grygier
84c8e305e8
Fix missing messages on sibling navigation ( #16408 )
...
* fix: resolve message disappearing issue when navigating between regenerated siblings by using current leaf nodes instead of cached sibling IDs
* chore: update webui build output
* chore: update webui build output
2025-10-03 12:51:40 +02:00
Aleksander Grygier
77233277c9
Capture model name only after first token (streaming) or completed request ( #16405 )
...
* feat: Capture model name only after first token (streaming) or completed request (non-streaming)
* chore: update webui build output
* chore: update webui build output
2025-10-03 11:30:39 +02:00
Concedo
e706d33367
Merge branch 'upstream' into concedo_experimental
...
# Conflicts:
# .devops/intel.Dockerfile
# .devops/rocm.Dockerfile
# .github/workflows/build.yml
# .github/workflows/release.yml
# CODEOWNERS
# ci/run.sh
# docs/backend/SYCL.md
# ggml/CMakeLists.txt
# ggml/src/ggml-cuda/fattn-wmma-f16.cu
# ggml/src/ggml-hip/CMakeLists.txt
# ggml/src/ggml-musa/CMakeLists.txt
# ggml/src/ggml-webgpu/ggml-webgpu.cpp
# ggml/src/ggml-webgpu/wgsl-shaders/rms_norm.wgsl
# tests/test-backend-ops.cpp
# tests/test-barrier.cpp
# tests/test-chat.cpp
2025-10-03 16:44:33 +08:00
Concedo
1731a3212c
Merge commit ' ded67b9444' into concedo_experimental
...
# Conflicts:
# .devops/rocm.Dockerfile
# .github/workflows/build.yml
# .github/workflows/docker.yml
# .github/workflows/release.yml
# CODEOWNERS
# common/CMakeLists.txt
# common/arg.cpp
# ggml/src/ggml-opencl/ggml-opencl.cpp
# ggml/src/ggml-opencl/kernels/get_rows.cl
# ggml/src/ggml-opencl/kernels/pad.cl
# ggml/src/ggml-webgpu/ggml-webgpu.cpp
# ggml/src/ggml-webgpu/wgsl-shaders/embed_wgsl.py
# tests/test-arg-parser.cpp
# tests/test-backend-ops.cpp
# tools/run/run.cpp
2025-10-03 16:15:27 +08:00
Aleksander Grygier
136bda78c5
webui : Fix messages payload sent to chat completions ( #16402 )
...
Check Pre-Tokenizer Hashes / pre-tokenizer-hashes (push) Has been cancelled
Python check requirements.txt / check-requirements (push) Has been cancelled
Python Type-Check / pyright type-check (push) Has been cancelled
* fix: Include just the currently active message branches instead of all in chat completions request
* chore: Build webui static output
* chore: Formatting
* chore: update webui build output
2025-10-03 10:11:34 +03:00
Pascal
5113efd34c
fix: track viewportHeight via window.innerHeight to avoid unwanted scrolling ( #16356 )
...
Use <svelte:window bind:innerHeight> instead of manual resize listener
Co-authored-by: Aleksander Grygier <aleksander.grygier@gmail.com>
2025-10-03 08:01:31 +02:00
Aleksander Grygier
764799279f
Conversation action dialogs as singletons from Chat Sidebar + apply conditional rendering for Actions Dropdown for Chat Conversation Items ( #16369 )
...
* fix: Render Conversation action dialogs as singletons from Chat Sidebar level
* chore: update webui build output
* fix: Render Actions Dropdown conditionally only when user hovers conversation item + remove unused markup
* chore: Update webui static build
* fix: Always truncate conversation names
* chore: Update webui static build
2025-10-01 18:18:10 +02:00
Aleksander Grygier
2a9b63383a
Improve code block color theming ( #16325 )
...
* feat: Improve code block theming
* chore: update webui build output
* chore: Update webui static build
2025-10-01 15:54:42 +02:00
Aleksander Grygier
4f1575921c
Add optional setting for showing "Model used:" information ( #16337 )
...
* feat: Add a setting to include model name used to generate the message
* feat: UI improvements
* feat: Save model info along with the database message entry creation
* chore: Build webui static output
2025-10-01 12:08:16 +02:00
Aleksander Grygier
aa9538a63a
webui: Remove running llama-server within WebUI dev.sh script ( #16363 )
2025-10-01 08:40:26 +03:00
Pascal
16b0ca0d2e
Chatapi ignore empty sampling ( #16330 )
...
* fix: skip empty sampling fields instead of coercing to 0 in chat API options
* chore: update webui build output
2025-09-30 19:18:54 +02:00
Concedo
20c802a198
Merge branch 'upstream' into concedo_experimental
...
# Conflicts:
# .github/workflows/build.yml
# CODEOWNERS
# ggml/CMakeLists.txt
# ggml/src/ggml-cpu/CMakeLists.txt
# ggml/src/ggml-cpu/kleidiai/kleidiai.cpp
# scripts/sync-ggml.last
# tests/test-backend-ops.cpp
2025-09-30 22:28:53 +08:00
Pascal
5f7e166cbf
Fix thinking blocks with quotes + add handling [THINK]...[/THINK] blocks ( #16326 )
...
* fix: prevent reasoning blocks with quotes from being truncated
* chore: update webui build output
* feat: Improve thinking content parsing
* test: Adds ChatMessage component stories for different thinking blocks
* chore: update webui build output
* fix: ChatMessage story fix
---------
Co-authored-by: Aleksander Grygier <aleksander.grygier@gmail.com>
2025-09-29 18:49:47 +02:00
Concedo
4f2b951547
Merge branch 'upstream' into concedo_experimental
...
# Conflicts:
# .devops/musa.Dockerfile
# .github/workflows/build-linux-cross.yml
# .github/workflows/build-riscv-native.yml
# .github/workflows/build.yml
# .github/workflows/docker.yml
# CODEOWNERS
# ci/run.sh
# ggml/CMakeLists.txt
# ggml/src/ggml-blas/CMakeLists.txt
# ggml/src/ggml-cpu/CMakeLists.txt
# scripts/sync-ggml.last
# tests/test-backend-ops.cpp
# tools/perplexity/perplexity.cpp
# tools/server/README.md
2025-09-30 00:36:38 +08:00
Aleksander Grygier
3a2bdcda0b
Improve Mobile UI for dialogs and action dropdowns ( #16222 )
...
* fix: Always show conversation item actions
* feat: Improve Alert Dialog and Dialog mobile UI
* feat: Add settings reset to default confirmation
* fix: Close Edit dialog on save
* chore: update webui build output
* webui: implement proper z-index system and scroll management
- Add CSS variable for centralized z-index control
- Fix dropdown positioning with Settings dialog conflicts
- Prevent external scroll interference with proper event handling
- Clean up hardcoded z-index values for maintainable architecture
* webui: ensured the settings dialog enforces dynamic viewport height on mobile while retaining existing desktop sizing overrides
* feat: Use `dvh` instead of computed px height for dialogs max height on mobile
* chore: update webui build output
* feat: Improve Settings fields UI
* chore: update webui build output
* chore: update webui build output
---------
Co-authored-by: Pascal <admin@serveurperso.com>
2025-09-29 10:37:20 +02:00
Pascal
66bb7985c3
fix: preserved zero values in chat settings inputs and textareas by switching to nullish coalescing for field values and default placeholders ( #16312 )
2025-09-29 09:08:41 +02:00
Imad Saddik
2811c65286
Fixed a few typos in the README of the LLaMA.cpp HTTP Server [no ci] ( #16297 )
2025-09-28 13:04:46 +02:00
Aleksander Grygier
4807e8f96a
Show message actions by default ( #16289 )
2025-09-27 19:56:40 +02:00
Adrien Gallouët
234e2ff8ed
server : remove old LLAMA_SERVER_SSL ( #16290 )
...
Signed-off-by: Adrien Gallouët <angt@huggingface.co>
2025-09-27 19:17:08 +03:00
Concedo
b120e107f9
Merge branch 'upstream' into concedo_experimental
...
# Conflicts:
# .clang-tidy
# .devops/musa.Dockerfile
# .github/workflows/build-linux-cross.yml
# .github/workflows/build.yml
# .github/workflows/docker.yml
# .gitignore
# CODEOWNERS
# CONTRIBUTING.md
# README.md
# build-xcframework.sh
# ci/README-MUSA.md
# ci/run.sh
# common/CMakeLists.txt
# docs/docker.md
# examples/CMakeLists.txt
# examples/eval-callback/CMakeLists.txt
# examples/model-conversion/Makefile
# examples/model-conversion/README.md
# examples/model-conversion/logits.cpp
# examples/model-conversion/scripts/causal/compare-logits.py
# examples/model-conversion/scripts/causal/run-org-model.py
# examples/model-conversion/scripts/embedding/compare-embeddings-logits.sh
# examples/model-conversion/scripts/embedding/run-converted-model.sh
# examples/model-conversion/scripts/embedding/run-original-model.py
# examples/model-conversion/scripts/utils/check-nmse.py
# examples/model-conversion/scripts/utils/inspect-org-model.py
# examples/model-conversion/scripts/utils/semantic_check.py
# ggml/CMakeLists.txt
# ggml/include/ggml-zdnn.h
# ggml/src/ggml-opencl/ggml-opencl.cpp
# ggml/src/ggml-opencl/kernels/set_rows.cl
# ggml/src/ggml-rpc/ggml-rpc.cpp
# ggml/src/ggml-sycl/ggml-sycl.cpp
# ggml/src/ggml-sycl/set_rows.cpp
# ggml/src/ggml-webgpu/ggml-webgpu.cpp
# ggml/src/ggml-zdnn/ggml-zdnn.cpp
# tests/CMakeLists.txt
# tests/test-backend-ops.cpp
# tests/test-quantize-perf.cpp
# tests/test-tokenizers-repo.sh
# tools/perplexity/perplexity.cpp
# tools/server/tests/README.md
2025-09-27 17:09:14 +08:00
Aleksander Grygier
807e8c6d31
Enhance text file detection logic for file attachments ( #16199 )
...
* feat: Enhances text file detection logic
* chore: Build static `webui` output
* chore: update webui build output
2025-09-26 19:25:29 +02:00
Aleksander Grygier
1a18927894
Allow viewing conversations even when llama server is down ( #16255 )
...
* webui: allow viewing conversations and sending messages even if llama-server is down
- Cached llama.cpp server properties in browser localStorage on startup, persisting successful fetches and reloading them when refresh attempts fail so the chat UI continues to render while the backend is unavailable.
- Cleared the stored server properties when resetting the store to prevent stale capability data after cache-backed operation.
- Kept the original error-splash behavior when no cached props exist so fresh installs still surface a clear failure state instead of rendering stale data.
* feat: Add UI for `props` endpoint unavailable + cleanup logic
* webui: extend cached props fallback to offline errors
Treat connection failures (refused, DNS, timeout, fetch) the same way as
server 5xx so the warning banner shows up when cache is available, instead
of falling back to a full error screen.
* webui: Left the chat form enabled when a server warning is present so operators can keep sending messages
e.g., to restart the backend over llama-swap, even while cached /props data is in use
* chore: update webui build output
---------
Co-authored-by: Pascal <admin@serveurperso.com>
2025-09-26 18:35:42 +02:00
Isaac McFadyen
e0539eb6ae
webui: switch to hash-based routing (alternative of #16079 ) ( #16157 )
...
* Switched web UI to hash-based routing
* Added hash to missed goto function call
* Removed outdated SPA handling code
* Fixed broken sidebar home link
2025-09-26 18:36:48 +03:00
Aleksander Grygier
5d0a40f390
Always show message actions for mobile UI + improvements for user message sizing ( #16076 )
2025-09-26 15:59:07 +02:00
Daniel Bevenius
d0991da39d
server : add support for external server for tests ( #16243 )
...
This commit adds support for using an externally started llama-server
instance for the server tests. This can be enabled by setting the
DEBUG_EXTERNAL environment variable.
The motivation for this is to allow debugging of the server itself
when investigating a test failure. Instructions for how to do this are
added to the README.md file in the tests directory.
2025-09-25 11:36:47 +02:00
Douglas Hanley
b5bd037832
llama : add support for qwen3 reranker ( #15824 )
2025-09-25 11:53:09 +03:00
Concedo
efe546390b
Merge branch 'upstream' into concedo_experimental
...
# Conflicts:
# .github/workflows/build.yml
# CODEOWNERS
# CONTRIBUTING.md
# README.md
# ci/run.sh
# examples/embedding/README.md
# tests/test-backend-ops.cpp
2025-09-22 21:25:29 +08:00
Quentin Bramas
138c87ce8b
webui : fix handling incomplete chunks ( #16107 )
2025-09-22 11:53:13 +03:00
Concedo
0dc6b9f418
Merge branch 'upstream' into concedo_experimental
...
# Conflicts:
# ggml/CMakeLists.txt
# ggml/src/CMakeLists.txt
# ggml/src/ggml-cann/ggml-cann.cpp
# ggml/src/ggml-cpu/amx/amx.cpp
# ggml/src/ggml-cuda/CMakeLists.txt
# ggml/src/ggml-opencl/CMakeLists.txt
# ggml/src/ggml-opencl/ggml-opencl.cpp
# ggml/src/ggml-opencl/kernels/cvt.cl
# ggml/src/ggml-rpc/ggml-rpc.cpp
# ggml/src/ggml-sycl/ggml-sycl.cpp
# ggml/src/ggml-webgpu/ggml-webgpu.cpp
# ggml/src/ggml-webgpu/wgsl-shaders/embed_wgsl.py
# ggml/src/ggml-webgpu/wgsl-shaders/mul_mat.tmpl.wgsl
# ggml/src/ggml-webgpu/wgsl-shaders/set_rows.wgsl
# ggml/src/ggml-zdnn/ggml-zdnn.cpp
# scripts/sync-ggml.last
# tests/test-backend-ops.cpp
# tests/test-chat.cpp
# tools/llama-bench/README.md
# tools/llama-bench/llama-bench.cpp
2025-09-21 11:38:47 +08:00
Concedo
326f6f3fad
not sure if working on metal
2025-09-21 11:35:02 +08:00
Concedo
3e72aaff5b
Merge commit ' 8f8f2274ee' into concedo_experimental
...
# Conflicts:
# .devops/rocm.Dockerfile
# .github/workflows/build.yml
# .github/workflows/release.yml
# CMakeLists.txt
# examples/simple/simple.cpp
# ggml/src/ggml-cann/common.h
# ggml/src/ggml-cann/ggml-cann.cpp
# ggml/src/ggml-opencl/kernels/tsembd.cl
# ggml/src/ggml-sycl/binbcast.cpp
# ggml/src/ggml-sycl/binbcast.hpp
# ggml/src/ggml-sycl/ggml-sycl.cpp
# ggml/src/ggml-sycl/tsembd.cpp
# ggml/src/ggml-zdnn/ggml-zdnn.cpp
# src/llama-model.cpp
# tools/batched-bench/CMakeLists.txt
# tools/cvector-generator/CMakeLists.txt
# tools/export-lora/CMakeLists.txt
# tools/gguf-split/CMakeLists.txt
# tools/imatrix/CMakeLists.txt
# tools/llama-bench/CMakeLists.txt
# tools/llama-bench/llama-bench.cpp
# tools/main/CMakeLists.txt
# tools/main/README.md
# tools/mtmd/CMakeLists.txt
# tools/perplexity/CMakeLists.txt
# tools/perplexity/perplexity.cpp
# tools/quantize/CMakeLists.txt
# tools/rpc/rpc-server.cpp
# tools/run/CMakeLists.txt
# tools/run/run.cpp
# tools/tokenize/CMakeLists.txt
# tools/tts/CMakeLists.txt
2025-09-21 08:58:23 +08:00
Benni
459c0c2c1a
server: fix SSE and OpenAI compatibility for error messages when streaming ( #16109 )
...
* server: fix SSE and OpenAI compatibility for error messages when streaming
* server: remove obsolete event parameter and use required data fieldname instead
2025-09-20 07:56:30 +02:00
Aleksander Grygier
4067f07fc5
feat: Improve mobile UI for Settings Dialog ( #16084 )
...
* feat: Improve mobile UI for Settings Dialog
* chore: update webui build output
* fix: Linting errors
* chore: update webui build output
2025-09-19 09:52:27 +02:00
Radoslav Gerganov
2b6b55a59f
server : include usage statistics only when user request them ( #16052 )
...
* server : include usage statistics only when user request them
When serving the OpenAI compatible API, we should check if
{"stream_options": {"include_usage": true} is set in the request when
deciding whether we should send usage statistics
closes : #16048
* add unit test
2025-09-18 10:36:57 +00:00
Aleksander Grygier
a7a98e0fff
SvelteKit-based WebUI ( #14839 )
2025-09-17 19:29:13 +02:00
Sigbjørn Skjæret
6c019cb04e
server : only attempt to enable thinking if using jinja ( #15967 )
2025-09-14 21:17:04 +02:00
Concedo
1dbd2fc259
Merge branch 'upstream' into concedo_experimental
...
# Conflicts:
# docs/build-s390x.md
# docs/ops.md
# docs/ops/zDNN.csv
# ggml/include/ggml-zdnn.h
# ggml/src/ggml-sycl/binbcast.cpp
# ggml/src/ggml-sycl/concat.cpp
# ggml/src/ggml-sycl/conv.cpp
# ggml/src/ggml-sycl/convert.cpp
# ggml/src/ggml-sycl/cpy.cpp
# ggml/src/ggml-sycl/dmmv.cpp
# ggml/src/ggml-sycl/dpct/helper.hpp
# ggml/src/ggml-sycl/element_wise.cpp
# ggml/src/ggml-sycl/getrows.cpp
# ggml/src/ggml-sycl/ggml-sycl.cpp
# ggml/src/ggml-sycl/gla.cpp
# ggml/src/ggml-sycl/im2col.cpp
# ggml/src/ggml-sycl/mmq.cpp
# ggml/src/ggml-sycl/mmvq.cpp
# ggml/src/ggml-sycl/norm.cpp
# ggml/src/ggml-sycl/rope.cpp
# ggml/src/ggml-sycl/set_rows.cpp
# ggml/src/ggml-sycl/softmax.cpp
# ggml/src/ggml-sycl/tsembd.cpp
# ggml/src/ggml-sycl/wkv.cpp
# ggml/src/ggml-zdnn/ggml-zdnn-impl.h
# ggml/src/ggml-zdnn/ggml-zdnn.cpp
# tools/llama-bench/llama-bench.cpp
2025-09-13 12:25:30 +08:00
Georgi Gerganov
f088b6a84f
server : adjust prompt similarity thold + add logs ( #15913 )
...
ggml-ci
2025-09-12 17:02:55 +03:00
Concedo
6463f5c26b
Merge branch 'upstream' into concedo_experimental
...
# Conflicts:
# .github/workflows/build.yml
# .github/workflows/release.yml
# CONTRIBUTING.md
# docs/backend/CANN.md
# examples/eval-callback/eval-callback.cpp
# examples/model-conversion/requirements.txt
# examples/model-conversion/scripts/causal/run-org-model.py
# ggml/src/ggml-cann/aclnn_ops.cpp
# ggml/src/ggml-cann/common.h
# ggml/src/ggml-cann/ggml-cann.cpp
# ggml/src/ggml-cpu/CMakeLists.txt
# ggml/src/ggml-cpu/kleidiai/kleidiai.cpp
# ggml/src/ggml-cuda/CMakeLists.txt
# ggml/src/ggml-opencl/ggml-opencl.cpp
# ggml/src/ggml-rpc/ggml-rpc.cpp
# ggml/src/ggml-sycl/ggml-sycl.cpp
# ggml/src/ggml-webgpu/ggml-webgpu.cpp
# ggml/src/ggml-zdnn/ggml-zdnn.cpp
# models/templates/README.md
# requirements/requirements-convert_hf_to_gguf.txt
# requirements/requirements-convert_legacy_llama.txt
# requirements/requirements-tool_bench.txt
# tests/.gitignore
# tests/test-backend-ops.cpp
# tests/test-chat-parser.cpp
# tests/test-chat.cpp
# tests/test-json-schema-to-grammar.cpp
# tests/test-tokenizer-random.py
2025-09-11 22:34:45 +08:00
Concedo
5de51b77c1
Merge branch 'upstream' into concedo_experimental
...
# Conflicts:
# .github/workflows/close-issue.yml
# docs/build-s390x.md
# examples/convert-llama2c-to-ggml/convert-llama2c-to-ggml.cpp
# ggml/CMakeLists.txt
# ggml/src/ggml-cann/ggml-cann.cpp
# ggml/src/ggml-cpu/CMakeLists.txt
# ggml/src/ggml-cpu/kleidiai/kleidiai.cpp
# ggml/src/ggml-cuda/fattn-tile-f16.cu
# ggml/src/ggml-cuda/fattn.cu
# ggml/src/ggml-webgpu/ggml-webgpu.cpp
# scripts/tool_bench.py
# tests/test-backend-ops.cpp
# tools/batched-bench/batched-bench.cpp
# tools/server/README.md
2025-09-11 22:28:19 +08:00
Daniel Bevenius
70cd37dbbe
requirements : update transformers/torch for Embedding Gemma ( #15828 )
...
* requirements : update transformers/torch for Embedding Gemma
This commit updates the requirements to support converting
Embedding Gemma 300m models.
The motivation for this change is that during development I had a local
copy of the transformers package which is what I used for converting
the models. This was a mistake on my part and I should have also updated
my transformers version to the official release.
I had checked the requirements/requirements-convert_legacy_llama.txt
file and noted that the version was >=4.45.1,<5.0.0 and came to the
conculusion that no updated would be needed, this assumed that
Embedding Gemma would be in a transformers release at the time
Commit fb15d649ed ("llama : add support
for EmbeddingGemma 300m (#15798 )) was merged. So anyone wanting to
convert themselves would be able to do so. However, Embedding Gemma is
a preview release and this commit updates the requirements to use this
preview release.
* resolve additional python dependencies
* fix pyright errors in tokenizer test and remove unused import
2025-09-09 06:06:52 +02:00
Aldehir Rojas
7057faf64b
json : support enum values within allOf ( #15830 )
2025-09-08 16:14:32 -05:00