koboldcpp/tools/server
Concedo 2771e16fbc Merge branch 'upstream' into concedo_experimental
# Conflicts:
#	.devops/intel.Dockerfile
#	.devops/nix/package.nix
#	.gitignore
#	docs/backend/SYCL.md
#	docs/ops.md
#	docs/ops/SYCL.csv
#	ggml/CMakeLists.txt
#	ggml/src/ggml-cuda/fattn.cu
#	ggml/src/ggml-cuda/ggml-cuda.cu
#	ggml/src/ggml-sycl/CMakeLists.txt
#	ggml/src/ggml-sycl/common.hpp
#	ggml/src/ggml-sycl/convert.cpp
#	ggml/src/ggml-sycl/dequantize.hpp
#	ggml/src/ggml-sycl/fattn-common.hpp
#	ggml/src/ggml-sycl/getrows.cpp
#	ggml/src/ggml-sycl/ggml-sycl.cpp
#	ggml/src/ggml-sycl/im2col.cpp
#	ggml/src/ggml-sycl/im2col.hpp
#	ggml/src/ggml-sycl/mmvq.cpp
#	ggml/src/ggml-sycl/quants.hpp
#	ggml/src/ggml-sycl/vecdotq.hpp
#	ggml/src/ggml-virtgpu/ggml-backend-device.cpp
#	scripts/sync-ggml.last
#	scripts/sync_vendor.py
#	tests/test-backend-ops.cpp
2026-05-11 16:18:28 +08:00
..
bench Merge branch 'upstream' into concedo_experimental 2026-03-22 23:39:13 +08:00
public webui: fix LLM title generation for agentic conversations (#22840) 2026-05-08 16:36:04 +02:00
tests Merge branch 'upstream' into concedo_experimental 2026-05-11 16:18:28 +08:00
webui Merge commit '66001722aa' into concedo_experimental 2026-05-11 15:40:10 +08:00
chat-llama2.sh scripts : make the shell scripts cross-platform (#14341) 2025-06-30 10:17:18 +02:00
chat.mjs llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
chat.sh scripts : make the shell scripts cross-platform (#14341) 2025-06-30 10:17:18 +02:00
README-dev.md server: (webui) no more gzip compression (#21073) 2026-03-31 15:44:26 +02:00
server-chat.cpp server: (router) Forward form-data to model server (Fixes #22044) (#22118) 2026-04-27 23:55:00 +02:00
server-chat.h server: (router) Forward form-data to model server (Fixes #22044) (#22118) 2026-04-27 23:55:00 +02:00
server-common.cpp parser: fix structured output bug (#22302) 2026-04-24 23:19:55 +02:00
server-common.h common/chat, server: refactor, move all conversion functions to common, add tests (#20690) 2026-04-22 10:28:45 +02:00
server-context.cpp backend sampling: support returning post-sampling probs (#22622) 2026-05-10 19:12:02 +02:00
server-context.h server: (router) expose child model info from router's /v1/models (#22683) 2026-05-08 14:42:15 +02:00
server-cors-proxy.h server: (router) Forward form-data to model server (Fixes #22044) (#22118) 2026-04-27 23:55:00 +02:00
server-http.cpp server: support Vertex AI compatible API (#22545) 2026-05-08 15:23:04 +02:00
server-http.h server: support Vertex AI compatible API (#22545) 2026-05-08 15:23:04 +02:00
server-models.cpp server: (router) expose child model info from router's /v1/models (#22683) 2026-05-08 14:42:15 +02:00
server-models.h server: (router) expose child model info from router's /v1/models (#22683) 2026-05-08 14:42:15 +02:00
server-queue.cpp server : print warning when HTTP timeout exceeded (#22907) 2026-05-10 22:00:18 +03:00
server-queue.h server: allow router to report child instances sleep status (#20849) 2026-03-22 18:33:52 +01:00
server-task.cpp spec : refactor params (#22397) 2026-04-28 09:07:33 +03:00
server-task.h server : speculative checkpointing (#19493) 2026-04-19 10:24:06 +03:00
server-tools.cpp server : validate --tools CLI argument against known tool names (#22538) 2026-05-05 06:35:27 +03:00
server-tools.h server: add built-in tools backend support (#20898) 2026-03-27 10:07:11 +01:00
server.cpp server: support Vertex AI compatible API (#22545) 2026-05-08 15:23:04 +02:00