koboldcpp/tools/server
Concedo f7923b261f need to fix cuda compile. Merge branch 'upstream' into concedo_experimental
# Conflicts:
#	.github/workflows/python-type-check.yml
#	examples/speculative-simple/README.md
#	examples/speculative-simple/speculative-simple.cpp
#	ggml/src/ggml-cuda/im2col.cu
#	ggml/src/ggml-opencl/CMakeLists.txt
#	ggml/src/ggml-opencl/ggml-opencl.cpp
#	ggml/src/ggml-opencl/kernels/cvt.cl
#	tests/test-backend-ops.cpp
#	tools/cli/README.md
#	tools/mtmd/CMakeLists.txt
#	tools/server/README.md
2026-05-12 20:47:07 +08:00
..
bench Merge branch 'upstream' into concedo_experimental 2026-03-22 23:39:13 +08:00
public webui: fix LLM title generation for agentic conversations (#22840) 2026-05-08 16:36:04 +02:00
tests need to fix cuda compile. Merge branch 'upstream' into concedo_experimental 2026-05-12 20:47:07 +08:00
webui Merge commit '66001722aa' into concedo_experimental 2026-05-11 15:40:10 +08:00
chat-llama2.sh scripts : make the shell scripts cross-platform (#14341) 2025-06-30 10:17:18 +02:00
chat.mjs llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
chat.sh scripts : make the shell scripts cross-platform (#14341) 2025-06-30 10:17:18 +02:00
README-dev.md server: (webui) no more gzip compression (#21073) 2026-03-31 15:44:26 +02:00
server-chat.cpp server: (router) Forward form-data to model server (Fixes #22044) (#22118) 2026-04-27 23:55:00 +02:00
server-chat.h server: (router) Forward form-data to model server (Fixes #22044) (#22118) 2026-04-27 23:55:00 +02:00
server-common.cpp parser: fix structured output bug (#22302) 2026-04-24 23:19:55 +02:00
server-common.h common/chat, server: refactor, move all conversion functions to common, add tests (#20690) 2026-04-22 10:28:45 +02:00
server-context.cpp docs: fix metrics endpoint description in server README (#22879) 2026-05-11 18:32:26 +02:00
server-context.h server: (router) expose child model info from router's /v1/models (#22683) 2026-05-08 14:42:15 +02:00
server-cors-proxy.h server: (router) Forward form-data to model server (Fixes #22044) (#22118) 2026-04-27 23:55:00 +02:00
server-http.cpp server: support Vertex AI compatible API (#22545) 2026-05-08 15:23:04 +02:00
server-http.h server: support Vertex AI compatible API (#22545) 2026-05-08 15:23:04 +02:00
server-models.cpp server: (router) expose child model info from router's /v1/models (#22683) 2026-05-08 14:42:15 +02:00
server-models.h server: (router) expose child model info from router's /v1/models (#22683) 2026-05-08 14:42:15 +02:00
server-queue.cpp server : print warning when HTTP timeout exceeded (#22907) 2026-05-10 22:00:18 +03:00
server-queue.h server: allow router to report child instances sleep status (#20849) 2026-03-22 18:33:52 +01:00
server-task.cpp spec : parallel drafting support (#22838) 2026-05-11 19:09:43 +03:00
server-task.h spec : parallel drafting support (#22838) 2026-05-11 19:09:43 +03:00
server-tools.cpp server : validate --tools CLI argument against known tool names (#22538) 2026-05-05 06:35:27 +03:00
server-tools.h server: add built-in tools backend support (#20898) 2026-03-27 10:07:11 +01:00
server.cpp server: support Vertex AI compatible API (#22545) 2026-05-08 15:23:04 +02:00