koboldcpp/tools/server
Concedo b6f6338bba Merge branch 'upstream' into concedo_experimental
# Conflicts:
#	.github/workflows/build-linux-cross.yml
#	.github/workflows/build.yml
#	CODEOWNERS
#	ggml/CMakeLists.txt
#	ggml/src/ggml-cuda/fattn.cu
#	ggml/src/ggml-webgpu/CMakeLists.txt
#	ggml/src/ggml-webgpu/ggml-webgpu.cpp
#	ggml/src/ggml-webgpu/wgsl-shaders/mul_mat.tmpl.wgsl
#	tests/test-backend-ops.cpp
#	tests/test-chat-template.cpp
#	tools/llama-bench/llama-bench.cpp
#	tools/rpc/README.md
#	tools/server/README.md
2025-10-09 01:33:27 +08:00
..
bench Merge branch 'upstream' into concedo_experimental 2025-08-23 11:35:28 +08:00
public webui : added download action (#13552) (#16282) 2025-10-07 11:11:08 +02:00
public_legacy json : support enum values within allOf (#15830) 2025-09-08 16:14:32 -05:00
public_simplechat Merge branch 'upstream' into concedo_experimental 2025-05-03 12:15:36 +08:00
tests Merge branch 'upstream' into concedo_experimental 2025-09-27 17:09:14 +08:00
themes Merge branch 'upstream' into concedo_experimental 2025-05-03 12:15:36 +08:00
webui webui : added download action (#13552) (#16282) 2025-10-07 11:11:08 +02:00
chat-llama2.sh scripts : make the shell scripts cross-platform (#14341) 2025-06-30 10:17:18 +02:00
chat.mjs llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
chat.sh scripts : make the shell scripts cross-platform (#14341) 2025-06-30 10:17:18 +02:00
server.cpp server : fix cancel pending task (#16467) 2025-10-08 11:20:18 +03:00
utils.hpp llama : add support for qwen3 reranker (#15824) 2025-09-25 11:53:09 +03:00