koboldcpp/tools/server
Concedo d5876024ec Merge commit 'f4586ee598' into concedo_experimental
# Conflicts:
#	README.md
#	docs/multimodal/minicpmo2.6.md
#	docs/multimodal/minicpmv2.6.md
#	ggml/src/ggml-cann/aclnn_ops.cpp
#	ggml/src/ggml-cann/ggml-cann.cpp
#	ggml/src/ggml-cpu/kleidiai/kleidiai.cpp
#	ggml/src/ggml-cuda/CMakeLists.txt
#	ggml/src/ggml-opencl/ggml-opencl.cpp
#	ggml/src/ggml-opencl/kernels/add.cl
#	ggml/src/ggml-sycl/ggml-sycl.cpp
#	tools/perplexity/perplexity.cpp
#	tools/server/README.md
2025-08-14 21:29:52 +08:00
..
bench Merge branch 'upstream' into concedo_experimental 2025-05-03 12:15:36 +08:00
public server : allow specifying reasoning_format in HTTP request (#15238) 2025-08-11 14:48:41 +02:00
public_legacy llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
public_simplechat Merge branch 'upstream' into concedo_experimental 2025-05-03 12:15:36 +08:00
tests Merge branch 'upstream' into concedo_experimental 2025-08-02 10:25:10 +08:00
themes Merge branch 'upstream' into concedo_experimental 2025-05-03 12:15:36 +08:00
webui server : allow specifying reasoning_format in HTTP request (#15238) 2025-08-11 14:48:41 +02:00
chat-llama2.sh scripts : make the shell scripts cross-platform (#14341) 2025-06-30 10:17:18 +02:00
chat.mjs llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
chat.sh scripts : make the shell scripts cross-platform (#14341) 2025-06-30 10:17:18 +02:00
server.cpp server : allow specifying reasoning_format in HTTP request (#15238) 2025-08-11 14:48:41 +02:00
utils.hpp scripts: benchmark for HTTP server throughput (#14668) 2025-07-14 13:14:30 +02:00