koboldcpp/tools/server
Concedo 868cb6aff7 Merge commit 'e121edc432' into concedo_experimental
# Conflicts:
#	.github/workflows/release.yml
#	common/CMakeLists.txt
#	docs/function-calling.md
#	ggml/src/ggml-sycl/binbcast.cpp
#	models/templates/README.md
#	scripts/tool_bench.py
#	src/llama-kv-cache.cpp
#	tests/CMakeLists.txt
#	tests/test-chat.cpp
#	tools/mtmd/clip.h
#	tools/rpc/rpc-server.cpp
#	tools/server/README.md
2025-05-28 00:20:45 +08:00
..
bench Merge branch 'upstream' into concedo_experimental 2025-05-03 12:15:36 +08:00
public webui : bump max upload file size to 500MB (#13779) 2025-05-25 18:02:18 +01:00
public_legacy llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
public_simplechat Merge branch 'upstream' into concedo_experimental 2025-05-03 12:15:36 +08:00
tests Merge commit 'e121edc432' into concedo_experimental 2025-05-28 00:20:45 +08:00
themes Merge branch 'upstream' into concedo_experimental 2025-05-03 12:15:36 +08:00
webui webui : bump max upload file size to 500MB (#13779) 2025-05-25 18:02:18 +01:00
chat-llama2.sh llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
chat.mjs llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
chat.sh llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
httplib.h llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
server.cpp server: add --reasoning-budget 0 to disable thinking (incl. qwen3 w/ enable_thinking:false) (#13771) 2025-05-26 00:30:51 +01:00
utils.hpp server: add --reasoning-budget 0 to disable thinking (incl. qwen3 w/ enable_thinking:false) (#13771) 2025-05-26 00:30:51 +01:00