koboldcpp/tools/server
Concedo 6eea7b88d2 Merge branch 'upstream' into concedo_experimental
# Conflicts:
#	.github/workflows/build.yml
#	README.md
#	ggml/src/ggml-cann/ggml-cann.cpp
#	ggml/src/ggml-opencl/ggml-opencl.cpp
#	ggml/src/ggml-sycl/ggml-sycl.cpp
#	tests/test-backend-ops.cpp
#	tests/test-chat-template.cpp
2025-08-06 10:51:29 +08:00
..
bench Merge branch 'upstream' into concedo_experimental 2025-05-03 12:15:36 +08:00
public llama : add gpt-oss (#15091) 2025-08-05 22:10:36 +03:00
public_legacy llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
public_simplechat Merge branch 'upstream' into concedo_experimental 2025-05-03 12:15:36 +08:00
tests Merge branch 'upstream' into concedo_experimental 2025-08-02 10:25:10 +08:00
themes Merge branch 'upstream' into concedo_experimental 2025-05-03 12:15:36 +08:00
webui llama : add gpt-oss (#15091) 2025-08-05 22:10:36 +03:00
chat-llama2.sh scripts : make the shell scripts cross-platform (#14341) 2025-06-30 10:17:18 +02:00
chat.mjs llama : move end-user examples to tools directory (#13249) 2025-05-02 20:27:13 +02:00
chat.sh scripts : make the shell scripts cross-platform (#14341) 2025-06-30 10:17:18 +02:00
server.cpp server: enable token array inputs for OAI API (#15001) 2025-08-02 10:12:41 +02:00
utils.hpp scripts: benchmark for HTTP server throughput (#14668) 2025-07-14 13:14:30 +02:00