koboldcpp/examples/server
Concedo bb13925f39 Merge branch 'upstream' into concedo_experimental
# Conflicts:
#	CMakePresets.json
#	Makefile
#	Package.swift
#	ci/run.sh
#	common/CMakeLists.txt
#	examples/CMakeLists.txt
#	flake.lock
#	ggml/src/CMakeLists.txt
#	ggml/src/ggml-backend.cpp
#	ggml/src/ggml.c
#	pocs/vdot/q8dot.cpp
#	pocs/vdot/vdot.cpp
#	tests/test-backend-ops.cpp
#	tests/test-grad0.cpp
#	tests/test-quantize-fns.cpp
#	tests/test-quantize-perf.cpp
#	tests/test-rope.cpp
2024-11-04 16:54:53 +08:00
..
bench Merge branch 'upstream' into concedo_experimental 2024-09-19 14:53:57 +08:00
public llama : remove Tail-Free sampling (#10071) 2024-10-29 10:42:05 +02:00
public_simplechat Merge commit 'df270ef745' into concedo_experimental 2024-09-09 17:10:08 +08:00
tests note: also has support for completion tokens count 2024-11-01 00:44:14 +08:00
themes note: also has support for completion tokens count 2024-11-01 00:44:14 +08:00
chat-llama2.sh chmod : make scripts executable (#2675) 2023-08-23 17:29:09 +03:00
chat.mjs json-schema-to-grammar improvements (+ added to server) (#5978) 2024-03-21 11:50:43 +00:00
chat.sh server : fix context shift (#5195) 2024-01-30 20:17:30 +02:00
deps.sh build: generate hex dump of server assets during build (#6661) 2024-04-21 18:48:53 +01:00
httplib.h Server: version bump for httplib and json (#6169) 2024-03-20 13:30:36 +01:00
server.cpp Merge branch 'upstream' into concedo_experimental 2024-11-04 16:54:53 +08:00
utils.hpp server : fix slot selection by lru (#10126) 2024-11-02 18:34:56 +02:00