koboldcpp/include
Concedo ce7f9c9a2c Merge branch 'upstream' into concedo_experimental
# Conflicts:
#	.devops/full-rocm.Dockerfile
#	.devops/llama-cli-rocm.Dockerfile
#	.devops/llama-server-rocm.Dockerfile
#	.github/workflows/build.yml
#	.github/workflows/python-type-check.yml
#	CMakeLists.txt
#	CONTRIBUTING.md
#	README.md
#	ci/run.sh
#	examples/embedding/embedding.cpp
#	examples/server/README.md
#	flake.lock
#	ggml/include/ggml.h
#	ggml/src/ggml.c
#	requirements/requirements-convert_legacy_llama.txt
#	scripts/sync-ggml.last
#	src/llama-vocab.cpp
#	src/llama.cpp
#	tests/test-backend-ops.cpp
#	tests/test-grad0.cpp
#	tests/test-tokenizer-0.cpp
2024-10-02 01:00:57 +08:00
..
CL wip dont use 2023-04-21 00:35:54 +08:00
vulkan changes required to get vulkan working on windows 2024-01-25 18:29:45 +08:00
cblas.h wip dont use 2023-04-21 00:35:54 +08:00
clblast.h Revert "clblast up ver" 2024-02-21 14:35:38 +08:00
clblast_c.h Revert "clblast up ver" 2024-02-21 14:35:38 +08:00
clblast_half.h upgraded clblast 2023-05-25 10:18:12 +08:00
clblast_netlib_c.h Not working, don't use. testing a merge 2023-05-16 12:33:24 +08:00
llama.h Merge branch 'upstream' into concedo_experimental 2024-10-02 01:00:57 +08:00
openblas_config.h wip dont use 2023-04-21 00:35:54 +08:00