Commit graph

25 commits

Author SHA1 Message Date
Concedo
a530afa1e4 Merge commit '280345968d' into concedo_experimental
# Conflicts:
#	.devops/full-cuda.Dockerfile
#	.devops/llama-cpp-cuda.srpm.spec
#	.devops/main-cuda.Dockerfile
#	.devops/nix/package.nix
#	.devops/server-cuda.Dockerfile
#	.github/workflows/build.yml
#	CMakeLists.txt
#	Makefile
#	README.md
#	ci/run.sh
#	docs/token_generation_performance_tips.md
#	flake.lock
#	llama.cpp
#	scripts/LlamaConfig.cmake.in
#	scripts/compare-commits.sh
#	scripts/server-llm.sh
#	tests/test-quantize-fns.cpp
2024-04-07 20:27:17 +08:00
Concedo
680a41ed71 refactor identifiers 2024-01-20 17:26:11 +08:00
Concedo
693f3f0b00 try to use allocator for cuda ggml v3 2024-01-20 12:53:31 +08:00
Concedo
db14de5c32 fossilize ggml library ver 3, to support ggjtv3 2024-01-20 10:49:25 +08:00
Concedo
230a638512 Merge branch 'master' into concedo_experimental
# Conflicts:
#	.github/workflows/docker.yml
#	CMakeLists.txt
#	Makefile
#	README.md
#	llama.cpp
#	tests/test-grad0.cpp
2023-12-22 14:40:13 +08:00
Concedo
a6eb9b8010 Fix GPT2 not loading due to graph too small 2023-11-26 23:06:42 +08:00
Concedo
dc4078c039 fixed segfault with all non-gguf models 2023-11-20 22:31:56 +08:00
Concedo
45ad1b97f8 max nodes 8192 2023-11-18 11:02:35 +08:00
Concedo
35a97e14b2 Merge branch 'master' into concedo_experimental
# Conflicts:
#	Makefile
#	README.md
#	docs/token_generation_performance_tips.md
#	grammars/README.md
#	scripts/sync-ggml.sh
#	tests/CMakeLists.txt
#	tests/test-grad0.cpp
#	tests/test-opt.cpp
2023-11-15 16:59:53 +08:00
Concedo
8c14c81b33 hopefully this fixes the dotnet nonsense 2023-11-03 11:23:56 +08:00
Concedo
42eabf2f2f rope fixes 2023-11-02 20:41:16 +08:00
Concedo
bc4ff72317 not working merge 2023-11-02 17:52:40 +08:00
Concedo
cc5b282350 Merge branch 'master' into concedo_experimental
# Conflicts:
#	CMakeLists.txt
#	Makefile
#	build.zig
#	flake.lock
#	flake.nix
#	ggml.c
2023-10-31 20:44:04 +08:00
Concedo
5db89b90b7 Merge branch 'master' into concedo_experimental
# Conflicts:
#	.gitignore
#	CMakeLists.txt
#	Makefile
#	README.md
#	build.zig
#	ggml-opencl.cpp
#	tests/CMakeLists.txt
#	tests/test-double-float.cpp
#	tests/test-sampling.cpp
2023-10-25 23:58:15 +08:00
Concedo
c9983a72d6 prevent lora with clblast 2023-10-25 15:18:03 +08:00
Concedo
f288c6b5e3 Merge branch 'master' into concedo_experimental
# Conflicts:
#	CMakeLists.txt
#	Makefile
#	build.zig
#	scripts/sync-ggml.sh
2023-10-10 00:09:46 +08:00
Concedo
202e28a76a do not offload rope for old cublas (+1 squashed commits)
Squashed commits:

[ca72a66f] fix allocr (+1 squashed commits)

Squashed commits:

[22a0e30e] updated lite
2023-09-30 18:18:36 +08:00
Concedo
b84e210f0d merge new rope param nonsense 2023-09-30 11:33:30 +08:00
Concedo
d7fed4732f fix for typical sampler 2023-09-01 15:24:00 +08:00
Concedo
0c3a265187 fixed incorrect buffer size values 2023-09-01 01:31:09 +08:00
vxiiduu
f2985a070b
Add support for 34B GGML models 2023-09-01 01:29:09 +10:00
askmyteapot
8263fd7bdb
Update llama_v3.cpp (#393)
Fixing C2065 compiler error. 
Missed '3' on 3 separate identifiers (kB > kB3, MB > MB3)
2023-08-23 22:15:48 +08:00
Concedo
af170fc2db Merge branch 'master' into concedo_experimental
# Conflicts:
#	README.md
#	llama.cpp
#	scripts/sync-ggml.sh
#	tests/test-tokenizer-0.cpp
2023-08-23 17:08:09 +08:00
Concedo
981c9131f0 gguf for llama is working 2023-08-23 16:07:07 +08:00
Concedo
39cc83e8c9 incomplete merge, compiles but generates rubbish 2023-08-22 23:12:47 +08:00