Concedo
|
a530afa1e4
|
Merge commit '280345968d ' into concedo_experimental
# Conflicts:
# .devops/full-cuda.Dockerfile
# .devops/llama-cpp-cuda.srpm.spec
# .devops/main-cuda.Dockerfile
# .devops/nix/package.nix
# .devops/server-cuda.Dockerfile
# .github/workflows/build.yml
# CMakeLists.txt
# Makefile
# README.md
# ci/run.sh
# docs/token_generation_performance_tips.md
# flake.lock
# llama.cpp
# scripts/LlamaConfig.cmake.in
# scripts/compare-commits.sh
# scripts/server-llm.sh
# tests/test-quantize-fns.cpp
|
2024-04-07 20:27:17 +08:00 |
|
Concedo
|
680a41ed71
|
refactor identifiers
|
2024-01-20 17:26:11 +08:00 |
|
Concedo
|
693f3f0b00
|
try to use allocator for cuda ggml v3
|
2024-01-20 12:53:31 +08:00 |
|
Concedo
|
db14de5c32
|
fossilize ggml library ver 3, to support ggjtv3
|
2024-01-20 10:49:25 +08:00 |
|
Concedo
|
230a638512
|
Merge branch 'master' into concedo_experimental
# Conflicts:
# .github/workflows/docker.yml
# CMakeLists.txt
# Makefile
# README.md
# llama.cpp
# tests/test-grad0.cpp
|
2023-12-22 14:40:13 +08:00 |
|
Concedo
|
a6eb9b8010
|
Fix GPT2 not loading due to graph too small
|
2023-11-26 23:06:42 +08:00 |
|
Concedo
|
dc4078c039
|
fixed segfault with all non-gguf models
|
2023-11-20 22:31:56 +08:00 |
|
Concedo
|
45ad1b97f8
|
max nodes 8192
|
2023-11-18 11:02:35 +08:00 |
|
Concedo
|
35a97e14b2
|
Merge branch 'master' into concedo_experimental
# Conflicts:
# Makefile
# README.md
# docs/token_generation_performance_tips.md
# grammars/README.md
# scripts/sync-ggml.sh
# tests/CMakeLists.txt
# tests/test-grad0.cpp
# tests/test-opt.cpp
|
2023-11-15 16:59:53 +08:00 |
|
Concedo
|
8c14c81b33
|
hopefully this fixes the dotnet nonsense
|
2023-11-03 11:23:56 +08:00 |
|
Concedo
|
42eabf2f2f
|
rope fixes
|
2023-11-02 20:41:16 +08:00 |
|
Concedo
|
bc4ff72317
|
not working merge
|
2023-11-02 17:52:40 +08:00 |
|
Concedo
|
cc5b282350
|
Merge branch 'master' into concedo_experimental
# Conflicts:
# CMakeLists.txt
# Makefile
# build.zig
# flake.lock
# flake.nix
# ggml.c
|
2023-10-31 20:44:04 +08:00 |
|
Concedo
|
5db89b90b7
|
Merge branch 'master' into concedo_experimental
# Conflicts:
# .gitignore
# CMakeLists.txt
# Makefile
# README.md
# build.zig
# ggml-opencl.cpp
# tests/CMakeLists.txt
# tests/test-double-float.cpp
# tests/test-sampling.cpp
|
2023-10-25 23:58:15 +08:00 |
|
Concedo
|
c9983a72d6
|
prevent lora with clblast
|
2023-10-25 15:18:03 +08:00 |
|
Concedo
|
f288c6b5e3
|
Merge branch 'master' into concedo_experimental
# Conflicts:
# CMakeLists.txt
# Makefile
# build.zig
# scripts/sync-ggml.sh
|
2023-10-10 00:09:46 +08:00 |
|
Concedo
|
202e28a76a
|
do not offload rope for old cublas (+1 squashed commits)
Squashed commits:
[ca72a66f] fix allocr (+1 squashed commits)
Squashed commits:
[22a0e30e] updated lite
|
2023-09-30 18:18:36 +08:00 |
|
Concedo
|
b84e210f0d
|
merge new rope param nonsense
|
2023-09-30 11:33:30 +08:00 |
|
Concedo
|
d7fed4732f
|
fix for typical sampler
|
2023-09-01 15:24:00 +08:00 |
|
Concedo
|
0c3a265187
|
fixed incorrect buffer size values
|
2023-09-01 01:31:09 +08:00 |
|
vxiiduu
|
f2985a070b
|
Add support for 34B GGML models
|
2023-09-01 01:29:09 +10:00 |
|
askmyteapot
|
8263fd7bdb
|
Update llama_v3.cpp (#393)
Fixing C2065 compiler error.
Missed '3' on 3 separate identifiers (kB > kB3, MB > MB3)
|
2023-08-23 22:15:48 +08:00 |
|
Concedo
|
af170fc2db
|
Merge branch 'master' into concedo_experimental
# Conflicts:
# README.md
# llama.cpp
# scripts/sync-ggml.sh
# tests/test-tokenizer-0.cpp
|
2023-08-23 17:08:09 +08:00 |
|
Concedo
|
981c9131f0
|
gguf for llama is working
|
2023-08-23 16:07:07 +08:00 |
|
Concedo
|
39cc83e8c9
|
incomplete merge, compiles but generates rubbish
|
2023-08-22 23:12:47 +08:00 |
|