Concedo
|
dbe72b959e
|
tidy up and refactor code to support old flags
|
2024-05-10 16:50:53 +08:00 |
|
Concedo
|
caf9f21888
|
update lite
|
2024-05-09 15:03:52 +08:00 |
|
Concedo
|
bd0ef8f366
|
add characterhub.org support
|
2024-05-08 19:53:18 +08:00 |
|
Concedo
|
165a56088b
|
Merge branch 'upstream' into concedo_experimental
# Conflicts:
# ggml.c
# scripts/compare-llama-bench.py
# tests/test-backend-ops.cpp
|
2024-05-08 18:19:28 +08:00 |
|
Concedo
|
173c7272d5
|
EOS bypass mode added
|
2024-05-06 18:01:49 +08:00 |
|
Concedo
|
3667cc0113
|
fixed stableui btn (+4 squashed commit)
Squashed commit:
[1d4714f1] update default amount to gen
[6eacba33] updated lite
[033589af] added first ver sdui
[16f66d57] updated lite
|
2024-05-06 00:55:16 +08:00 |
|
Concedo
|
17a24d753c
|
Merge branch 'upstream' into concedo_experimental
# Conflicts:
# .devops/main-intel.Dockerfile
# .devops/main-vulkan.Dockerfile
# .devops/server-intel.Dockerfile
# .devops/server-vulkan.Dockerfile
# .github/workflows/bench.yml
# .github/workflows/build.yml
# .github/workflows/python-lint.yml
# .github/workflows/server.yml
# .gitignore
# Makefile
# README-sycl.md
# README.md
# ci/run.sh
# flake.lock
# llama.cpp
# models/ggml-vocab-falcon.gguf
# models/ggml-vocab-llama-spm.gguf
# models/ggml-vocab-mpt.gguf
# models/ggml-vocab-stablelm.gguf
# models/ggml-vocab-starcoder.gguf
# requirements.txt
# scripts/check-requirements.sh
# tests/CMakeLists.txt
# tests/test-backend-ops.cpp
# tests/test-grammar-integration.cpp
# tests/test-tokenizer-0-bpe.py
# tests/test-tokenizer-0-spm.py
# tests/test-tokenizer-1-spm.cpp
|
2024-04-30 21:04:17 +08:00 |
|
Concedo
|
4025efb497
|
anti wi keys
|
2024-04-30 00:25:37 +08:00 |
|
Concedo
|
46c9785bb5
|
default max ctx to true max ctx if found
|
2024-04-28 15:49:45 +08:00 |
|
Concedo
|
8faa6ed84d
|
clarify some descriptions
|
2024-04-28 15:16:02 +08:00 |
|
Concedo
|
6c3fd5b685
|
updated lite (+2 squashed commit)
Squashed commit:
[d10a731e] update lite
[2554b8e6] update docs
|
2024-04-28 10:40:57 +08:00 |
|
Concedo
|
c230b78906
|
refactored a lot of code, remove bantokens, move it to api
|
2024-04-27 17:57:13 +08:00 |
|
Concedo
|
4ec8a9c57b
|
expose stop reason in generation
|
2024-04-27 01:12:12 +08:00 |
|
Concedo
|
a681cdd9ef
|
Merge branch 'upstream' into concedo_experimental
# Conflicts:
# common/sampling.h
# llama.h
# tests/test-chat-template.cpp
|
2024-04-24 21:29:07 +08:00 |
|
Concedo
|
69dcffa4ec
|
updated lite and colab
|
2024-04-21 16:48:48 +08:00 |
|
Concedo
|
b01820dec7
|
auto rope scaling changes
|
2024-04-19 23:08:55 +08:00 |
|
Concedo
|
5b6ac9cc6e
|
add llama3 template
|
2024-04-19 22:15:35 +08:00 |
|
Concedo
|
bb7eb36134
|
test copying from install
|
2024-04-16 16:49:38 +08:00 |
|
Concedo
|
010b2a5a71
|
fixed bad handling for malformed requests
|
2024-04-14 20:37:08 +08:00 |
|
Concedo
|
cfbe8cffbd
|
updated lite, added extra check in pyinstaller cleanup
|
2024-04-14 00:22:21 +08:00 |
|
Concedo
|
bf320dca90
|
update lite
|
2024-04-10 21:51:46 +08:00 |
|
Concedo
|
d54af7fa31
|
updated swagger json link fix
|
2024-04-09 14:55:27 +08:00 |
|
Concedo
|
81ac0e5656
|
Merge branch 'upstream' into concedo_experimental
# Conflicts:
# .devops/full-cuda.Dockerfile
# .devops/full-rocm.Dockerfile
# .devops/full.Dockerfile
# .devops/llama-cpp-clblast.srpm.spec
# .devops/llama-cpp-cuda.srpm.spec
# .devops/llama-cpp.srpm.spec
# .devops/nix/package.nix
# .devops/server-cuda.Dockerfile
# .devops/server-intel.Dockerfile
# .devops/server-rocm.Dockerfile
# .devops/server-vulkan.Dockerfile
# .devops/server.Dockerfile
# .github/workflows/build.yml
# .github/workflows/code-coverage.yml
# .github/workflows/docker.yml
# .github/workflows/editorconfig.yml
# .github/workflows/gguf-publish.yml
# .github/workflows/nix-ci-aarch64.yml
# .github/workflows/nix-ci.yml
# .github/workflows/python-check-requirements.yml
# .github/workflows/python-lint.yml
# .github/workflows/server.yml
# .github/workflows/zig-build.yml
# CMakeLists.txt
# Makefile
# README-sycl.md
# README.md
# ci/run.sh
# examples/gguf-split/gguf-split.cpp
# flake.lock
# flake.nix
# llama.cpp
# scripts/compare-llama-bench.py
# scripts/sync-ggml-am.sh
# scripts/sync-ggml.last
# scripts/sync-ggml.sh
# tests/CMakeLists.txt
# tests/test-backend-ops.cpp
# tests/test-chat-template.cpp
|
2024-04-07 22:07:27 +08:00 |
|
Concedo
|
d8b808454d
|
updated lite
|
2024-04-07 19:45:04 +08:00 |
|
Concedo
|
bec16d182b
|
Merge commit '2f34b865b6' into concedo_experimental
# Conflicts:
# .clang-tidy
# CMakeLists.txt
# Makefile
# ggml-cuda.cu
|
2024-04-07 18:30:35 +08:00 |
|
Concedo
|
743687020d
|
fixed img2img
|
2024-04-06 17:29:44 +08:00 |
|
Concedo
|
8131616454
|
updated lite
|
2024-03-20 00:13:44 +08:00 |
|
Concedo
|
8b360b661c
|
Merge branch 'upstream' into concedo_experimental
# Conflicts:
# Makefile
# README.md
# common/common.h
|
2024-03-17 23:03:12 +08:00 |
|
Concedo
|
7a2de82c96
|
updated lite
|
2024-03-13 18:27:19 +08:00 |
|
Concedo
|
a9435163ab
|
fixed uploading non square images
|
2024-03-13 14:19:51 +08:00 |
|
Concedo
|
85287c7701
|
handle uploading non square images
|
2024-03-13 13:57:14 +08:00 |
|
Concedo
|
6c6ad93f01
|
added basic support for password protection (+2 squashed commit)
Squashed commit:
[ff91ca72] added basic support for password protection
[91b0b208] updated docs
|
2024-03-12 19:47:12 +08:00 |
|
Concedo
|
9229ea664e
|
if no existing filepath, do not use cwd, use last path instead
|
2024-03-11 22:19:38 +08:00 |
|
Concedo
|
95c8090967
|
updated lite
|
2024-03-11 21:59:18 +08:00 |
|
Concedo
|
2dc647f892
|
updated lite (+1 squashed commits)
Squashed commits:
[f33ea44a] updated lite
|
2024-03-11 20:10:34 +08:00 |
|
Concedo
|
d59ec68753
|
added interrogate endpoint (+1 squashed commits)
Squashed commits:
[7bf96261] added interrogate endpoint
|
2024-03-11 18:50:18 +08:00 |
|
Concedo
|
484d90c330
|
llava support is now fully functioning
|
2024-03-11 15:55:32 +08:00 |
|
Concedo
|
4682918965
|
better resolution scaling, support non-square images (+1 squashed commits)
Squashed commits:
[6efee4f2] better resolution scaling
|
2024-03-09 12:38:24 +08:00 |
|
Concedo
|
b4ca54401d
|
Merge branch 'master' into concedo_experimental
# Conflicts:
# Makefile
# README.md
# ci/run.sh
# ggml-quants.c
# ggml.c
# grammars/json.gbnf
# grammars/json_arr.gbnf
# llama.cpp
# scripts/compare-llama-bench.py
|
2024-03-09 10:58:49 +08:00 |
|
Concedo
|
e64210ef16
|
updated lite
|
2024-03-05 21:32:32 +08:00 |
|
Concedo
|
e0d6fbb1ec
|
updated lite
|
2024-03-04 18:01:52 +08:00 |
|
Concedo
|
b8bc500a4b
|
updated readme and lite
|
2024-03-03 23:31:52 +08:00 |
|
Concedo
|
0c59c1ed90
|
allow specifying width and height
|
2024-03-03 15:44:15 +08:00 |
|
Concedo
|
fa1d8b8d95
|
updated lite
|
2024-03-02 22:33:02 +08:00 |
|
Concedo
|
4c0beef598
|
updated lite, added personal notes
|
2024-03-02 18:53:15 +08:00 |
|
Concedo
|
e53d21d748
|
sanitize SD prompt to avoid segfault
|
2024-03-02 12:05:59 +08:00 |
|
Concedo
|
59c5448ac8
|
fixed colab (+1 squashed commits)
Squashed commits:
[1d1c686f] updated colab and docs
|
2024-03-02 10:09:07 +08:00 |
|
Concedo
|
2d9a90b652
|
try to fix ci compile errors (+1 squashed commits)
Squashed commits:
[d0d49663] fixed log multiline (+1 squashed commits)
Squashed commits:
[81a8befe] try to fix linux build error (+1 squashed commits)
Squashed commits:
[22850dda] try to fix build (+1 squashed commits)
Squashed commits:
[b8294611] missing type
|
2024-03-01 23:38:15 +08:00 |
|
Concedo
|
55af5446ad
|
Merge branch 'master' into concedo_experimental
# Conflicts:
# README.md
# ci/run.sh
# llama.cpp
# scripts/sync-ggml.last
|
2024-03-01 17:41:37 +08:00 |
|
Concedo
|
80011ed8aa
|
KCPP SD: add warn and step restriction., updated lite, handle quant mode
|
2024-03-01 16:41:19 +08:00 |
|