Concedo
8823e8b06d
added presence penalty into lite ui
2023-12-23 10:39:40 +08:00
Concedo
77463e0e9c
batch size improvements
2023-12-22 15:27:40 +08:00
Concedo
3f863eed72
add presence penalty
2023-12-19 23:18:56 +08:00
Concedo
7469f202ea
use lowvram flag for offload qkv
2023-12-08 18:16:14 +08:00
Concedo
ec21fa7712
Merge branch 'master' into concedo_experimental
...
# Conflicts:
# .github/workflows/build.yml
# .gitignore
# CMakeLists.txt
# Makefile
# Package.swift
# README.md
# ggml-cuda.cu
# llama.cpp
# llama.h
# scripts/sync-ggml.sh
# tests/CMakeLists.txt
2023-12-08 17:42:26 +08:00
Concedo
c7511526a2
noscript mode is done
2023-12-07 00:52:25 +08:00
Concedo
6570a2005b
token count includes ids
2023-12-03 15:44:53 +08:00
Concedo
c142c5634a
fixed segfault with clblast by reversing commit in issue https://github.com/ggerganov/llama.cpp/issues/4296
2023-12-03 00:56:00 +08:00
Concedo
12f66eaa1d
adjust fragmentation fix
2023-12-02 15:59:08 +08:00
Concedo
a012342a77
updated docs, shifted kv extra space to be subtracted from user's ctx value instead of added on load.
2023-11-30 14:19:40 +08:00
Concedo
ba5c33319b
Allocate a small amount of extra context for GGUF to deal with KV fragmentation causing issues in some scenarios.
2023-11-28 20:55:14 +08:00
Concedo
bffa78116d
explore quiet mode
2023-11-26 23:57:27 +08:00
Concedo
a6eb9b8010
Fix GPT2 not loading due to graph too small
2023-11-26 23:06:42 +08:00
Concedo
eb42c73953
revert auto rope scaling for already-ropetuned models - just use their values
2023-11-24 14:20:36 +08:00
Concedo
4d7c14be73
fix stop seq escaping newline
2023-11-20 22:35:45 +08:00
Concedo
cf646fa809
try to scale custom roped models
2023-11-19 16:24:13 +08:00
Concedo
8b919b5b57
allow customized rope to use model set values
2023-11-15 16:21:52 +08:00
Concedo
be92cfa125
added preloadstory
2023-11-10 13:05:22 +08:00
Concedo
fb3bcac368
handle memory separately for kcpp
2023-11-07 17:15:14 +08:00
Concedo
1e7088a80b
autopick cublas in gui if possible, better layer picking logic
2023-11-05 01:35:27 +08:00
Concedo
ae2cd56de8
kobold integration of min_p sampler (+1 squashed commits)
...
Squashed commits:
[8ad2e349] kobold integration for min_p sampler
2023-11-01 19:08:45 +08:00
Concedo
cc5b282350
Merge branch 'master' into concedo_experimental
...
# Conflicts:
# CMakeLists.txt
# Makefile
# build.zig
# flake.lock
# flake.nix
# ggml.c
2023-10-31 20:44:04 +08:00
Concedo
9eba77c6a0
finally got something workable
2023-10-30 23:30:21 +08:00
Concedo
7f050b5d16
tweak numbers
2023-10-29 22:46:19 +08:00
Concedo
7924592a83
context shift feature done
2023-10-29 18:21:39 +08:00
Concedo
338d6c265d
fixes to smartcontextpro
2023-10-29 10:42:37 +08:00
Concedo
20ef442c2a
fixed for smartcontext
2023-10-28 19:09:22 +08:00
Concedo
15f525c580
revamped smart context for llama models
2023-10-28 12:59:08 +08:00
Concedo
0f46534866
wip
2023-10-26 21:58:51 +08:00
Concedo
5db89b90b7
Merge branch 'master' into concedo_experimental
...
# Conflicts:
# .gitignore
# CMakeLists.txt
# Makefile
# README.md
# build.zig
# ggml-opencl.cpp
# tests/CMakeLists.txt
# tests/test-double-float.cpp
# tests/test-sampling.cpp
2023-10-25 23:58:15 +08:00
Concedo
839fc6dac8
handle freq_base_train
2023-10-24 23:44:22 +08:00
Concedo
cff75061fe
fixed some old models failing due to tokenizer changes, update lite (+1 squashed commits)
...
Squashed commits:
[9dee81ec] fixed some old models failing due to tokenizer changes, update lite tooltip (+3 squashed commit)
Squashed commit:
[5ab95a79] fixes
[a561d5e2] fixed some old models failing due to tokenizer changes
[95e65daf] lite updates
2023-10-22 11:04:59 +08:00
kalomaze
ddce116ec9
Fix for Top K disabling ( #480 )
...
* Update gpttype_adapter.cpp
* use n_vocab instead of 32000 for when top k is off
2023-10-19 23:20:44 +08:00
Concedo
c1ca1de2ac
fixed support for old falcon models
2023-10-18 17:20:44 +08:00
Concedo
700951dbd4
Merge branch 'master' into concedo_experimental
...
# Conflicts:
# README.md
2023-10-18 16:33:09 +08:00
Concedo
d74eab0e63
actually for this round, do not include deprecated params. i dont want to have to deal with them (+2 squashed commit)
...
Squashed commit:
[df2691c2] show context limit
[7c74f52a] prevent old scripts from crashing
2023-10-10 19:20:33 +08:00
Concedo
0cc740115d
updated lite, improve horde worker (+1 squashed commits)
...
Squashed commits:
[a7c25999] improve horde worker
2023-10-03 23:44:27 +08:00
Concedo
d10470a1e3
Breaking Change: Remove deprecated commands
2023-10-03 17:16:09 +08:00
Concedo
0c47e79537
updated the API routing path and fixed a bug with threads
2023-10-02 11:05:19 +08:00
Concedo
bc841ec302
flag to retain grammar, fix makefile (+2 squashed commit)
...
Squashed commit:
[d5cd3f28] flag to retain grammar, fix makefile
[b3352963] updated lite to v73
2023-10-01 14:39:56 +08:00
Concedo
5e6450161a
functional merge
2023-09-30 12:31:57 +08:00
Concedo
b84e210f0d
merge new rope param nonsense
2023-09-30 11:33:30 +08:00
Concedo
ca8b315202
increase context for gguf to 32k, horde worker stats, fixed glitch in horde launcher ui, oai freq penalty, updated lite
2023-09-28 23:50:08 +08:00
Concedo
de4e27904d
clear reader copy on new gen
2023-09-23 00:13:19 +08:00
Concedo
c03409c1f6
grammar sampling added for lite
2023-09-19 00:13:30 +08:00
Concedo
8c453d1e4e
added grammar sampling
2023-09-18 23:02:00 +08:00
Ycros
f6ba36dff6
Reduce warnings. ( #439 )
2023-09-16 18:52:09 +08:00
Concedo
6667fdcec8
add option for 4th gpu, also fixed missing case in auto rope scaling
2023-09-11 11:43:54 +08:00
Concedo
7fb809b94b
fixed auto rope scaling (+1 squashed commits)
...
Squashed commits:
[b1767874] wip
2023-09-07 14:45:08 +08:00
Concedo
81abd3cb1f
Merge remote-tracking branch 'elbios/concat_output_mutex' into concedo_experimental
2023-09-01 15:24:13 +08:00