Commit graph

186 commits

Author SHA1 Message Date
Concedo
8823e8b06d added presence penalty into lite ui 2023-12-23 10:39:40 +08:00
Concedo
77463e0e9c batch size improvements 2023-12-22 15:27:40 +08:00
Concedo
3f863eed72 add presence penalty 2023-12-19 23:18:56 +08:00
Concedo
7469f202ea use lowvram flag for offload qkv 2023-12-08 18:16:14 +08:00
Concedo
ec21fa7712 Merge branch 'master' into concedo_experimental
# Conflicts:
#	.github/workflows/build.yml
#	.gitignore
#	CMakeLists.txt
#	Makefile
#	Package.swift
#	README.md
#	ggml-cuda.cu
#	llama.cpp
#	llama.h
#	scripts/sync-ggml.sh
#	tests/CMakeLists.txt
2023-12-08 17:42:26 +08:00
Concedo
c7511526a2 noscript mode is done 2023-12-07 00:52:25 +08:00
Concedo
6570a2005b token count includes ids 2023-12-03 15:44:53 +08:00
Concedo
c142c5634a fixed segfault with clblast by reversing commit in issue https://github.com/ggerganov/llama.cpp/issues/4296 2023-12-03 00:56:00 +08:00
Concedo
12f66eaa1d adjust fragmentation fix 2023-12-02 15:59:08 +08:00
Concedo
a012342a77 updated docs, shifted kv extra space to be subtracted from user's ctx value instead of added on load. 2023-11-30 14:19:40 +08:00
Concedo
ba5c33319b Allocate a small amount of extra context for GGUF to deal with KV fragmentation causing issues in some scenarios. 2023-11-28 20:55:14 +08:00
Concedo
bffa78116d explore quiet mode 2023-11-26 23:57:27 +08:00
Concedo
a6eb9b8010 Fix GPT2 not loading due to graph too small 2023-11-26 23:06:42 +08:00
Concedo
eb42c73953 revert auto rope scaling for already-ropetuned models - just use their values 2023-11-24 14:20:36 +08:00
Concedo
4d7c14be73 fix stop seq escaping newline 2023-11-20 22:35:45 +08:00
Concedo
cf646fa809 try to scale custom roped models 2023-11-19 16:24:13 +08:00
Concedo
8b919b5b57 allow customized rope to use model set values 2023-11-15 16:21:52 +08:00
Concedo
be92cfa125 added preloadstory 2023-11-10 13:05:22 +08:00
Concedo
fb3bcac368 handle memory separately for kcpp 2023-11-07 17:15:14 +08:00
Concedo
1e7088a80b autopick cublas in gui if possible, better layer picking logic 2023-11-05 01:35:27 +08:00
Concedo
ae2cd56de8 kobold integration of min_p sampler (+1 squashed commits)
Squashed commits:

[8ad2e349] kobold integration for min_p sampler
2023-11-01 19:08:45 +08:00
Concedo
cc5b282350 Merge branch 'master' into concedo_experimental
# Conflicts:
#	CMakeLists.txt
#	Makefile
#	build.zig
#	flake.lock
#	flake.nix
#	ggml.c
2023-10-31 20:44:04 +08:00
Concedo
9eba77c6a0 finally got something workable 2023-10-30 23:30:21 +08:00
Concedo
7f050b5d16 tweak numbers 2023-10-29 22:46:19 +08:00
Concedo
7924592a83 context shift feature done 2023-10-29 18:21:39 +08:00
Concedo
338d6c265d fixes to smartcontextpro 2023-10-29 10:42:37 +08:00
Concedo
20ef442c2a fixed for smartcontext 2023-10-28 19:09:22 +08:00
Concedo
15f525c580 revamped smart context for llama models 2023-10-28 12:59:08 +08:00
Concedo
0f46534866 wip 2023-10-26 21:58:51 +08:00
Concedo
5db89b90b7 Merge branch 'master' into concedo_experimental
# Conflicts:
#	.gitignore
#	CMakeLists.txt
#	Makefile
#	README.md
#	build.zig
#	ggml-opencl.cpp
#	tests/CMakeLists.txt
#	tests/test-double-float.cpp
#	tests/test-sampling.cpp
2023-10-25 23:58:15 +08:00
Concedo
839fc6dac8 handle freq_base_train 2023-10-24 23:44:22 +08:00
Concedo
cff75061fe fixed some old models failing due to tokenizer changes, update lite (+1 squashed commits)
Squashed commits:

[9dee81ec] fixed some old models failing due to tokenizer changes, update lite tooltip (+3 squashed commit)

Squashed commit:

[5ab95a79] fixes

[a561d5e2] fixed some old models failing due to tokenizer changes

[95e65daf] lite updates
2023-10-22 11:04:59 +08:00
kalomaze
ddce116ec9
Fix for Top K disabling (#480)
* Update gpttype_adapter.cpp

* use n_vocab instead of 32000 for when top k is off
2023-10-19 23:20:44 +08:00
Concedo
c1ca1de2ac fixed support for old falcon models 2023-10-18 17:20:44 +08:00
Concedo
700951dbd4 Merge branch 'master' into concedo_experimental
# Conflicts:
#	README.md
2023-10-18 16:33:09 +08:00
Concedo
d74eab0e63 actually for this round, do not include deprecated params. i dont want to have to deal with them (+2 squashed commit)
Squashed commit:

[df2691c2] show context limit

[7c74f52a] prevent old scripts from crashing
2023-10-10 19:20:33 +08:00
Concedo
0cc740115d updated lite, improve horde worker (+1 squashed commits)
Squashed commits:

[a7c25999] improve horde worker
2023-10-03 23:44:27 +08:00
Concedo
d10470a1e3 Breaking Change: Remove deprecated commands 2023-10-03 17:16:09 +08:00
Concedo
0c47e79537 updated the API routing path and fixed a bug with threads 2023-10-02 11:05:19 +08:00
Concedo
bc841ec302 flag to retain grammar, fix makefile (+2 squashed commit)
Squashed commit:

[d5cd3f28] flag to retain grammar, fix makefile

[b3352963] updated lite to v73
2023-10-01 14:39:56 +08:00
Concedo
5e6450161a functional merge 2023-09-30 12:31:57 +08:00
Concedo
b84e210f0d merge new rope param nonsense 2023-09-30 11:33:30 +08:00
Concedo
ca8b315202 increase context for gguf to 32k, horde worker stats, fixed glitch in horde launcher ui, oai freq penalty, updated lite 2023-09-28 23:50:08 +08:00
Concedo
de4e27904d clear reader copy on new gen 2023-09-23 00:13:19 +08:00
Concedo
c03409c1f6 grammar sampling added for lite 2023-09-19 00:13:30 +08:00
Concedo
8c453d1e4e added grammar sampling 2023-09-18 23:02:00 +08:00
Ycros
f6ba36dff6
Reduce warnings. (#439) 2023-09-16 18:52:09 +08:00
Concedo
6667fdcec8 add option for 4th gpu, also fixed missing case in auto rope scaling 2023-09-11 11:43:54 +08:00
Concedo
7fb809b94b fixed auto rope scaling (+1 squashed commits)
Squashed commits:

[b1767874] wip
2023-09-07 14:45:08 +08:00
Concedo
81abd3cb1f Merge remote-tracking branch 'elbios/concat_output_mutex' into concedo_experimental 2023-09-01 15:24:13 +08:00