From f289fb494a22e55c02ebee27891aafc9be7ea862 Mon Sep 17 00:00:00 2001 From: Concedo <39025047+LostRuins@users.noreply.github.com> Date: Sun, 28 Jul 2024 20:29:39 +0800 Subject: [PATCH] bump size of some payload arr sequences from 16 to 24 --- expose.h | 6 +++--- koboldcpp.py | 6 +++--- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/expose.h b/expose.h index b620c1b2e..42d0ff31e 100644 --- a/expose.h +++ b/expose.h @@ -1,11 +1,11 @@ #pragma once #include -const int stop_token_max = 16; +const int stop_token_max = 24; const int ban_token_max = 16; const int tensor_split_max = 16; -const int logit_bias_max = 16; -const int dry_seq_break_max = 16; +const int logit_bias_max = 24; +const int dry_seq_break_max = 24; const int images_max = 4; // match kobold's sampler list and order diff --git a/koboldcpp.py b/koboldcpp.py index b3f01b13e..42c17b347 100644 --- a/koboldcpp.py +++ b/koboldcpp.py @@ -18,11 +18,11 @@ from concurrent.futures import ThreadPoolExecutor # constants sampler_order_max = 7 -stop_token_max = 16 +stop_token_max = 24 ban_token_max = 16 tensor_split_max = 16 -logit_bias_max = 16 -dry_seq_break_max = 16 +logit_bias_max = 24 +dry_seq_break_max = 24 images_max = 4 bias_min_value = -100.0 bias_max_value = 100.0