diff --git a/gpttype_adapter.cpp b/gpttype_adapter.cpp index ec769aa28..8da5600de 100644 --- a/gpttype_adapter.cpp +++ b/gpttype_adapter.cpp @@ -3263,28 +3263,28 @@ generation_outputs gpttype_generate(const generation_inputs inputs) } //need to add a cursed hack to get coherency for GLM4, by ensuring injection for both sop and gmask - if (file_format == FileFormat::GGUF_GENERIC && file_format_meta.model_architecture == GGUFArch::ARCH_GLM4) { - std::string temp = gpttype_get_chat_template(); - if (temp.find("[gMASK]") != std::string::npos) { - if (addedmemory == "") { - if (kcpp_data->prompt.rfind("[gMASK]", 0) == 0) { //check startswith - kcpp_data->prompt.erase(0, 7); - } - if (kcpp_data->prompt.rfind("", 0) == 0) { //check startswith - kcpp_data->prompt.erase(0, 5); - } - addedmemory = "[gMASK]"; - } else { - if (addedmemory.rfind("[gMASK]", 0) == 0) { //check startswith - addedmemory.erase(0, 7); - } - if (addedmemory.rfind("", 0) == 0) { //check startswith - addedmemory.erase(0, 5); - } - addedmemory = "[gMASK]" + addedmemory; - } - } - } + // if (file_format == FileFormat::GGUF_GENERIC && file_format_meta.model_architecture == GGUFArch::ARCH_GLM4) { + // std::string temp = gpttype_get_chat_template(); + // if (temp.find("[gMASK]") != std::string::npos) { + // if (addedmemory == "") { + // if (kcpp_data->prompt.rfind("[gMASK]", 0) == 0) { //check startswith + // kcpp_data->prompt.erase(0, 7); + // } + // if (kcpp_data->prompt.rfind("", 0) == 0) { //check startswith + // kcpp_data->prompt.erase(0, 5); + // } + // addedmemory = "[gMASK]"; + // } else { + // if (addedmemory.rfind("[gMASK]", 0) == 0) { //check startswith + // addedmemory.erase(0, 7); + // } + // if (addedmemory.rfind("", 0) == 0) { //check startswith + // addedmemory.erase(0, 5); + // } + // addedmemory = "[gMASK]" + addedmemory; + // } + // } + // } bool stream_sse = inputs.stream_sse; bool allow_regular_prints = (!is_quiet && debugmode!=-1); diff --git a/koboldcpp.py b/koboldcpp.py index fa914e732..b35065bfc 100644 --- a/koboldcpp.py +++ b/koboldcpp.py @@ -4433,8 +4433,8 @@ def show_gui(): except Exception: pass else: - quick_gpuname_label.configure(text="") - gpuname_label.configure(text="") + quick_gpuname_label.configure(text="(dGPUs only, tensor split sets ratio)") + gpuname_label.configure(text="(dGPUs only, tensor split sets ratio)") gpu_choice_var.trace("w", changed_gpu_choice_var) gpulayers_var.trace("w", changed_gpulayers_estimate)