fixes for vulkan multigpu

This commit is contained in:
Concedo 2024-02-09 14:42:27 +08:00
parent fe424a5466
commit 992eea71d7
3 changed files with 34 additions and 18 deletions

View file

@ -961,7 +961,12 @@ ModelLoadResult gpttype_load_model(const load_model_inputs inputs, FileFormat in
} }
#endif #endif
model_params.main_gpu = cu_parseinfo_maindevice; model_params.main_gpu = cu_parseinfo_maindevice;
#if defined(GGML_USE_CUBLAS)
model_params.split_mode = llama_split_mode::LLAMA_SPLIT_ROW; model_params.split_mode = llama_split_mode::LLAMA_SPLIT_ROW;
#else
model_params.split_mode = llama_split_mode::LLAMA_SPLIT_LAYER;
#endif
llama_ctx_params.n_batch = kcpp_params->n_batch; llama_ctx_params.n_batch = kcpp_params->n_batch;
llama_ctx_params.n_threads = kcpp_params->n_threads; llama_ctx_params.n_threads = kcpp_params->n_threads;

View file

@ -6,7 +6,7 @@ It requires no dependencies, installation or setup.
Just copy this single static HTML file anywhere and open it in a browser, or from a webserver. Just copy this single static HTML file anywhere and open it in a browser, or from a webserver.
Please go to https://github.com/LostRuins/lite.koboldai.net for updates on Kobold Lite. Please go to https://github.com/LostRuins/lite.koboldai.net for updates on Kobold Lite.
Kobold Lite is under the AGPL v3.0 License unless otherwise exempted. Please do not remove this line. Kobold Lite is under the AGPL v3.0 License unless otherwise exempted. Please do not remove this line.
Current version: 109 Current version: 110
-Concedo -Concedo
--> -->
@ -3036,12 +3036,14 @@ Current version: 109
}) })
.catch((error) => { .catch((error) => {
console.error('Error:', error); console.error('Error:', error);
clear_poll_flags();
render_gametext();
if(error.name!="AbortError") //aborts are silent if(error.name!="AbortError") //aborts are silent
{ {
flush_streaming_text();
msgbox("Error while submitting prompt: " + error); msgbox("Error while submitting prompt: " + error);
} }
clear_poll_flags();
render_gametext();
}); });
} }
} }
@ -3110,28 +3112,26 @@ Current version: 109
} }
}, },
abort(error) { abort(error) {
console.error('Error:', error); console.error('Error:', error);
clear_poll_flags();
render_gametext();
if(error.name!="AbortError") //aborts are silent. slightly diff logic if(error.name!="AbortError") //aborts are silent. slightly diff logic
{ {
flush_streaming_text();
msgbox("Error while submitting prompt: " + error); msgbox("Error while submitting prompt: " + error);
} }
clear_poll_flags();
render_gametext();
}, },
})); }));
}) })
.catch((error) => { .catch((error) => {
console.error('Error:', error); console.error('Error:', error);
clear_poll_flags();
render_gametext();
if(error.name!="AbortError") //aborts are silent. slightly diff logic if(error.name!="AbortError") //aborts are silent. slightly diff logic
{ {
flush_streaming_text();
msgbox("Error while submitting prompt: " + error); msgbox("Error while submitting prompt: " + error);
} }
clear_poll_flags();
render_gametext();
}); });
} }
@ -8371,10 +8371,8 @@ Current version: 109
} }
} }
function abort_generation() { function flush_streaming_text()
let id_to_cancel = pending_response_id; {
//flush any streaming text first
if(is_using_custom_ep() && pending_response_id != "" && (synchro_pending_stream != "" || synchro_polled_response != "")) if(is_using_custom_ep() && pending_response_id != "" && (synchro_pending_stream != "" || synchro_polled_response != ""))
{ {
//apply a short delay of 1s before button reenables //apply a short delay of 1s before button reenables
@ -8391,6 +8389,13 @@ Current version: 109
horde_poll_nearly_completed = false; horde_poll_nearly_completed = false;
poll_pending_response(); poll_pending_response();
} }
}
function abort_generation() {
let id_to_cancel = pending_response_id;
//flush any streaming text first
flush_streaming_text();
console.log("Generation " + pending_response_id + " aborted"); console.log("Generation " + pending_response_id + " aborted");
clear_poll_flags(); clear_poll_flags();
@ -8908,10 +8913,16 @@ Current version: 109
} }
if (localsettings.opmode == 4) { if (localsettings.opmode == 4) {
if(localsettings.inject_timestamps_instruct)
if(localsettings.inject_timestamps_instruct && pending_context_preinjection=="" && truncated_context!="")
{
let endmatcher = (localsettings.placeholder_tags?instructendplaceholder:get_instruct_endtag(false));
if(truncated_context.toLowerCase().trim().endsWith(endmatcher.toLowerCase().trim()))
{ {
pending_context_preinjection += "["+(new Date().toLocaleTimeString([], {year: 'numeric', month: 'numeric', day: 'numeric', hour: '2-digit', minute: '2-digit'}))+"] "; pending_context_preinjection += "["+(new Date().toLocaleTimeString([], {year: 'numeric', month: 'numeric', day: 'numeric', hour: '2-digit', minute: '2-digit'}))+"] ";
} }
}
truncated_context += pending_context_preinjection; truncated_context += pending_context_preinjection;
} }

View file

@ -464,7 +464,7 @@ maxhordelen = 256
modelbusy = threading.Lock() modelbusy = threading.Lock()
requestsinqueue = 0 requestsinqueue = 0
defaultport = 5001 defaultport = 5001
KcppVersion = "1.57" KcppVersion = "1.57.1"
showdebug = True showdebug = True
showsamplerwarning = True showsamplerwarning = True
showmaxctxwarning = True showmaxctxwarning = True