mirror of
https://github.com/LostRuins/koboldcpp.git
synced 2025-09-10 17:14:36 +00:00
updated lite
This commit is contained in:
parent
5ab2fd6118
commit
e958b2f78b
2 changed files with 67 additions and 32 deletions
93
klite.embd
93
klite.embd
|
@ -12,7 +12,7 @@ Current version indicated by LITEVER below.
|
|||
-->
|
||||
|
||||
<script>
|
||||
const LITEVER = 176;
|
||||
const LITEVER = 177;
|
||||
const urlParams = new URLSearchParams(window.location.search);
|
||||
var localflag = true;
|
||||
const STORAGE_PREFIX = (localflag?"e_":"")+"kaihordewebui_";
|
||||
|
@ -8230,6 +8230,8 @@ Current version indicated by LITEVER below.
|
|||
if(ep_should_always_use_chat_completions() || dropdown.selectedIndex==dropdown.options.length-1)
|
||||
{
|
||||
document.getElementById("useoaichatcompl").checked = true;
|
||||
} else if (document.getElementById("custom_oai_endpoint").value.toLowerCase().includes("featherless.ai")) {
|
||||
document.getElementById("useoaichatcompl").checked = false; //use completions for a better experience
|
||||
} else {
|
||||
document.getElementById("useoaichatcompl").checked = !non_completions;
|
||||
}
|
||||
|
@ -8240,21 +8242,7 @@ Current version indicated by LITEVER below.
|
|||
{
|
||||
let desired_oai_key = document.getElementById("custom_oai_key").value.trim();
|
||||
let desired_oai_ep = document.getElementById("custom_oai_endpoint").value.trim();
|
||||
if(desired_oai_ep!="" && desired_oai_ep.slice(-1)=="/")
|
||||
{
|
||||
desired_oai_ep = desired_oai_ep.slice(0, -1);
|
||||
}
|
||||
if(!desired_oai_ep.includes("://")) //user did not add http/https
|
||||
{
|
||||
let is_local = is_local_url(desired_oai_ep);
|
||||
desired_oai_ep = (is_local?"http://":"https://") + desired_oai_ep;
|
||||
}
|
||||
if (document.getElementById("oaiaddversion").checked)
|
||||
{
|
||||
if(desired_oai_ep!="" && desired_oai_ep.length > 4 && !desired_oai_ep.slice(-4).toLowerCase().includes("/v") && !desired_oai_ep.toLowerCase().includes("/v1/")) {
|
||||
desired_oai_ep = desired_oai_ep + "/v1";
|
||||
}
|
||||
}
|
||||
desired_oai_ep = transform_oai_ep(desired_oai_ep);
|
||||
|
||||
let oaiheaders = {
|
||||
'Authorization': 'Bearer '+desired_oai_key,
|
||||
|
@ -8340,6 +8328,24 @@ Current version indicated by LITEVER below.
|
|||
}
|
||||
|
||||
let openrouter_fetch_attempted = false;
|
||||
let oai_custom_fetch_attempted = false;
|
||||
function try_fetch_oai_models_auto()
|
||||
{
|
||||
//only for apis that don't gate the model list
|
||||
if (document.getElementById("custom_oai_endpoint").value!="" &&
|
||||
document.getElementById("custom_oai_endpoint").value.toLowerCase().includes("featherless.ai"))
|
||||
{
|
||||
if(!oai_custom_fetch_attempted)
|
||||
{
|
||||
oai_custom_fetch_attempted = true;
|
||||
let dropdown = document.getElementById("custom_oai_model");
|
||||
if(dropdown.options.length < 40)
|
||||
{
|
||||
oai_fetch_models(); //autofetch models
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
function customapi_dropdown(force_autotoggle_chatcompl = false)
|
||||
{
|
||||
let epchoice = document.getElementById("customapidropdown").value;
|
||||
|
@ -8384,6 +8390,7 @@ Current version indicated by LITEVER below.
|
|||
} else {
|
||||
document.getElementById("custom_oai_endpoint").value = (localsettings.saved_oai_addr ? localsettings.saved_oai_addr : default_oai_base);
|
||||
}
|
||||
try_fetch_oai_models_auto();
|
||||
}
|
||||
else if(epchoice==7)
|
||||
{
|
||||
|
@ -8473,6 +8480,31 @@ Current version indicated by LITEVER below.
|
|||
}
|
||||
}
|
||||
|
||||
function transform_oai_ep(desired_oai_ep)
|
||||
{
|
||||
if(desired_oai_ep!="" && desired_oai_ep.slice(-1)=="/")
|
||||
{
|
||||
desired_oai_ep = desired_oai_ep.slice(0, -1);
|
||||
}
|
||||
if(!desired_oai_ep.includes("://")) //user did not add http/https
|
||||
{
|
||||
let is_local = is_local_url(desired_oai_ep);
|
||||
desired_oai_ep = (is_local?"http://":"https://") + desired_oai_ep;
|
||||
}
|
||||
if (document.getElementById("oaiaddversion").checked)
|
||||
{
|
||||
if(desired_oai_ep!="" && desired_oai_ep.length > 4 && !desired_oai_ep.slice(-4).toLowerCase().includes("/v") && !desired_oai_ep.toLowerCase().includes("/v1/")) {
|
||||
desired_oai_ep = desired_oai_ep + "/v1";
|
||||
}
|
||||
}
|
||||
//fix common url typos if detected
|
||||
if(desired_oai_ep!="" && desired_oai_ep.toLowerCase().startsWith("https://featherless.ai"))
|
||||
{
|
||||
desired_oai_ep = desired_oai_ep.replace("https://featherless.ai","https://api.featherless.ai")
|
||||
}
|
||||
return desired_oai_ep;
|
||||
}
|
||||
|
||||
function connect_custom_endpoint()
|
||||
{
|
||||
mainmenu_untab(false);
|
||||
|
@ -8788,21 +8820,8 @@ Current version indicated by LITEVER below.
|
|||
desired_oai_ep = document.getElementById("custom_oai_endpoint").value = default_oai_base;
|
||||
}
|
||||
|
||||
if(desired_oai_ep!="" && desired_oai_ep.slice(-1)=="/")
|
||||
{
|
||||
desired_oai_ep = desired_oai_ep.slice(0, -1);
|
||||
}
|
||||
if(!desired_oai_ep.includes("://")) //user did not add http/https
|
||||
{
|
||||
let is_local = is_local_url(desired_oai_ep);
|
||||
desired_oai_ep = (is_local?"http://":"https://") + desired_oai_ep;
|
||||
}
|
||||
if (document.getElementById("oaiaddversion").checked)
|
||||
{
|
||||
if(desired_oai_ep!="" && desired_oai_ep.length > 4 && !desired_oai_ep.slice(-4).toLowerCase().includes("/v") && !desired_oai_ep.toLowerCase().includes("/v1/")) {
|
||||
desired_oai_ep = desired_oai_ep + "/v1";
|
||||
}
|
||||
}
|
||||
desired_oai_ep = transform_oai_ep(desired_oai_ep);
|
||||
|
||||
if(desired_oai_key!="" && desired_oai_ep!="")
|
||||
{
|
||||
dismiss_endpoint_container();
|
||||
|
@ -12421,6 +12440,16 @@ Current version indicated by LITEVER below.
|
|||
//mistral api does not support presence pen
|
||||
oai_payload.presence_penalty = scaled_rep_pen;
|
||||
}
|
||||
if(targetep.toLowerCase().includes("featherless.ai"))
|
||||
{
|
||||
//featherless api supports additional fields, include them
|
||||
oai_payload.top_k = (submit_payload.params.top_k<1?300:submit_payload.params.top_k);
|
||||
oai_payload.min_p = localsettings.min_p;
|
||||
if(submit_payload.params.sampler_seed>=1)
|
||||
{
|
||||
oai_payload.seed = submit_payload.params.sampler_seed;
|
||||
}
|
||||
}
|
||||
if(submit_payload.params.logit_bias && JSON.stringify(submit_payload.params.logit_bias) != '{}')
|
||||
{
|
||||
oai_payload.logit_bias = submit_payload.params.logit_bias;
|
||||
|
@ -17709,7 +17738,7 @@ Current version indicated by LITEVER below.
|
|||
<span class="color_green" style="font-weight: bold;">Please input MistralAI Key.</span><br><br>
|
||||
</span>
|
||||
|
||||
<input class="form-control" type="text" id="custom_oai_endpoint" placeholder="OpenAI API URL" value="">
|
||||
<input class="form-control" type="text" id="custom_oai_endpoint" placeholder="OpenAI API URL" value="" onblur="try_fetch_oai_models_auto()">
|
||||
<input class="form-control" type="password" id="custom_oai_key" placeholder="OpenAI API Key (Required)" value="" onfocus="focus_api_keys()" onblur="blur_api_keys()"><br>
|
||||
Model Choice:<br>
|
||||
<select style="padding:4px;display:inline;width:calc(100% - 220px)" class="form-control" id="custom_oai_model" onchange="oai_model_change(true)">
|
||||
|
|
|
@ -2492,6 +2492,7 @@ def show_gui():
|
|||
fetch_gpu_properties(False,True,True)
|
||||
else:
|
||||
fetch_gpu_properties(True,True,True)
|
||||
found_new_backend = False
|
||||
#autopick cublas if suitable, requires at least 3.5GB VRAM to auto pick
|
||||
#we do not want to autoselect hip/cublas if the user has already changed their desired backend!
|
||||
if exitcounter < 100 and MaxMemory[0]>3500000000 and (("Use CuBLAS" in runopts and CUDevicesNames[0]!="") or "Use hipBLAS (ROCm)" in runopts) and (any(CUDevicesNames) or any(CLDevicesNames)) and runmode_untouched:
|
||||
|
@ -2499,17 +2500,22 @@ def show_gui():
|
|||
runopts_var.set("Use CuBLAS")
|
||||
gpu_choice_var.set("1")
|
||||
print("Auto Selected CUDA Backend...\n")
|
||||
found_new_backend = True
|
||||
elif "Use hipBLAS (ROCm)" in runopts:
|
||||
runopts_var.set("Use hipBLAS (ROCm)")
|
||||
gpu_choice_var.set("1")
|
||||
print("Auto Selected HIP Backend...\n")
|
||||
found_new_backend = True
|
||||
elif exitcounter < 100 and (1 in VKIsDGPU) and runmode_untouched and "Use Vulkan" in runopts:
|
||||
for i in range(0,len(VKIsDGPU)):
|
||||
if VKIsDGPU[i]==1:
|
||||
runopts_var.set("Use Vulkan")
|
||||
gpu_choice_var.set(str(i+1))
|
||||
print("Auto Selected Vulkan Backend...\n")
|
||||
found_new_backend = True
|
||||
break
|
||||
if not found_new_backend:
|
||||
print("Auto Selected Default Backend...\n")
|
||||
changed_gpu_choice_var()
|
||||
|
||||
def on_picked_model_file(filepath):
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue