\r\n",
"LoadImgModel = False #@param {type:\"boolean\"}\r\n",
- "ImgModel = \"https://huggingface.co/koboldcpp/imgmodel/resolve/main/imgmodel_older_q4_0.gguf\" #@param [\"https://huggingface.co/koboldcpp/imgmodel/resolve/main/imgmodel_older_q4_0.gguf\"]{allow-input: true}\r\n",
+ "ImgModel = \"https://huggingface.co/koboldcpp/imgmodel/resolve/main/imgmodel_ftuned_q4_0.gguf\" #@param [\"https://huggingface.co/koboldcpp/imgmodel/resolve/main/imgmodel_ftuned_q4_0.gguf\"]{allow-input: true}\r\n",
"SCommand = \"\"\r\n",
"\r\n",
"import os\r\n",
diff --git a/klite.embd b/klite.embd
index c33ec11c6..352a71625 100644
--- a/klite.embd
+++ b/klite.embd
@@ -7,7 +7,7 @@ Just copy this single static HTML file anywhere and open it in a browser, or fro
Please go to https://github.com/LostRuins/lite.koboldai.net for updates on Kobold Lite.
If you are submitting a pull request for Lite, PLEASE use the above repo, not the KoboldCpp one.
Kobold Lite is under the AGPL v3.0 License unless otherwise exempted. Please do not remove this line.
-Current version: 128
+Current version: 129
-Concedo
-->
@@ -10343,7 +10343,7 @@ Current version: 128
}
];
}
- else if(mdlname=="gemini-pro" || mdlname=="gemini-1.5-pro-latest")
+ else if(mdlname=="gemini-pro" || mdlname=="gemini-1.5-pro-latest" || mdlname=="gemini-ultra")
{
if(localsettings.opmode==1)
{
@@ -14036,6 +14036,7 @@ Current version: 128
+
Please input Gemini or PaLM API Key.
diff --git a/koboldcpp.py b/koboldcpp.py
index 4108c3a5f..5754ef662 100644
--- a/koboldcpp.py
+++ b/koboldcpp.py
@@ -1115,7 +1115,7 @@ Enter Prompt:
response_body = (json.dumps([]).encode())
- elif self.path=="/api" or self.path=="/docs":
+ elif self.path=="/api" or self.path=="/docs" or self.path.startswith(('/api/?json=','/api?json=','/docs/?json=','/docs?json=')):
content_type = 'text/html'
if self.embedded_kcpp_docs is None:
response_body = (f"KoboldCpp API is running!\n\nAPI usage reference can be found at the wiki: https://github.com/LostRuins/koboldcpp/wiki").encode()