diff --git a/MIT_LICENSE_GGML_LLAMACPP_ONLY b/MIT_LICENSE_GGML_LLAMACPP_ONLY index 22fd48829..cf9797087 100644 --- a/MIT_LICENSE_GGML_LLAMACPP_ONLY +++ b/MIT_LICENSE_GGML_LLAMACPP_ONLY @@ -23,4 +23,4 @@ SOFTWARE. =================================== Note that the above license applies ONLY to the GGML library and llama.cpp by ggerganov which are licensed under the MIT License -Kobold Lite by Concedo and the provided python ctypes bindings in koboldcpp dlls are licensed under the AGPL v3.0 License \ No newline at end of file +KoboldAI Lite by Concedo and the provided python ctypes bindings in koboldcpp dlls are licensed under the AGPL v3.0 License \ No newline at end of file diff --git a/README.md b/README.md index fe082829a..849854a7b 100644 --- a/README.md +++ b/README.md @@ -1,6 +1,6 @@ # koboldcpp -KoboldCpp is an easy-to-use AI text-generation software for GGML and GGUF models. It's a single self contained distributable from Concedo, that builds off llama.cpp, and adds a versatile Kobold API endpoint, additional format support, Stable Diffusion image generation, backward compatibility, as well as a fancy UI with persistent stories, editing tools, save formats, memory, world info, author's note, characters, scenarios and everything Kobold and Kobold Lite have to offer. +KoboldCpp is an easy-to-use AI text-generation software for GGML and GGUF models. It's a single self contained distributable from Concedo, that builds off llama.cpp, and adds a versatile Kobold API endpoint, additional format support, Stable Diffusion image generation, backward compatibility, as well as a fancy UI with persistent stories, editing tools, save formats, memory, world info, author's note, characters, scenarios and everything KoboldAI and KoboldAI Lite have to offer. ![Preview](media/preview.png) ![Preview](media/preview2.png) @@ -20,7 +20,7 @@ KoboldCpp is an easy-to-use AI text-generation software for GGML and GGUF models - **(Nvidia Only) GPU Acceleration**: If you're on Windows with an Nvidia GPU you can get CUDA support out of the box using the `--usecublas` flag, make sure you select the correct .exe with CUDA support. - **Any GPU Acceleration**: As a slightly slower alternative, try CLBlast with `--useclblast` flags for a slightly slower but more GPU compatible speedup. - **GPU Layer Offloading**: Want even more speedup? Combine one of the above GPU flags with `--gpulayers` to offload entire layers to the GPU! **Much faster, but uses more VRAM**. Experiment to determine number of layers to offload, and reduce by a few if you run out of memory. -- **Increasing Context Size**: Try `--contextsize 4096` to 2x your context size! without much perplexity gain. Note that you'll have to increase the max context in the Kobold Lite UI as well (click and edit the number text field). +- **Increasing Context Size**: Try `--contextsize 4096` to 2x your context size! without much perplexity gain. Note that you'll have to increase the max context in the KoboldAI Lite UI as well (click and edit the number text field). - If you are having crashes or issues, you can try turning off BLAS with the `--noblas` flag. You can also try running in a non-avx2 compatibility mode with `--noavx2`. Lastly, you can try turning off mmap with `--nommap`. For more information, be sure to run the program with the `--help` flag, or [check the wiki](https://github.com/LostRuins/koboldcpp/wiki). @@ -122,7 +122,7 @@ There are some community made AUR packages available: [CUBLAS](https://aur.archl ## License - The original GGML library and llama.cpp by ggerganov are licensed under the MIT License -- However, Kobold Lite is licensed under the AGPL v3.0 License +- However, KoboldAI Lite is licensed under the AGPL v3.0 License - The other files are also under the AGPL v3.0 License unless otherwise stated ## Notes diff --git a/kcpp_docs.embd b/kcpp_docs.embd index e27f9230e..1f402feca 100644 --- a/kcpp_docs.embd +++ b/kcpp_docs.embd @@ -607,7 +607,7 @@ "description": "Successful request" } }, - "description": "Retrieves the KoboldCpp preloaded story, --preloadstory configures a prepared story json save file to be hosted on the server, which frontends (such as Kobold Lite) can access over the API.", + "description": "Retrieves the KoboldCpp preloaded story, --preloadstory configures a prepared story json save file to be hosted on the server, which frontends (such as KoboldAI Lite) can access over the API.", "summary": "Retrieves the KoboldCpp preloaded story", "tags": [ "api/extra" diff --git a/klite.embd b/klite.embd index 81003b4bc..85d7bb4b6 100644 --- a/klite.embd +++ b/klite.embd @@ -1,12 +1,12 @@ @@ -3400,7 +3400,7 @@ Current version: 146 if ("Notification" in window) { // Request permission to show notifications if (Notification.permission === "granted" || notify_allowed) { - var notification = new Notification("Kobold Lite", { + var notification = new Notification("KoboldAI Lite", { body: "Text Generation Completed!" }); } else { @@ -4295,7 +4295,7 @@ Current version: 146 document.getElementById("customkoboldendpoint").value = protocol + localmodehost + ":" + localmodeport + sublocalpathname; connect_custom_endpoint(); document.getElementById("lastreq").innerHTML = document.getElementById("lastreq2").innerHTML = - `You're using Kobold Lite Embedded.`; + `You're using KoboldAI Lite Embedded.`; read_url_params_data(); } @@ -4605,7 +4605,7 @@ Current version: 146 function set_dalle_key() { - inputBox("Enter DALL-E API Key.\n\nNote: DALL-E is known to rephrase and rewrite submitted image prompts before generating, for censorship purposes. There is nothing Kobold Lite can do about that. ","DALL-E API Key",localsettings.saved_dalle_key,"Input DALL-E API Key", ()=>{ + inputBox("Enter DALL-E API Key.\n\nNote: DALL-E is known to rephrase and rewrite submitted image prompts before generating, for censorship purposes. There is nothing KoboldAI Lite can do about that. ","DALL-E API Key",localsettings.saved_dalle_key,"Input DALL-E API Key", ()=>{ let userinput = getInputBoxValue(); userinput = userinput.trim(); if (userinput != null && userinput!="") { @@ -4615,7 +4615,7 @@ Current version: 146 } function set_dalle_url() { - inputBox("Enter DALL-E API URL.\n\nNote: DALL-E is known to rephrase and rewrite submitted image prompts before generating, for censorship purposes. There is nothing Kobold Lite can do about that. ","DALL-E API URL",localsettings.saved_dalle_url,"Input DALL-E API URL", ()=>{ + inputBox("Enter DALL-E API URL.\n\nNote: DALL-E is known to rephrase and rewrite submitted image prompts before generating, for censorship purposes. There is nothing KoboldAI Lite can do about that. ","DALL-E API URL",localsettings.saved_dalle_url,"Input DALL-E API URL", ()=>{ let userinput = getInputBoxValue(); userinput = userinput.trim(); if (userinput != null && userinput!="") { @@ -15400,7 +15400,7 @@ Current version: 146

Loading...

- +
@@ -16404,7 +16404,7 @@ Current version: 146
Run In Background ?Prevents the browser from suspending Kobold Lite by playing a silent audio track. This setting cannot be saved.
+ class="helptext">Prevents the browser from suspending KoboldAI Lite by playing a silent audio track. This setting cannot be saved.
diff --git a/koboldcpp.py b/koboldcpp.py index 55257ccc2..129a761ff 100644 --- a/koboldcpp.py +++ b/koboldcpp.py @@ -6,7 +6,7 @@ # and adds a versatile Kobold API endpoint, additional format support, # backward compatibility, as well as a fancy UI with persistent stories, # editing tools, save formats, memory, world info, author's note, characters, -# scenarios and everything Kobold and Kobold Lite have to offer. +# scenarios and everything Kobold and KoboldAI Lite have to offer. import ctypes import os, math, re @@ -1165,7 +1165,7 @@ class ServerRequestHandler(http.server.SimpleHTTPRequestHandler):

KoboldCpp NoScript Mode

KoboldCpp can be used without Javascript enabled, however this is not recommended. -
If you have Javascript, please use Kobold Lite WebUI instead.


+
If you have Javascript, please use KoboldAI Lite WebUI instead.


Enter Prompt:
@@ -1201,7 +1201,7 @@ Enter Prompt:
if self.path in ["", "/?"] or self.path.startswith(('/?','?')): #it's possible for the root url to have ?params without / content_type = 'text/html' if embedded_kailite is None: - response_body = (f"Embedded Kobold Lite is not found.
You will have to connect via the main KoboldAI client, or use this URL to connect.").encode() + response_body = (f"Embedded KoboldAI Lite is not found.
You will have to connect via the main KoboldAI client, or use this URL to connect.").encode() else: response_body = embedded_kailite @@ -3478,14 +3478,14 @@ def main(launch_args,start_server=True): with open(os.path.join(basepath, "klite.embd"), mode='rb') as f: embedded_kailite = f.read() # patch it with extra stuff - origStr = "Sorry, Kobold Lite requires Javascript to function." - patchedStr = "Sorry, Kobold Lite requires Javascript to function.
You can use KoboldCpp NoScript mode instead." + origStr = "Sorry, KoboldAI Lite requires Javascript to function." + patchedStr = "Sorry, KoboldAI Lite requires Javascript to function.
You can use KoboldCpp NoScript mode instead." embedded_kailite = embedded_kailite.decode("UTF-8","ignore") embedded_kailite = embedded_kailite.replace(origStr, patchedStr) embedded_kailite = embedded_kailite.encode() - print("Embedded Kobold Lite loaded.") + print("Embedded KoboldAI Lite loaded.") except Exception as e: - print("Could not find Kobold Lite. Embedded Kobold Lite will not be available.") + print("Could not find KoboldAI Lite. Embedded KoboldAI Lite will not be available.") try: basepath = os.path.abspath(os.path.dirname(__file__)) @@ -3698,7 +3698,7 @@ if __name__ == '__main__': advparser.add_argument("--remotetunnel", help="Uses Cloudflare to create a remote tunnel, allowing you to access koboldcpp remotely over the internet even behind a firewall.", action='store_true') advparser.add_argument("--highpriority", help="Experimental flag. If set, increases the process CPU priority, potentially speeding up generation. Use caution.", action='store_true') advparser.add_argument("--foreground", help="Windows only. Sends the terminal to the foreground every time a new prompt is generated. This helps avoid some idle slowdown issues.", action='store_true') - advparser.add_argument("--preloadstory", help="Configures a prepared story json save file to be hosted on the server, which frontends (such as Kobold Lite) can access over the API.", default="") + advparser.add_argument("--preloadstory", help="Configures a prepared story json save file to be hosted on the server, which frontends (such as KoboldAI Lite) can access over the API.", default="") advparser.add_argument("--quiet", help="Enable quiet mode, which hides generation inputs and outputs in the terminal. Quiet mode is automatically enabled when running a horde worker.", action='store_true') advparser.add_argument("--ssl", help="Allows all content to be served over SSL instead. A valid UNENCRYPTED SSL cert and key .pem files must be provided", metavar=('[cert_pem]', '[key_pem]'), nargs='+') advparser.add_argument("--nocertify", help="Allows insecure SSL connections. Use this if you have cert errors and need to bypass certificate restrictions.", action='store_true')