mirror of
https://github.com/LostRuins/koboldcpp.git
synced 2026-04-28 03:30:20 +00:00
modify llama.cpp branding on lcpp ui (+1 squashed commits)
Squashed commits: [067343edf] modify llama.cpp branding on lcpp ui
This commit is contained in:
parent
d27949f22a
commit
8c17541cc0
6 changed files with 6 additions and 6 deletions
|
|
@ -3833,7 +3833,7 @@ Change Mode<br>
|
|||
elif self.path.endswith(('/.well-known/serviceinfo')):
|
||||
response_body = (json.dumps({"version":"0.2","software":{"name":"KoboldCpp","version":KcppVersion,"repository":"https://github.com/LostRuins/koboldcpp","homepage":"https://github.com/LostRuins/koboldcpp","logo":"https://raw.githubusercontent.com/LostRuins/koboldcpp/refs/heads/concedo/niko.ico"},"api":{"koboldai":{"name":"KoboldAI API","rel_url":"/api","documentation":"https://lite.koboldai.net/koboldcpp_api","version":KcppVersion},"openai":{"name":"OpenAI API","rel_url ":"/v1","documentation":"https://openai.com/documentation/api","version":KcppVersion}}}).encode())
|
||||
|
||||
elif self.path=="/props":
|
||||
elif self.path.split("?")[0]=="/props":
|
||||
response_body = (json.dumps({
|
||||
"chat_template": cached_chat_template,
|
||||
"id": 0,
|
||||
|
|
|
|||
|
|
@ -437,7 +437,7 @@
|
|||
>
|
||||
<div class="w-full max-w-[48rem] px-4">
|
||||
<div class="mb-10 text-center" in:fade={{ duration: 300 }}>
|
||||
<h1 class="mb-4 text-3xl font-semibold tracking-tight">llama.cpp</h1>
|
||||
<h1 class="mb-4 text-3xl font-semibold tracking-tight">llama.cpp UI (KoboldCpp Backend)</h1>
|
||||
|
||||
<p class="text-lg text-muted-foreground">
|
||||
{serverStore.props?.modalities?.audio
|
||||
|
|
|
|||
|
|
@ -103,7 +103,7 @@
|
|||
<ScrollArea class="h-[100vh]">
|
||||
<Sidebar.Header class=" top-0 z-10 gap-6 bg-sidebar/50 px-4 py-4 pb-2 backdrop-blur-lg md:sticky">
|
||||
<a href="#/" onclick={handleMobileSidebarItemClick}>
|
||||
<h1 class="inline-flex items-center gap-1 px-2 text-xl font-semibold">llama.cpp</h1>
|
||||
<h1 class="inline-flex items-center gap-1 px-2 text-xl font-semibold">llama.cpp UI</h1>
|
||||
</a>
|
||||
|
||||
<ChatSidebarActions {handleMobileSidebarItemClick} bind:isSearchModeActive bind:searchQuery />
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@
|
|||
message?: string;
|
||||
}
|
||||
|
||||
let { class: className = '', message = 'Initializing connection to llama.cpp server...' }: Props =
|
||||
let { class: className = '', message = 'Initializing connection to KoboldCpp server...' }: Props =
|
||||
$props();
|
||||
</script>
|
||||
|
||||
|
|
|
|||
|
|
@ -79,7 +79,7 @@
|
|||
</script>
|
||||
|
||||
<svelte:head>
|
||||
<title>llama.cpp - AI Chat Interface</title>
|
||||
<title>llama.cpp UI (KoboldCpp Backend)</title>
|
||||
</svelte:head>
|
||||
|
||||
<ChatScreen showCenteredEmpty={true} />
|
||||
|
|
|
|||
|
|
@ -164,7 +164,7 @@
|
|||
</script>
|
||||
|
||||
<svelte:head>
|
||||
<title>{activeConversation()?.name || 'Chat'} - llama.cpp</title>
|
||||
<title>{activeConversation()?.name || 'Chat'} - llama.cpp UI (KoboldCpp Backend)</title>
|
||||
</svelte:head>
|
||||
|
||||
<ChatScreen />
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue