diff --git a/klite.embd b/klite.embd index 93b69c72d..780260405 100644 --- a/klite.embd +++ b/klite.embd @@ -946,7 +946,7 @@ Current version: 145 .statusicon:hover .statustext, .statusicon.statustoggled .statustext { display: inline-block; - width: 250px; + width: 280px; background-color: #1f2931; color: #ffffff; font-size: 11pt; @@ -11075,6 +11075,26 @@ Current version: 145 } if (document.getElementById("useoaichatcompl").checked) { + let mainoaibody = submit_payload.prompt; //can be string or array + if(insertAIVisionImages.length>0) + { + mainoaibody = [ + { + "type": "text", + "text": mainoaibody + } + ]; + for(let i=0;iOpenAI API (Conditional)`:`Unsupported`; + } + else if(savedmeta.visionmode==3) { visionstatus = ((!savedmeta.visionmode || savedmeta.visionmode==0)?`Inactive`:(hasllava?`Active`:`Unsupported`)); } @@ -11967,14 +11992,15 @@ Current version: 145 visionstatus = ((!savedmeta.visionmode || savedmeta.visionmode==0)?`Inactive`:(savedmeta.desc?`Active`:`Analyzing`)); } - let togglebtn = ` + `; document.getElementById("zoomedimgdesc").innerHTML = ` - AI Vision: `+visionstatus+` ?This allows the AI to visually recognize this image and react to it. On KoboldCpp, LLaVA models can be used. Otherwise, uses Horde or Local A1111 for image interrogation if enabled. + AI Vision: `+visionstatus+` ?Allows the AI to see and react to this image. On KoboldCpp, LLaVA models can be used. Horde or Local A1111 use image interrogation if enabled. For OpenAI API, only works with Vision Models like Gpt4o. `+togglebtn+`
@@ -13267,7 +13293,7 @@ Current version: 145 { return "\n(Attached Image: " + foundmeta.desc + ")\n"; } - else if(foundmeta.visionmode==3) + else if(foundmeta.visionmode==3 || foundmeta.visionmode==4) { let parts = inner.split(','); if (parts.length === 2 && parts[0].startsWith('data:image')) {