From 60d234550b5748a2f9eeab103edfe951ca90907e Mon Sep 17 00:00:00 2001 From: Concedo <39025047+LostRuins@users.noreply.github.com> Date: Tue, 12 Mar 2024 20:09:49 +0800 Subject: [PATCH] fix colab --- colab.ipynb | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/colab.ipynb b/colab.ipynb index f775f76b9..08090b24e 100644 --- a/colab.ipynb +++ b/colab.ipynb @@ -54,8 +54,8 @@ "ForceRebuild = False #@param {type:\"boolean\"}\r\n", "#@markdown
\r\n", "LoadLLaVAmmproj = False #@param {type:\"boolean\"}\r\n", - "LLaVAmmproj = \"https://huggingface.co/concedo/koboldcpp-mmproj/resolve/main/llama-13b-mmproj-v1.5.Q4_1.gguf\" #@param [\"https://huggingface.co/concedo/koboldcpp-mmproj/resolve/main/llama-13b-mmproj-v1.5.Q4_1.gguf\"]{allow-input: true}\r\n", - "Vcommand = \"\"\r\n", + "LLaVAmmproj = \"https://huggingface.co/concedo/koboldcpp-mmproj/resolve/main/llama-13b-mmproj-v1.5.Q4_1.gguf\" #@param [\"https://huggingface.co/concedo/koboldcpp-mmproj/resolve/main/llama-13b-mmproj-v1.5.Q4_1.gguf\",\"https://huggingface.co/concedo/koboldcpp-mmproj/resolve/main/mistral-7b-mmproj-v1.5-Q4_1.gguf\",\"https://huggingface.co/concedo/koboldcpp-mmproj/resolve/main/llama-7b-mmproj-v1.5-Q4_0.gguf\"]{allow-input: true}\r\n", + "VCommand = \"\"\r\n", "#@markdown
\r\n", "LoadImgModel = False #@param {type:\"boolean\"}\r\n", "ImgModel = \"https://huggingface.co/koboldcpp/imgmodel/resolve/main/imgmodel_older_q4_0.gguf\" #@param [\"https://huggingface.co/koboldcpp/imgmodel/resolve/main/imgmodel_older_q4_0.gguf\"]{allow-input: true}\r\n", @@ -73,7 +73,7 @@ "if ForceRebuild:\r\n", " kvers = \"force_rebuild\"\r\n", "if LLaVAmmproj and LoadLLaVAmmproj:\r\n", - " Vcommand = \"--mmproj vmodel.gguf\"\r\n", + " VCommand = \"--mmproj vmodel.gguf\"\r\n", "else:\r\n", " SCommand = \"\"\r\n", "if ImgModel and LoadImgModel:\r\n", @@ -92,7 +92,7 @@ " !aria2c -x 10 -o vmodel.gguf --summary-interval=5 --download-result=default --allow-overwrite=true --file-allocation=none $LLaVAmmproj\r\n", "if SCommand:\r\n", " !aria2c -x 10 -o imodel.gguf --summary-interval=5 --download-result=default --allow-overwrite=true --file-allocation=none $ImgModel\r\n", - "!python koboldcpp.py model.gguf --usecublas 0 mmq --multiuser --gpulayers $Layers --contextsize $ContextSize --quiet --remotetunnel $Vcommand $SCommand\r\n" + "!python koboldcpp.py model.gguf --usecublas 0 mmq --multiuser --gpulayers $Layers --contextsize $ContextSize --quiet --remotetunnel $VCommand $SCommand\r\n" ] } ],