From 5ebc532ca9dd8ed3c45655d61e4c5eea3ec3874d Mon Sep 17 00:00:00 2001 From: Concedo <39025047+LostRuins@users.noreply.github.com> Date: Mon, 3 Jun 2024 14:55:12 +0800 Subject: [PATCH] update colab --- colab.ipynb | 12 +++++++++++- 1 file changed, 11 insertions(+), 1 deletion(-) diff --git a/colab.ipynb b/colab.ipynb index 6ffb84af5..1123fcc94 100644 --- a/colab.ipynb +++ b/colab.ipynb @@ -60,6 +60,10 @@ "LoadImgModel = False #@param {type:\"boolean\"}\r\n", "ImgModel = \"https://huggingface.co/koboldcpp/imgmodel/resolve/main/imgmodel_ftuned_q4_0.gguf\" #@param [\"https://huggingface.co/koboldcpp/imgmodel/resolve/main/imgmodel_ftuned_q4_0.gguf\"]{allow-input: true}\r\n", "SCommand = \"\"\r\n", + "#@markdown
\r\n", + "LoadSpeechModel = False #@param {type:\"boolean\"}\r\n", + "SpeechModel = \"https://huggingface.co/koboldcpp/whisper/resolve/main/whisper-base.en-q5_1.bin\" #@param [\"https://huggingface.co/koboldcpp/whisper/resolve/main/whisper-base.en-q5_1.bin\"]{allow-input: true}\r\n", + "WCommand = \"\"\r\n", "\r\n", "import os\r\n", "if not os.path.isfile(\"/opt/bin/nvidia-smi\"):\r\n", @@ -80,6 +84,10 @@ " SCommand = \"--sdmodel imodel.gguf --sdthreads 4 --sdquant --sdclamped\"\r\n", "else:\r\n", " SCommand = \"\"\r\n", + "if SpeechModel and LoadSpeechModel:\r\n", + " WCommand = \"--whispermodel wmodel.bin\"\r\n", + "else:\r\n", + " WCommand = \"\"\r\n", "!echo Finding prebuilt binary for {kvers}\r\n", "!wget -O dlfile.tmp https://kcppcolab.concedo.workers.dev/?{kvers} && mv dlfile.tmp koboldcpp_cublas.so\r\n", "!test -f koboldcpp_cublas.so && echo Prebuilt Binary Exists || echo Prebuilt Binary Does Not Exist\r\n", @@ -92,7 +100,9 @@ " !aria2c -x 10 -o vmodel.gguf --summary-interval=5 --download-result=default --allow-overwrite=true --file-allocation=none $LLaVAmmproj\r\n", "if SCommand:\r\n", " !aria2c -x 10 -o imodel.gguf --summary-interval=5 --download-result=default --allow-overwrite=true --file-allocation=none $ImgModel\r\n", - "!python koboldcpp.py model.gguf --usecublas 0 mmq --multiuser --gpulayers $Layers --contextsize $ContextSize --quiet --remotetunnel $VCommand $SCommand\r\n" + "if WCommand:\r\n", + " !aria2c -x 10 -o wmodel.bin --summary-interval=5 --download-result=default --allow-overwrite=true --file-allocation=none $SpeechModel\r\n", + "!python koboldcpp.py model.gguf --usecublas 0 mmq --multiuser --gpulayers $Layers --contextsize $ContextSize --quiet --remotetunnel $VCommand $SCommand $WCommand\r\n" ] } ],