mirror of
https://github.com/LostRuins/koboldcpp.git
synced 2025-09-10 17:14:36 +00:00
fix colab
This commit is contained in:
parent
6c6ad93f01
commit
60d234550b
1 changed files with 4 additions and 4 deletions
|
@ -54,8 +54,8 @@
|
||||||
"ForceRebuild = False #@param {type:\"boolean\"}\r\n",
|
"ForceRebuild = False #@param {type:\"boolean\"}\r\n",
|
||||||
"#@markdown <hr>\r\n",
|
"#@markdown <hr>\r\n",
|
||||||
"LoadLLaVAmmproj = False #@param {type:\"boolean\"}\r\n",
|
"LoadLLaVAmmproj = False #@param {type:\"boolean\"}\r\n",
|
||||||
"LLaVAmmproj = \"https://huggingface.co/concedo/koboldcpp-mmproj/resolve/main/llama-13b-mmproj-v1.5.Q4_1.gguf\" #@param [\"https://huggingface.co/concedo/koboldcpp-mmproj/resolve/main/llama-13b-mmproj-v1.5.Q4_1.gguf\"]{allow-input: true}\r\n",
|
"LLaVAmmproj = \"https://huggingface.co/concedo/koboldcpp-mmproj/resolve/main/llama-13b-mmproj-v1.5.Q4_1.gguf\" #@param [\"https://huggingface.co/concedo/koboldcpp-mmproj/resolve/main/llama-13b-mmproj-v1.5.Q4_1.gguf\",\"https://huggingface.co/concedo/koboldcpp-mmproj/resolve/main/mistral-7b-mmproj-v1.5-Q4_1.gguf\",\"https://huggingface.co/concedo/koboldcpp-mmproj/resolve/main/llama-7b-mmproj-v1.5-Q4_0.gguf\"]{allow-input: true}\r\n",
|
||||||
"Vcommand = \"\"\r\n",
|
"VCommand = \"\"\r\n",
|
||||||
"#@markdown <hr>\r\n",
|
"#@markdown <hr>\r\n",
|
||||||
"LoadImgModel = False #@param {type:\"boolean\"}\r\n",
|
"LoadImgModel = False #@param {type:\"boolean\"}\r\n",
|
||||||
"ImgModel = \"https://huggingface.co/koboldcpp/imgmodel/resolve/main/imgmodel_older_q4_0.gguf\" #@param [\"https://huggingface.co/koboldcpp/imgmodel/resolve/main/imgmodel_older_q4_0.gguf\"]{allow-input: true}\r\n",
|
"ImgModel = \"https://huggingface.co/koboldcpp/imgmodel/resolve/main/imgmodel_older_q4_0.gguf\" #@param [\"https://huggingface.co/koboldcpp/imgmodel/resolve/main/imgmodel_older_q4_0.gguf\"]{allow-input: true}\r\n",
|
||||||
|
@ -73,7 +73,7 @@
|
||||||
"if ForceRebuild:\r\n",
|
"if ForceRebuild:\r\n",
|
||||||
" kvers = \"force_rebuild\"\r\n",
|
" kvers = \"force_rebuild\"\r\n",
|
||||||
"if LLaVAmmproj and LoadLLaVAmmproj:\r\n",
|
"if LLaVAmmproj and LoadLLaVAmmproj:\r\n",
|
||||||
" Vcommand = \"--mmproj vmodel.gguf\"\r\n",
|
" VCommand = \"--mmproj vmodel.gguf\"\r\n",
|
||||||
"else:\r\n",
|
"else:\r\n",
|
||||||
" SCommand = \"\"\r\n",
|
" SCommand = \"\"\r\n",
|
||||||
"if ImgModel and LoadImgModel:\r\n",
|
"if ImgModel and LoadImgModel:\r\n",
|
||||||
|
@ -92,7 +92,7 @@
|
||||||
" !aria2c -x 10 -o vmodel.gguf --summary-interval=5 --download-result=default --allow-overwrite=true --file-allocation=none $LLaVAmmproj\r\n",
|
" !aria2c -x 10 -o vmodel.gguf --summary-interval=5 --download-result=default --allow-overwrite=true --file-allocation=none $LLaVAmmproj\r\n",
|
||||||
"if SCommand:\r\n",
|
"if SCommand:\r\n",
|
||||||
" !aria2c -x 10 -o imodel.gguf --summary-interval=5 --download-result=default --allow-overwrite=true --file-allocation=none $ImgModel\r\n",
|
" !aria2c -x 10 -o imodel.gguf --summary-interval=5 --download-result=default --allow-overwrite=true --file-allocation=none $ImgModel\r\n",
|
||||||
"!python koboldcpp.py model.gguf --usecublas 0 mmq --multiuser --gpulayers $Layers --contextsize $ContextSize --quiet --remotetunnel $Vcommand $SCommand\r\n"
|
"!python koboldcpp.py model.gguf --usecublas 0 mmq --multiuser --gpulayers $Layers --contextsize $ContextSize --quiet --remotetunnel $VCommand $SCommand\r\n"
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
],
|
],
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue