diff --git a/README.md b/README.md index d58ba3faa..f87855609 100644 --- a/README.md +++ b/README.md @@ -61,8 +61,9 @@ Finally, obtain and load a GGUF model. See [here](#Obtaining-a-GGUF-model) - KoboldCpp can now be used on RunPod cloud GPUs! This is an easy way to get started without installing anything in a minute or two, and is very scalable, capable of running 70B+ models at afforable cost. [Try our RunPod image here!](https://koboldai.org/runpodcpp). Alternatively, you can also try [SimplePod](https://koboldai.org/simplepod) for smaller models ## Docker -- The official docker can be found at https://hub.docker.com/r/koboldai/koboldcpp -- If you're building your own docker, remember to enable LLAMA_PORTABLE +- Caution: The KoboldCpp docker is intended for experts only! If you're not an experienced user, you're recommended to use the [precompiled binaries directly instead](https://github.com/LostRuins/koboldcpp/releases/latest) +- The docker uses a x86-64 Ubuntu Linux based environment interally, and expects a Nvidia or AMD GPU. It may perform suboptimally on some Windows and MacOS devices, and may outright fail for ARM. AVX/AVX2 feature detection may not work on some systems, resulting in the failsafe binaries being loaded (will become very slow). +- If you still want to proceed, the official docker can be found at https://hub.docker.com/r/koboldai/koboldcpp ## Obtaining a GGUF model - KoboldCpp uses GGUF models. They are not included with KoboldCpp, but you can download GGUF files from other places such as [Bartowski's Huggingface](https://huggingface.co/bartowski). Search for "GGUF" on huggingface.co for plenty of compatible models in the `.gguf` format.