updated the readme for more docker information to make it clearer what to expect. please don't use the docker on a M series macOS

This commit is contained in:
Concedo 2026-04-27 18:21:22 +08:00
parent 095cfd6354
commit 9b38d83377

View file

@ -61,8 +61,9 @@ Finally, obtain and load a GGUF model. See [here](#Obtaining-a-GGUF-model)
- KoboldCpp can now be used on RunPod cloud GPUs! This is an easy way to get started without installing anything in a minute or two, and is very scalable, capable of running 70B+ models at afforable cost. [Try our RunPod image here!](https://koboldai.org/runpodcpp). Alternatively, you can also try [SimplePod](https://koboldai.org/simplepod) for smaller models
## Docker
- The official docker can be found at https://hub.docker.com/r/koboldai/koboldcpp
- If you're building your own docker, remember to enable LLAMA_PORTABLE
- Caution: The KoboldCpp docker is intended for experts only! If you're not an experienced user, you're recommended to use the [precompiled binaries directly instead](https://github.com/LostRuins/koboldcpp/releases/latest)
- The docker uses a x86-64 Ubuntu Linux based environment interally, and expects a Nvidia or AMD GPU. It may perform suboptimally on some Windows and MacOS devices, and may outright fail for ARM. AVX/AVX2 feature detection may not work on some systems, resulting in the failsafe binaries being loaded (will become very slow).
- If you still want to proceed, the official docker can be found at https://hub.docker.com/r/koboldai/koboldcpp
## Obtaining a GGUF model
- KoboldCpp uses GGUF models. They are not included with KoboldCpp, but you can download GGUF files from other places such as [Bartowski's Huggingface](https://huggingface.co/bartowski). Search for "GGUF" on huggingface.co for plenty of compatible models in the `.gguf` format.