diff --git a/README.md b/README.md index bc5d603e0..63c356861 100644 --- a/README.md +++ b/README.md @@ -19,7 +19,7 @@ KoboldCpp is an easy-to-use AI text-generation software for GGML and GGUF models - Includes multiple modes (chat, adventure, instruct, storywriter) and UI Themes (aesthetic roleplay, classic writer, corporate assistant, messsenger) - Supports loading Tavern Character Cards, importing many different data formats from various sites, reading or exporting JSON savefiles and persistent stories. - Many other features including new samplers, regex support, websearch, RAG via TextDB and more. -- Ready-to-use binaries for Windows, MacOS, Linux, Android (via Termux), Colab, Docker, also supports other platforms if self-compiled (like Raspberry PI). +- Ready-to-use binaries for Windows, MacOS, Linux. Runs directly with Colab, Docker, also supports other platforms if self-compiled (like Android (via Termux) and Raspberry PI). - [Need help finding a model? Read this!](https://github.com/LostRuins/koboldcpp/wiki#getting-an-ai-model-file) ## Windows Usage (Precompiled Binary, Recommended) @@ -31,7 +31,7 @@ KoboldCpp is an easy-to-use AI text-generation software for GGML and GGUF models - You can also run it using the command line. For info, please check `koboldcpp.exe --help` ## Linux Usage (Precompiled Binary, Recommended) -On modern Linux systems, you should download the `koboldcpp-linux-x64-cuda1150` prebuilt PyInstaller binary on the **[releases page](https://github.com/LostRuins/koboldcpp/releases/latest)**. Simply download and run the binary (You may have to `chmod +x` it first). +On modern Linux systems, you should download the `koboldcpp-linux-x64-cuda1150` prebuilt PyInstaller binary for greatest compatibility on the **[releases page](https://github.com/LostRuins/koboldcpp/releases/latest)**. Simply download and run the binary (You may have to `chmod +x` it first). If you have a newer device, you can also try the `koboldcpp-linux-x64-cuda1210` instead for better speeds. Alternatively, you can also install koboldcpp to the current directory by running the following terminal command: ``` diff --git a/koboldcpp.py b/koboldcpp.py index 9bbfc1f76..3506e1b27 100644 --- a/koboldcpp.py +++ b/koboldcpp.py @@ -978,7 +978,6 @@ def fetch_gpu_properties(testCL,testCU,testVK): FetchedCUdevices = [] FetchedCUdeviceMem = [] FetchedCUfreeMem = [] - faileddetectvram = False AMDgpu = None try: # Get NVIDIA GPU names @@ -989,7 +988,6 @@ def fetch_gpu_properties(testCL,testCU,testVK): except Exception: FetchedCUdeviceMem = [] FetchedCUfreeMem = [] - faileddetectvram = True pass if len(FetchedCUdevices)==0: try: # Get AMD ROCm GPU names @@ -1011,7 +1009,6 @@ def fetch_gpu_properties(testCL,testCU,testVK): except Exception: FetchedCUdeviceMem = [] FetchedCUfreeMem = [] - faileddetectvram = True pass lowestcumem = 0 lowestfreecumem = 0 @@ -1030,14 +1027,13 @@ def fetch_gpu_properties(testCL,testCU,testVK): except Exception: lowestcumem = 0 lowestfreecumem = 0 - faileddetectvram = True - - if faileddetectvram: - print("Unable to detect VRAM, please set layers manually.") MaxMemory[0] = max(lowestcumem,MaxMemory[0]) MaxFreeMemory[0] = max(lowestfreecumem,MaxFreeMemory[0]) + if MaxMemory[0] < (1024*1024*256): + print("Unable to detect VRAM, please set layers manually.") + if testVK: try: # Get Vulkan names foundVkGPU = False