mirror of
https://github.com/LostRuins/koboldcpp.git
synced 2025-09-11 01:24:36 +00:00
updated readme, memory detection prints
This commit is contained in:
parent
11c4e7c2c4
commit
8e23a087e7
2 changed files with 5 additions and 9 deletions
|
@ -19,7 +19,7 @@ KoboldCpp is an easy-to-use AI text-generation software for GGML and GGUF models
|
|||
- Includes multiple modes (chat, adventure, instruct, storywriter) and UI Themes (aesthetic roleplay, classic writer, corporate assistant, messsenger)
|
||||
- Supports loading Tavern Character Cards, importing many different data formats from various sites, reading or exporting JSON savefiles and persistent stories.
|
||||
- Many other features including new samplers, regex support, websearch, RAG via TextDB and more.
|
||||
- Ready-to-use binaries for Windows, MacOS, Linux, Android (via Termux), Colab, Docker, also supports other platforms if self-compiled (like Raspberry PI).
|
||||
- Ready-to-use binaries for Windows, MacOS, Linux. Runs directly with Colab, Docker, also supports other platforms if self-compiled (like Android (via Termux) and Raspberry PI).
|
||||
- [Need help finding a model? Read this!](https://github.com/LostRuins/koboldcpp/wiki#getting-an-ai-model-file)
|
||||
|
||||
## Windows Usage (Precompiled Binary, Recommended)
|
||||
|
@ -31,7 +31,7 @@ KoboldCpp is an easy-to-use AI text-generation software for GGML and GGUF models
|
|||
- You can also run it using the command line. For info, please check `koboldcpp.exe --help`
|
||||
|
||||
## Linux Usage (Precompiled Binary, Recommended)
|
||||
On modern Linux systems, you should download the `koboldcpp-linux-x64-cuda1150` prebuilt PyInstaller binary on the **[releases page](https://github.com/LostRuins/koboldcpp/releases/latest)**. Simply download and run the binary (You may have to `chmod +x` it first).
|
||||
On modern Linux systems, you should download the `koboldcpp-linux-x64-cuda1150` prebuilt PyInstaller binary for greatest compatibility on the **[releases page](https://github.com/LostRuins/koboldcpp/releases/latest)**. Simply download and run the binary (You may have to `chmod +x` it first). If you have a newer device, you can also try the `koboldcpp-linux-x64-cuda1210` instead for better speeds.
|
||||
|
||||
Alternatively, you can also install koboldcpp to the current directory by running the following terminal command:
|
||||
```
|
||||
|
|
10
koboldcpp.py
10
koboldcpp.py
|
@ -978,7 +978,6 @@ def fetch_gpu_properties(testCL,testCU,testVK):
|
|||
FetchedCUdevices = []
|
||||
FetchedCUdeviceMem = []
|
||||
FetchedCUfreeMem = []
|
||||
faileddetectvram = False
|
||||
|
||||
AMDgpu = None
|
||||
try: # Get NVIDIA GPU names
|
||||
|
@ -989,7 +988,6 @@ def fetch_gpu_properties(testCL,testCU,testVK):
|
|||
except Exception:
|
||||
FetchedCUdeviceMem = []
|
||||
FetchedCUfreeMem = []
|
||||
faileddetectvram = True
|
||||
pass
|
||||
if len(FetchedCUdevices)==0:
|
||||
try: # Get AMD ROCm GPU names
|
||||
|
@ -1011,7 +1009,6 @@ def fetch_gpu_properties(testCL,testCU,testVK):
|
|||
except Exception:
|
||||
FetchedCUdeviceMem = []
|
||||
FetchedCUfreeMem = []
|
||||
faileddetectvram = True
|
||||
pass
|
||||
lowestcumem = 0
|
||||
lowestfreecumem = 0
|
||||
|
@ -1030,14 +1027,13 @@ def fetch_gpu_properties(testCL,testCU,testVK):
|
|||
except Exception:
|
||||
lowestcumem = 0
|
||||
lowestfreecumem = 0
|
||||
faileddetectvram = True
|
||||
|
||||
if faileddetectvram:
|
||||
print("Unable to detect VRAM, please set layers manually.")
|
||||
|
||||
MaxMemory[0] = max(lowestcumem,MaxMemory[0])
|
||||
MaxFreeMemory[0] = max(lowestfreecumem,MaxFreeMemory[0])
|
||||
|
||||
if MaxMemory[0] < (1024*1024*256):
|
||||
print("Unable to detect VRAM, please set layers manually.")
|
||||
|
||||
if testVK:
|
||||
try: # Get Vulkan names
|
||||
foundVkGPU = False
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue