mirror of
https://github.com/LostRuins/koboldcpp.git
synced 2026-05-06 16:21:49 +00:00
* sd: remove C++ support for enforcing fixed LoRA multipliers The logic at the Python level is enough. * sd: support changing preloaded LoRA multipliers We keep the same rules as before: - Any LoRA with multiplier 0 can be changed - If all LoRAs have multiplier != 0, they are fixed and optimized but tweak the corner case of LoRAs specified more than once to allow adjusting the multiplier if the same LoRA is also specified with a zero multiplier, as if they were two different LoRAs. So the following keeps working as before: - --sdlora /loras/lcm.gguf --sdloramult 1 : fixed as 1 - --sdlora /loras/lcm.gguf --sdloramult 0 : dynamic, default 0 - --sdlora /loras/ : dynamic, default 0 - --sdlora /loras/lcm.gguf /loras/lcm.gguf --sdloramult 1 1 : fixed as 2 But now we have: - --sdlora /loras/lcm.gguf /loras/lcm.gguf --sdloramult 1 0 : dynamic, default 1 - --sdlora /loras/lcm.gguf /loras/ --sdloramult 1 : dynamic, default 1 |
||
|---|---|---|
| .. | ||
| acestep | ||
| qwen3tts | ||
| sdcpp | ||
| tools | ||
| ttscpp | ||
| whispercpp | ||
| embeddings_adapter.cpp | ||
| ggml_v1.c | ||
| ggml_v1.h | ||
| ggml_v2-cuda-legacy.cu | ||
| ggml_v2-cuda-legacy.h | ||
| ggml_v2-cuda.cu | ||
| ggml_v2-cuda.h | ||
| ggml_v2.c | ||
| ggml_v2.h | ||
| ggml_v3-cuda.cu | ||
| ggml_v3-cuda.h | ||
| ggml_v3.c | ||
| ggml_v3.h | ||
| gpt2_v1.cpp | ||
| gpt2_v2.cpp | ||
| gpt2_v3.cpp | ||
| gptj_v1.cpp | ||
| gptj_v2.cpp | ||
| gptj_v3.cpp | ||
| llama-util.h | ||
| llama_v2-util.h | ||
| llama_v2.cpp | ||
| llama_v2.h | ||
| llama_v3.cpp | ||
| llama_v3.h | ||
| mpt_v3.cpp | ||
| neox_v2.cpp | ||
| neox_v3.cpp | ||
| otherarch.h | ||
| rwkv_v2.cpp | ||
| rwkv_v2.h | ||
| rwkv_v3.cpp | ||
| rwkv_v3.h | ||
| rwkv_vocab.cpp | ||
| tts_adapter.cpp | ||
| utils.cpp | ||
| utils.h | ||