Commit graph

42 commits

Author SHA1 Message Date
Concedo
fcb658453e remove duplicate bundling for oldpc versions 2025-06-22 23:35:22 +08:00
Concedo
1970d8c9e8 uvos said it might work 2025-06-12 16:44:46 +08:00
Concedo
5cdb2d3fc6 cleanup 2025-06-11 01:35:40 +08:00
henk717
f151648f03 Pyinstaller launcher and dependency updates
This PR adds a new launcher executable to the unpack feature, eliminating the need to have python and its dependencies in the unpacked version. It also does a few dependency changes to help future proof.
2025-06-10 23:08:02 +08:00
Concedo
28b35ca879 allow wmma flag for rocm 2025-06-10 01:23:48 +08:00
Concedo
6c5c8be48d try to make rocm work for the github ci, requires disabling rocwmma 2025-06-08 21:52:29 +08:00
henk717
5d8f499f03
Remove 32GB of rocm dependencies with this one special trick (#1585)
* One file to remove them all

* That one lib wasn't versioned
2025-06-08 11:16:15 +08:00
Concedo
a80dfa5c10 various minor fixes 2025-06-08 01:11:42 +08:00
Concedo
8b141d8647 stick to cu12.1 for linux for now 2025-06-06 17:38:28 +08:00
Concedo
eec5a8ad16 breaking change: due to cuda12 upgrade, release filenames will change. standardize them to windows naming for the future. (+1 squashed commits)
Squashed commits:

[75842919a] cuda12.4 test
2025-06-06 14:02:34 +08:00
henk717
b8883e254a
KoboldCpp.sh updates (#1562)
* YR makefile upstream

* Create make_portable_rocm_libs.sh

* update makefile, support llama portable, ditch all unnecessary changes

* Delete make_portable_rocm_libs.sh should not be needed

* koboldcpp.sh updates

* Small rocm fixes

* ROCm is now a cuda version not a command

* Don't commit temp file

* Don't commit temp file

* 1200 has errors, removing it for now

* Only rebuild rocm with rebuild

* Update kcpp-build-release-linux.yaml

* Fix rocm filename

* ROCm Linux CI

* We need more diskspace

* Workaround for lockfile getting stuck

Why do I have to do hacks like this....

* Update kcpp-build-release-linux-rocm.yaml

* Dont apt update rocm

You don't allow us to apt update? Better not break things github!

* Container maybe?

* Turns out we aren't root, so we use sudo

* Cleanup ROCm CI PR

* Build for Runpods GPU

* We also need rocblas

* More cleanup just in case

* Update kcpp-build-release-linux-rocm.yaml

---------

Co-authored-by: LostRuins Concedo <39025047+LostRuins@users.noreply.github.com>
2025-05-26 15:24:49 +08:00
Concedo
8bd6f9f9ae added a simple cross platform launch script for unpacked dirs 2025-05-22 22:09:46 +08:00
Concedo
be3e93c76a bundle AGPL license and llama.cpp's MIT license into binaries. clarified some licensing terms, updated readme (+1 squashed commits)
Squashed commits:

[61c152daf] bundle AGPL license and llama.cpp's MIT license into binaries. clarified some licensing terms, updated readme
2025-05-18 02:21:27 +08:00
Concedo
b951310ca5 tryout smaller binaries 2025-05-07 14:56:34 +08:00
Concedo
7a7bdeab6d json to gbnf endpoint added 2025-04-12 11:41:11 +08:00
Concedo
143b611274 updated workflows 2025-03-19 21:56:35 +08:00
Concedo
2c9ade61fe test automatic vk shader rebuilding 2025-03-13 19:34:15 +08:00
Concedo
7a5499e77b added one more backend for clblast noavx2 and clblast failsafe 2025-01-30 22:47:22 +08:00
Concedo
bf4a52383f change of plans, we can't bundle numpy 2025-01-19 22:53:38 +08:00
Concedo
fa7e661133 various fixes 2025-01-18 23:52:39 +08:00
Concedo
3732bb2686 taesd now supports flux and sd3 2025-01-08 22:35:50 +08:00
Concedo
fd073fc904 try fix ci builds 2024-11-23 18:37:09 +08:00
Concedo
afc575fbd8 cleanup, try to add version tagging 2024-11-23 12:59:06 +08:00
Concedo
53bf0fb32d removed openblas backend, merged into CPU (with llamafile for BLAS). GPU backend is now automatically selected when running from CLI unless noblas is specified. 2024-09-15 19:21:52 +08:00
Concedo
8412946b9f fix oldcpu build avx1 2024-07-15 23:42:22 +08:00
Concedo
21179d675b try ci for avx1, up ver (+2 squashed commit)
Squashed commit:

[74150175] up version

[97b6163c] try ci for avx1 linux
2024-07-15 23:07:07 +08:00
Concedo
066e7ac540 minor fixes: colab gpu backend, lite bugs, package python file with embd 2024-07-15 17:36:03 +08:00
Concedo
11f0643fa4 fix pyinstallers 2024-06-27 15:19:44 +08:00
Concedo
a0ecd0d8e6 update build job count, updated lite 2024-06-18 21:12:16 +08:00
Concedo
2dedea9a74 add to remaining pyinstallers 2024-05-24 16:21:26 +08:00
henk717
8762880452
Harden koboldcpp.sh against conda hijacking (#857)
* Harden against conda hijacks

Turns out conda can override the location of our packages, by ignoring the conda rc files we can prevent this.

* Harden even further

Turns out not everything has a --no-rc option, now we hardcode the path.

* Remove --no-rc in run

Not supported in run, breaks the commands.
2024-05-21 21:43:44 +08:00
Concedo
5ce2fdad24 taesd for sdxl, add lora loading done 2024-05-14 23:02:56 +08:00
Concedo
5d15f8f76a vae test 2024-05-14 19:17:01 +08:00
Concedo
3667cc0113 fixed stableui btn (+4 squashed commit)
Squashed commit:

[1d4714f1] update default amount to gen

[6eacba33] updated lite

[033589af] added first ver sdui

[16f66d57] updated lite
2024-05-06 00:55:16 +08:00
henk717
b6bfab128f
CUDA 12 CI (#815)
* Allow KCPP_CUDA to specify CUDA version

* CUDA 12 CI Linux

* CUDA 12 CI

* Fix KCPP_CUDA indent

* KCPP_CUDA ENV Fix

StackOverflow is bad for advice sometimes....

* Lowcase cuda on output filename

* Strip . from filename output
2024-05-03 17:12:57 +08:00
Concedo
f3a0e05d91 added noavx2 vulkan 2024-02-22 16:56:25 +08:00
henk717
c9ecd0f7c6
Vulkan Koboldcpp.sh support (#640)
* Add Vulkan Loader

* Build Vulkan

* Ship Vulkan
2024-01-26 01:46:46 +08:00
Concedo
79b73e28b2 update pyinstallers 2024-01-25 23:53:46 +08:00
Concedo
07c49f8554 update linux builder 2024-01-24 20:17:06 +08:00
Concedo
14de08586e added more compile flags to set apart the conda paths, and also for colab. updated readme for multitool 2024-01-21 17:38:33 +08:00
henk717
306754791f
Koboldcpp.sh Fix & Nocuda (#562)
* .sh script V1

* koboldcpp.sh polish

* koboldcpp.sh dist generator

* Include html's in dist

* RWKV in Linux Dist

* Lower dependency requirements

* Eliminate wget dependency

* More distinct binary name

I know its technically amd64, but I don't want to cause confusion among nvidia users.

* Use System OpenCL

Unsure how this will behave in the pyinstaller build, but pocl ended up CPU only. With a bit of luck the pyinstaller uses the one from the actual system if compiled in a system without opencl, while conda now includes it for that specific system.

* Add cblas dependency

Missing this causes compile failures on some system's

* ICD workaround

Ideally we find a better solution, but conda forces ICD and needs this for the successful compile. However, pyinstaller then embeds the ICD causing it to be limited to the system it was compiled for. By temporarily removing the ICD pyinstaller can't find it and everything remains functional. Ideally we do this on a pyinstaller level, but I could not find any good options to do so yet.

* Fix & Nocuda

---------

Co-authored-by: root <root@DESKTOP-DQ1QRAG>
2023-12-14 00:06:58 +08:00
henk717
e2cf3b7aca
koboldcpp.sh - The Mamba Multitool (#554)
* .sh script V1

* koboldcpp.sh polish

* koboldcpp.sh dist generator

* Include html's in dist

* RWKV in Linux Dist

* Lower dependency requirements

* Eliminate wget dependency

* More distinct binary name

I know its technically amd64, but I don't want to cause confusion among nvidia users.

* Use System OpenCL

Unsure how this will behave in the pyinstaller build, but pocl ended up CPU only. With a bit of luck the pyinstaller uses the one from the actual system if compiled in a system without opencl, while conda now includes it for that specific system.

* Add cblas dependency

Missing this causes compile failures on some system's

* ICD workaround

Ideally we find a better solution, but conda forces ICD and needs this for the successful compile. However, pyinstaller then embeds the ICD causing it to be limited to the system it was compiled for. By temporarily removing the ICD pyinstaller can't find it and everything remains functional. Ideally we do this on a pyinstaller level, but I could not find any good options to do so yet.

---------

Co-authored-by: root <root@DESKTOP-DQ1QRAG>
2023-12-10 21:30:17 +08:00