mirror of
https://github.com/LostRuins/koboldcpp.git
synced 2026-04-28 03:30:20 +00:00
CUDA: also store `node->src->data` ptrs for equality check (#21635)
python-check-requirements.yml #283 -Commit
d12cc3d1ca
pushed by
vrr
CUDA: also store `node->src->data` ptrs for equality check (#21635)
pre-tokenizer-hashes.yml #282 -Commit
d12cc3d1ca
pushed by
vrr
docs: fix typo in build.md (emdawbwebgpu -> emdawnwebgpu) (#21518)
update-ops-docs.yml #281 -Commit
0033f53a07
pushed by
vrr
docs: fix typo in build.md (emdawbwebgpu -> emdawnwebgpu) (#21518)
python-type-check.yml #280 -Commit
0033f53a07
pushed by
vrr
docs: fix typo in build.md (emdawbwebgpu -> emdawnwebgpu) (#21518)
python-check-requirements.yml #279 -Commit
0033f53a07
pushed by
vrr
docs: fix typo in build.md (emdawbwebgpu -> emdawnwebgpu) (#21518)
pre-tokenizer-hashes.yml #278 -Commit
0033f53a07
pushed by
vrr
convert : set "add bos" == True for Gemma 4 (#21500)
python-type-check.yml #277 -Commit
400ac8e194
pushed by
vrr
convert : set "add bos" == True for Gemma 4 (#21500)
python-check-requirements.yml #276 -Commit
400ac8e194
pushed by
vrr
convert : set "add bos" == True for Gemma 4 (#21500)
pre-tokenizer-hashes.yml #275 -Commit
400ac8e194
pushed by
vrr
ggml-webgpu: move from parameter buffer pool to single buffer with offsets (#21278)
python-type-check.yml #274 -Commit
d006858316
pushed by
vrr
docker : bump cuda12 to 12.9.1 (#20920)
update-ops-docs.yml #273 -Commit
277ff5fff7
pushed by
vrr
docker : bump cuda12 to 12.9.1 (#20920)
python-type-check.yml #272 -Commit
277ff5fff7
pushed by
vrr
docker : bump cuda12 to 12.9.1 (#20920)
python-check-requirements.yml #271 -Commit
277ff5fff7
pushed by
vrr
docker : bump cuda12 to 12.9.1 (#20920)
pre-tokenizer-hashes.yml #270 -Commit
277ff5fff7
pushed by
vrr
tests: allow exporting graph ops from HF file without downloading weights (#21182)
python-type-check.yml #269 -Commit
5803c8d115
pushed by
vrr
tests: allow exporting graph ops from HF file without downloading weights (#21182)
python-check-requirements.yml #268 -Commit
5803c8d115
pushed by
vrr
tests: allow exporting graph ops from HF file without downloading weights (#21182)
pre-tokenizer-hashes.yml #267 -Commit
5803c8d115
pushed by
vrr
ggml-webgpu: port all AOT operators to JIT (#20728)
python-type-check.yml #266 -Commit
825eb91a66
pushed by
vrr
webui: Fix branching logic on edit message (#21175)
python-type-check.yml #265 -Commit
389c7d4955
pushed by
vrr
webui: Fix branching logic on edit message (#21175)
python-check-requirements.yml #264 -Commit
389c7d4955
pushed by
vrr
webui: Fix branching logic on edit message (#21175)
pre-tokenizer-hashes.yml #263 -Commit
389c7d4955
pushed by
vrr
common/json-schema: fix: handle non-capturing groups (?:...) in JSON schema pattern converter (#21124)
python-type-check.yml #262 -Commit
e397d3885c
pushed by
vrr
common/json-schema: fix: handle non-capturing groups (?:...) in JSON schema pattern converter (#21124)
python-check-requirements.yml #261 -Commit
e397d3885c
pushed by
vrr
common/json-schema: fix: handle non-capturing groups (?:...) in JSON schema pattern converter (#21124)
pre-tokenizer-hashes.yml #260 -Commit
e397d3885c
pushed by
vrr
llama-bench: print `-n-cpu-moe` when offloaded layers > 1 (#20984)
python-type-check.yml #259 -Commit
9c600bcd4b
pushed by
vrr
llama-bench: print `-n-cpu-moe` when offloaded layers > 1 (#20984)
python-check-requirements.yml #258 -Commit
9c600bcd4b
pushed by
vrr
llama-bench: print `-n-cpu-moe` when offloaded layers > 1 (#20984)
pre-tokenizer-hashes.yml #257 -Commit
9c600bcd4b
pushed by
vrr
llama-bench: print `-n-cpu-moe` when offloaded layers > 1 (#20984)
copilot-setup-steps.yml #256 -Commit
9c600bcd4b
pushed by
vrr
misc : prefer ggml-org models in docs and examples (#20827)
python-type-check.yml #255 -Commit
3306dbaef7
pushed by
vrr
misc : prefer ggml-org models in docs and examples (#20827)
python-check-requirements.yml #254 -Commit
3306dbaef7
pushed by
vrr