Commit graph

96 commits

Author SHA1 Message Date
Concedo
f46d42343b swap colab default model 2026-04-21 22:31:55 +08:00
Concedo
96ec87127a updated colab, handle connection dropping during prompt processing 2026-04-21 21:46:13 +08:00
Concedo
2962e5bac4 updated colab image models 2026-04-18 18:02:17 +08:00
Concedo
78589974de updated colab 2026-04-18 16:41:27 +08:00
Concedo
b81103d6ba clean up colab a bit 2026-03-25 22:14:38 +08:00
Concedo
efc1db9ec8 add mirror for colab 2026-03-22 17:43:41 +08:00
Concedo
9d4653bcb9 colab: clip and vae to gpu (+1 squashed commits)
Squashed commits:

[d5de2f86d] colab: clip and vae to gpu
2026-03-22 01:10:55 +08:00
Concedo
79e39e1989 fixed a help menu bug, updated colab (+1 squashed commits)
Squashed commits:

[618478e00] fixed a help menu bug, updated colab
2026-03-22 01:00:30 +08:00
Concedo
9ba8c7a661 fixed colab 2026-03-21 10:21:18 +08:00
Concedo
2d349723d3 fixed colab 2026-03-20 18:19:59 +08:00
Concedo
d6aae073b6 fixed typo 2026-03-20 12:02:04 +08:00
Concedo
1f73eabb46 fixed colab 2026-03-20 11:45:10 +08:00
Concedo
3b9385a627 updated colab, wip model router 2026-03-15 00:38:29 +08:00
Concedo
d100c8660e added Tlacuilo 2026-02-23 10:48:56 +08:00
Concedo
fcd86bb570 lower mistral colab defaults 2026-02-12 23:17:06 +08:00
Concedo
df30473716 deduplicate repeated statements in colab, minor refactgor 2025-11-25 16:11:18 +08:00
Concedo
bd0d6c2da5 add estopia to colab list, pending minor refactor 2025-11-25 00:18:11 +08:00
henk717
e2232d9ad9
Make colab more user friendly (#1857) 2025-11-24 22:31:38 +08:00
Concedo
0e2b031159 colab cpus are too slow to run kokoro. swap back to outetts
Some checks failed
Copilot Setup Steps / copilot-setup-steps (push) Has been cancelled
2025-08-24 16:20:01 +08:00
Concedo
774a399068 updated colab 2025-08-24 16:07:58 +08:00
Concedo
e7eb6d3200 increase default ctx size to 8k, rename usecublas to usecuda 2025-07-13 18:27:42 +08:00
Concedo
dcf88d6e78 Revert "make tts use gpu by default. use --ttscpu to disable"
This reverts commit 669f80265b.
2025-06-08 17:08:04 +08:00
Concedo
669f80265b make tts use gpu by default. use --ttscpu to disable 2025-06-08 17:06:19 +08:00
Concedo
a80dfa5c10 various minor fixes 2025-06-08 01:11:42 +08:00
Concedo
2142d6ba68 Updated Colab to use internal downloader
fixed model command (+1 squashed commits)

Squashed commits:

[a4d8fd9f1] tryout new colab (+1 squashed commits)

Squashed commits:

[c97333d44] tryout new colab
2025-06-01 11:41:45 +08:00
Concedo
59c02aa1a6 embeddings model colab 2025-04-05 10:30:47 +08:00
Concedo
75e7902789 add localtunnel fallback (+1 squashed commits)
Squashed commits:

[ff0a63f6] add localtunnel fallback
2025-03-26 17:35:59 +08:00
Concedo
dbd8c680ba allow remote saving to google drive 2025-03-09 15:04:43 +08:00
Concedo
6b7d2349a7 Rewrite history to fix bad vulkan shader commits without increasing repo size
added dpe colab (+8 squashed commit)

Squashed commit:

[b8362da4] updated lite

[ed6c037d] move nsigma into the regular sampler stack

[ac5f61c6] relative filepath fixed

[05fe96ab] export template

[ed0a5a3e] nix_example.md: refactor (#1401)

* nix_example.md: add override example

* nix_example.md: drop graphics example, already basic nixos knowledge

* nix_example.md: format

* nix_example.md: Vulkan is disabled on macOS

Disabled in: 1ccd253acc

* nix_examples.md: nixpkgs.config.cuda{Arches -> Capabilities}

Fixes: https://github.com/LostRuins/koboldcpp/issues/1367

[675c62f7] AutoGuess: Phi 4 (mini) (#1402)

[4bf56982] phrasing

[b8c0df04] Add Rep Pen to Top N Sigma sampler chain (#1397)

- place after nsigma and before xtc (+3 squashed commit)

Squashed commit:

[87c52b97] disable VMM from HIP

[ee8906f3] edit description

[e85c0e69] Remove Unnecessary Rep Counting (#1394)

* stop counting reps

* fix range-based initializer

* strike that - reverse it
2025-03-05 00:02:20 +08:00
Concedo
5ee7cbe08c add cydonia to colab 2025-02-22 23:02:44 +08:00
Concedo
03def285db updated colab 2025-01-23 00:13:55 +08:00
Concedo
4d92b4e98e updated readme and colab 2025-01-14 00:31:52 +08:00
Concedo
dcfa1eca4e Merge commit '017cc5f446' into concedo_experimental
# Conflicts:
#	.github/ISSUE_TEMPLATE/010-bug-compilation.yml
#	.github/ISSUE_TEMPLATE/019-bug-misc.yml
#	CODEOWNERS
#	examples/batched-bench/batched-bench.cpp
#	examples/batched/batched.cpp
#	examples/convert-llama2c-to-ggml/convert-llama2c-to-ggml.cpp
#	examples/gritlm/gritlm.cpp
#	examples/llama-bench/llama-bench.cpp
#	examples/passkey/passkey.cpp
#	examples/quantize-stats/quantize-stats.cpp
#	examples/run/run.cpp
#	examples/simple-chat/simple-chat.cpp
#	examples/simple/simple.cpp
#	examples/tokenize/tokenize.cpp
#	ggml/CMakeLists.txt
#	ggml/src/ggml-metal/CMakeLists.txt
#	ggml/src/ggml-vulkan/CMakeLists.txt
#	scripts/sync-ggml.last
#	src/llama.cpp
#	tests/test-autorelease.cpp
#	tests/test-model-load-cancel.cpp
#	tests/test-tokenizer-0.cpp
#	tests/test-tokenizer-1-bpe.cpp
#	tests/test-tokenizer-1-spm.cpp
2025-01-08 23:15:21 +08:00
Concedo
1012281320 updated colab 2025-01-03 18:02:02 +08:00
Concedo
df7c2b9923 renamed some labels 2024-11-11 19:40:47 +08:00
Concedo
90f5cd0f67 wip logprobs data 2024-10-30 00:59:34 +08:00
Concedo
efc6939294 flashattn default true on colab 2024-10-14 18:50:02 +08:00
Concedo
1803382415 updated colab 2024-10-06 21:30:58 +08:00
Concedo
1df850c95c add magnum to colab models 2024-07-30 21:13:29 +08:00
Concedo
a441c27cb5 fixed broken link 2024-07-16 01:00:16 +08:00
Concedo
066e7ac540 minor fixes: colab gpu backend, lite bugs, package python file with embd 2024-07-15 17:36:03 +08:00
Concedo
5b605d03ea Merge branch 'upstream' into concedo_experimental
# Conflicts:
#	.github/ISSUE_TEMPLATE/config.yml
#	.gitignore
#	CMakeLists.txt
#	CONTRIBUTING.md
#	Makefile
#	README.md
#	ci/run.sh
#	common/common.h
#	examples/main-cmake-pkg/CMakeLists.txt
#	ggml/src/CMakeLists.txt
#	models/ggml-vocab-bert-bge.gguf.inp
#	models/ggml-vocab-bert-bge.gguf.out
#	models/ggml-vocab-deepseek-coder.gguf.inp
#	models/ggml-vocab-deepseek-coder.gguf.out
#	models/ggml-vocab-deepseek-llm.gguf.inp
#	models/ggml-vocab-deepseek-llm.gguf.out
#	models/ggml-vocab-falcon.gguf.inp
#	models/ggml-vocab-falcon.gguf.out
#	models/ggml-vocab-gpt-2.gguf.inp
#	models/ggml-vocab-gpt-2.gguf.out
#	models/ggml-vocab-llama-bpe.gguf.inp
#	models/ggml-vocab-llama-bpe.gguf.out
#	models/ggml-vocab-llama-spm.gguf.inp
#	models/ggml-vocab-llama-spm.gguf.out
#	models/ggml-vocab-mpt.gguf.inp
#	models/ggml-vocab-mpt.gguf.out
#	models/ggml-vocab-phi-3.gguf.inp
#	models/ggml-vocab-phi-3.gguf.out
#	models/ggml-vocab-starcoder.gguf.inp
#	models/ggml-vocab-starcoder.gguf.out
#	requirements.txt
#	requirements/requirements-convert_legacy_llama.txt
#	scripts/check-requirements.sh
#	scripts/pod-llama.sh
#	src/CMakeLists.txt
#	src/llama.cpp
#	tests/test-rope.cpp
2024-07-06 00:25:10 +08:00
Concedo
6b0756506b improvements to model downloader and chat completions adapter loader 2024-07-04 15:34:08 +08:00
Concedo
4f369b0a0a update colab 2024-06-27 15:41:06 +08:00
Concedo
967b6572a2 try to use GPU for whisper 2024-06-03 23:07:26 +08:00
Concedo
5ebc532ca9 update colab 2024-06-03 14:55:12 +08:00
Concedo
868446bd1a replace sdconfig and hordeconfig 2024-05-09 22:43:50 +08:00
Concedo
640f195140 add kobble tiny to readme 2024-05-03 18:13:39 +08:00
Concedo
69dcffa4ec updated lite and colab 2024-04-21 16:48:48 +08:00
Concedo
d54af7fa31 updated swagger json link fix 2024-04-09 14:55:27 +08:00