mirror of
https://github.com/LostRuins/koboldcpp.git
synced 2025-09-10 17:14:36 +00:00
# Conflicts: # common/CMakeLists.txt # common/arg.cpp # common/chat.cpp # examples/parallel/README.md # examples/parallel/parallel.cpp # ggml/cmake/common.cmake # ggml/src/CMakeLists.txt # ggml/src/ggml-cpu/CMakeLists.txt # ggml/src/ggml-sycl/ggml-sycl.cpp # ggml/src/ggml-sycl/rope.cpp # models/ggml-vocab-bert-bge.gguf.inp # models/ggml-vocab-bert-bge.gguf.out # models/ggml-vocab-command-r.gguf.inp # models/ggml-vocab-command-r.gguf.out # models/ggml-vocab-deepseek-coder.gguf.inp # models/ggml-vocab-deepseek-coder.gguf.out # models/ggml-vocab-deepseek-llm.gguf.inp # models/ggml-vocab-deepseek-llm.gguf.out # models/ggml-vocab-falcon.gguf.inp # models/ggml-vocab-falcon.gguf.out # models/ggml-vocab-gpt-2.gguf.inp # models/ggml-vocab-gpt-2.gguf.out # models/ggml-vocab-llama-bpe.gguf.inp # models/ggml-vocab-llama-bpe.gguf.out # models/ggml-vocab-llama-spm.gguf.inp # models/ggml-vocab-llama-spm.gguf.out # models/ggml-vocab-mpt.gguf.inp # models/ggml-vocab-mpt.gguf.out # models/ggml-vocab-phi-3.gguf.inp # models/ggml-vocab-phi-3.gguf.out # models/ggml-vocab-qwen2.gguf.inp # models/ggml-vocab-qwen2.gguf.out # models/ggml-vocab-refact.gguf.inp # models/ggml-vocab-refact.gguf.out # models/ggml-vocab-starcoder.gguf.inp # models/ggml-vocab-starcoder.gguf.out # requirements/requirements-gguf_editor_gui.txt # tests/CMakeLists.txt # tests/test-chat.cpp # tests/test-grammar-integration.cpp # tests/test-json-schema-to-grammar.cpp # tools/mtmd/CMakeLists.txt # tools/run/run.cpp # tools/server/CMakeLists.txt |
||
---|---|---|
.. | ||
llama-adapter.cpp | ||
llama-adapter.h | ||
llama-arch.cpp | ||
llama-arch.h | ||
llama-batch.cpp | ||
llama-batch.h | ||
llama-chat.cpp | ||
llama-chat.h | ||
llama-context.cpp | ||
llama-context.h | ||
llama-cparams.cpp | ||
llama-cparams.h | ||
llama-grammar.cpp | ||
llama-grammar.h | ||
llama-graph.cpp | ||
llama-graph.h | ||
llama-hparams.cpp | ||
llama-hparams.h | ||
llama-impl.cpp | ||
llama-impl.h | ||
llama-io.cpp | ||
llama-io.h | ||
llama-kv-cache.cpp | ||
llama-kv-cache.h | ||
llama-kv-cells.h | ||
llama-memory.cpp | ||
llama-memory.h | ||
llama-mmap.cpp | ||
llama-mmap.h | ||
llama-model-loader.cpp | ||
llama-model-loader.h | ||
llama-model-saver.cpp | ||
llama-model-saver.h | ||
llama-model.cpp | ||
llama-model.h | ||
llama-quant.cpp | ||
llama-quant.h | ||
llama-sampling.cpp | ||
llama-sampling.h | ||
llama-vocab.cpp | ||
llama-vocab.h | ||
llama.cpp | ||
unicode-data.cpp | ||
unicode-data.h | ||
unicode.cpp | ||
unicode.h |