mirror of
https://github.com/LostRuins/koboldcpp.git
synced 2026-05-08 01:41:37 +00:00
* gguf-py: Refactor and add file reading support * Replay changes from #3871 Credit to @cebtenzzre for that pull * Various type annotation fixes. * sort imports with isort (again) * Fix missing return statement in add_tensor * style cleanup with flake8 * fix NamedTuple and Enum usage * Fix an issue with state init in GGUFReader Move examples to an examples/ directory Clean up examples Add an example of modifying keys in a GGUF file Update documentation with info on examples Try to support people importing gguf/gguf.py directly * Damagage is not a word. * Clean up gguf-py/examples/modify_gguf.py whitespace Co-authored-by: Jared Van Bortel <cebtenzzre@gmail.com> * Update gguf-py/examples/modify_gguf.py formatting Co-authored-by: Jared Van Bortel <cebtenzzre@gmail.com> * Update gguf-py/gguf/gguf_reader.py type hint Co-authored-by: Jared Van Bortel <cebtenzzre@gmail.com> * Make examples executable, formatting changes * Add more information to GGUFReader and examples comments * Include a gguf Python package version bump * Add convert-gguf-endian.py script * cleanup * gguf-py : bump minor version * Reorganize scripts * Make GGUFReader endian detection less arbitrary * Add JSON dumping support to gguf-dump.py Which I kind of regret now * A few for gguf-dump.py cleanups * Murder accidental tuple in gguf-py/scripts/gguf-dump.py Co-authored-by: Jared Van Bortel <cebtenzzre@gmail.com> * cleanup * constants : remove unneeded type annotations * fix python 3.8 compat * Set up gguf- scripts in pyproject.toml * And include scripts/__init__.py, derp * convert.py: We can't currently support Q8_0 on big endian. * gguf-py: SpecialVocab: Always try available sources for special token ids gguf-py: SpecialVocab: Try to load merges from merges.txt if not in tokenizer.json gguf-py: SpecialVocab: Add 'add_bos_token' type bools to GGUF metadata u * cleanup * Promote add_X_token to GGUF metadata for BOS and EOS --------- Co-authored-by: Jared Van Bortel <jared@nomic.ai> Co-authored-by: Jared Van Bortel <cebtenzzre@gmail.com> |
||
|---|---|---|
| .. | ||
| baby-llama | ||
| batched | ||
| batched-bench | ||
| batched.swift | ||
| beam-search | ||
| benchmark | ||
| convert-llama2c-to-ggml | ||
| embedding | ||
| export-lora | ||
| finetune | ||
| gguf | ||
| infill | ||
| jeopardy | ||
| llama-bench | ||
| llava | ||
| main | ||
| main-cmake-pkg | ||
| metal | ||
| parallel | ||
| perplexity | ||
| quantize | ||
| quantize-stats | ||
| save-load-state | ||
| server | ||
| simple | ||
| speculative | ||
| train-text-from-scratch | ||
| alpaca.sh | ||
| chat-13B.bat | ||
| chat-13B.sh | ||
| chat-persistent.sh | ||
| chat-vicuna.sh | ||
| chat.sh | ||
| CMakeLists.txt | ||
| gpt4all.sh | ||
| json-schema-to-grammar.py | ||
| llama.vim | ||
| llama2-13b.sh | ||
| llama2.sh | ||
| llm.vim | ||
| make-ggml.py | ||
| Miku.sh | ||
| reason-act.sh | ||
| server-llama2-13B.sh | ||