mirror of
https://github.com/LostRuins/koboldcpp.git
synced 2025-09-11 09:34:37 +00:00
Merge branch 'upstream' into concedo_experimental
# Conflicts: # .github/workflows/build.yml # .github/workflows/close-issue.yml # .github/workflows/nix-ci-aarch64.yml # .github/workflows/nix-ci.yml # README.md # ci/run.sh # examples/server/README.md # ggml/src/ggml-cuda.cu # ggml/src/ggml-metal.m # scripts/sync-ggml.last # tests/test-backend-ops.cpp
This commit is contained in:
commit
da6cf261a8
28 changed files with 725 additions and 421 deletions
|
@ -2028,7 +2028,7 @@ struct server_context {
|
|||
continue;
|
||||
}
|
||||
|
||||
// prompt: <s>query</s><s>doc</s>
|
||||
// prompt: [BOS]query[EOS][SEP]doc[EOS]
|
||||
prompt_tokens.clear();
|
||||
prompt_tokens.push_back(llama_token_bos(model));
|
||||
{
|
||||
|
@ -2036,7 +2036,7 @@ struct server_context {
|
|||
prompt_tokens.insert(prompt_tokens.end(), part.begin(), part.end());
|
||||
}
|
||||
prompt_tokens.push_back(llama_token_eos(model));
|
||||
prompt_tokens.push_back(llama_token_bos(model));
|
||||
prompt_tokens.push_back(llama_token_sep(model));
|
||||
{
|
||||
const auto part = tokenize(slot.prompt[1], false);
|
||||
prompt_tokens.insert(prompt_tokens.end(), part.begin(), part.end());
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue