mirror of
https://github.com/LostRuins/koboldcpp.git
synced 2026-05-06 08:01:27 +00:00
completion : fix prompt cache for recurrent models (#19045)
This commit is contained in:
parent
1243f93a2d
commit
080b161995
2 changed files with 48 additions and 41 deletions
|
|
@ -2559,6 +2559,7 @@ size_t llama_context::state_write_data(llama_io_write_i & io) {
|
|||
}
|
||||
}
|
||||
|
||||
// [TAG_CONTEXT_STATE_LOGITS]
|
||||
// write logits
|
||||
{
|
||||
LLAMA_LOG_DEBUG("%s: - writing logits\n", __func__);
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue