mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2025-09-09 22:05:30 +00:00
Merge branch 'main' into feat-chunk-prefill-flashinfer
This commit is contained in:
commit
fa03ea48dd
3 changed files with 10 additions and 146 deletions
|
@ -328,6 +328,12 @@ class TransformersInterface(BackendInterfaceBase):
|
|||
|
||||
@torch.no_grad
|
||||
def generate(self):
|
||||
self.args.max_new_tokens = min(self.args.max_new_tokens, self.args.cache_lens - self.seq_length)
|
||||
if(self.args.max_new_tokens <= 0):
|
||||
logger.warning("max_new_tokens is less than 0")
|
||||
yield self.streamer.end()
|
||||
return
|
||||
logger.info(f"max_new_tokens: {self.args.max_new_tokens}")
|
||||
self.profiler.set_counter("decode", 0)
|
||||
for i in range(1, self.args.max_new_tokens):
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue