kvcache-ai-ktransformers/ktransformers/server/backend/interfaces
2025-02-18 11:15:17 +08:00
..
__init__.py Initial commit 2024-07-27 16:06:58 +08:00
exllamav2.py Initial commit 2024-07-27 16:06:58 +08:00
ktransformers.py fix: use 'cuda:0' by default if torch_device is 'cuda' 2025-02-18 11:15:17 +08:00
transformers.py fix: server: drop <think> tag in chat template 2025-02-17 14:25:27 +08:00