mirror of
https://github.com/carlrobertoh/ProxyAI.git
synced 2026-05-11 21:31:04 +00:00
fix: use /infill for llama.cpp code-completions (#513)
This commit is contained in:
parent
7d05d17797
commit
8de72b3301
3 changed files with 8 additions and 9 deletions
|
|
@ -122,7 +122,7 @@ public final class CompletionRequestService {
|
|||
CodeCompletionRequestFactory.buildCustomRequest(requestDetails),
|
||||
new OpenAITextCompletionEventSourceListener(eventListener));
|
||||
case LLAMA_CPP -> CompletionClientProvider.getLlamaClient()
|
||||
.getChatCompletionAsync(
|
||||
.getInfillAsync(
|
||||
CodeCompletionRequestFactory.buildLlamaRequest(requestDetails),
|
||||
eventListener);
|
||||
default ->
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue