Files
continue/core/llm/llms/LlamaCpp.ts
Tomasz Stefaniak 7715d5161e Merge pull request #5726 from ipaddicting/fix-5530
fix: reverted the URL of llama.cpp back to 'completion'.
2025-05-30 18:57:54 -07:00

58 lines
1.4 KiB
TypeScript

import { streamSse } from "@continuedev/fetch";
import { CompletionOptions, LLMOptions } from "../../index.js";
import { BaseLLM } from "../index.js";
class LlamaCpp extends BaseLLM {
static providerName = "llama.cpp";
static defaultOptions: Partial<LLMOptions> = {
apiBase: "http://127.0.0.1:8080/",
};
private _convertArgs(options: CompletionOptions, prompt: string) {
const finalOptions = {
n_predict: options.maxTokens,
frequency_penalty: options.frequencyPenalty,
presence_penalty: options.presencePenalty,
min_p: options.minP,
mirostat: options.mirostat,
stop: options.stop,
top_k: options.topK,
top_p: options.topP,
temperature: options.temperature,
};
return finalOptions;
}
protected async *_streamComplete(
prompt: string,
signal: AbortSignal,
options: CompletionOptions,
): AsyncGenerator<string> {
const headers = {
"Content-Type": "application/json",
Authorization: `Bearer ${this.apiKey}`,
...this.requestOptions?.headers,
};
const resp = await this.fetch(new URL("completion", this.apiBase), {
method: "POST",
headers,
body: JSON.stringify({
prompt,
stream: true,
...this._convertArgs(options, prompt),
}),
signal,
});
for await (const value of streamSse(resp)) {
if (value.content) {
yield value.content;
}
}
}
}
export default LlamaCpp;