Set model in llama.cpp config

This commit is contained in:
Nicolas Mowen
2026-02-20 14:13:28 -07:00
parent b6e17b032a
commit ee80475528

View File

@@ -67,6 +67,7 @@ class LlamaCppClient(GenAIClient):
# Build request payload with llama.cpp native options
payload = {
"model": self.genai_config.model,
"messages": [
{
"role": "user",
@@ -134,6 +135,7 @@ class LlamaCppClient(GenAIClient):
openai_tool_choice = "required"
payload = {
"model": self.genai_config.model,
"messages": messages,
}