Set model in llama.cpp config

This commit is contained in:
Nicolas Mowen 2026-02-20 14:13:28 -07:00
parent 657500143b
commit 51a26d3365

View File

@ -67,6 +67,7 @@ class LlamaCppClient(GenAIClient):
# Build request payload with llama.cpp native options
payload = {
"model": self.genai_config.model,
"messages": [
{
"role": "user",
@ -134,6 +135,7 @@ class LlamaCppClient(GenAIClient):
openai_tool_choice = "required"
payload = {
"model": self.genai_config.model,
"messages": messages,
}