From 471bef3984088c5ec8921490f4bbf0db905c5f76 Mon Sep 17 00:00:00 2001 From: Nicolas Mowen Date: Thu, 26 Feb 2026 15:28:58 -0700 Subject: [PATCH] Set model --- frigate/genai/llama_cpp.py | 12 +++++++----- 1 file changed, 7 insertions(+), 5 deletions(-) diff --git a/frigate/genai/llama_cpp.py b/frigate/genai/llama_cpp.py index fc3569379..dd6dc6f71 100644 --- a/frigate/genai/llama_cpp.py +++ b/frigate/genai/llama_cpp.py @@ -226,15 +226,17 @@ class LlamaCppClient(GenAIClient): to_encode = jpeg_bytes if jpeg_bytes is not None else img encoded = base64.b64encode(to_encode).decode("utf-8") # prompt_string must contain <__media__> placeholder for image tokenization - content.append({ - "prompt_string": "<__media__>\n", - "multimodal_data": [encoded], - }) + content.append( + { + "prompt_string": "<__media__>\n", + "multimodal_data": [encoded], + } + ) try: response = requests.post( f"{self.provider}/embeddings", - json={"content": content}, + json={"model": self.genai_config.model, "content": content}, timeout=self.timeout, ) response.raise_for_status()