Add option to set runtime options for LLM providers

This commit is contained in:
Nicolas Mowen 2026-01-12 20:22:13 -07:00
parent 4f331b4f5e
commit 5b08dcb197
6 changed files with 19 additions and 3 deletions

View File

@ -26,3 +26,6 @@ class GenAIConfig(FrigateBaseModel):
provider_options: dict[str, Any] = Field( provider_options: dict[str, Any] = Field(
default={}, title="GenAI Provider extra options." default={}, title="GenAI Provider extra options."
) )
runtime_options: dict[str, Any] = Field(
default={}, title="Options to pass during inference calls."
)

View File

@ -64,6 +64,7 @@ class OpenAIClient(GenAIClient):
}, },
], ],
timeout=self.timeout, timeout=self.timeout,
**self.genai_config.runtime_options,
) )
except Exception as e: except Exception as e:
logger.warning("Azure OpenAI returned an error: %s", str(e)) logger.warning("Azure OpenAI returned an error: %s", str(e))

View File

@ -35,10 +35,14 @@ class GeminiClient(GenAIClient):
for img in images for img in images
] + [prompt] ] + [prompt]
try: try:
# Merge runtime_options into generation_config if provided
generation_config_dict = {"candidate_count": 1}
generation_config_dict.update(self.genai_config.runtime_options)
response = self.provider.generate_content( response = self.provider.generate_content(
data, data,
generation_config=genai.types.GenerationConfig( generation_config=genai.types.GenerationConfig(
candidate_count=1, **generation_config_dict
), ),
request_options=genai.types.RequestOptions( request_options=genai.types.RequestOptions(
timeout=self.timeout, timeout=self.timeout,

View File

@ -58,11 +58,15 @@ class OllamaClient(GenAIClient):
) )
return None return None
try: try:
ollama_options = {
**self.provider_options,
**self.genai_config.runtime_options,
}
result = self.provider.generate( result = self.provider.generate(
self.genai_config.model, self.genai_config.model,
prompt, prompt,
images=images if images else None, images=images if images else None,
**self.provider_options, **ollama_options,
) )
logger.debug( logger.debug(
f"Ollama tokens used: eval_count={result.get('eval_count')}, prompt_eval_count={result.get('prompt_eval_count')}" f"Ollama tokens used: eval_count={result.get('eval_count')}, prompt_eval_count={result.get('prompt_eval_count')}"

View File

@ -61,6 +61,7 @@ class OpenAIClient(GenAIClient):
}, },
], ],
timeout=self.timeout, timeout=self.timeout,
**self.genai_config.runtime_options,
) )
if ( if (
result is not None result is not None

View File

@ -15,6 +15,9 @@
}, },
"provider_options": { "provider_options": {
"label": "GenAI Provider extra options." "label": "GenAI Provider extra options."
},
"runtime_options": {
"label": "Options to pass during inference calls."
} }
} }
} }