mirror of
https://github.com/blakeblackshear/frigate.git
synced 2026-01-22 20:18:30 +03:00
Add option to set runtime options for LLM providers
This commit is contained in:
parent
4f331b4f5e
commit
5b08dcb197
@ -26,3 +26,6 @@ class GenAIConfig(FrigateBaseModel):
|
|||||||
provider_options: dict[str, Any] = Field(
|
provider_options: dict[str, Any] = Field(
|
||||||
default={}, title="GenAI Provider extra options."
|
default={}, title="GenAI Provider extra options."
|
||||||
)
|
)
|
||||||
|
runtime_options: dict[str, Any] = Field(
|
||||||
|
default={}, title="Options to pass during inference calls."
|
||||||
|
)
|
||||||
|
|||||||
@ -64,6 +64,7 @@ class OpenAIClient(GenAIClient):
|
|||||||
},
|
},
|
||||||
],
|
],
|
||||||
timeout=self.timeout,
|
timeout=self.timeout,
|
||||||
|
**self.genai_config.runtime_options,
|
||||||
)
|
)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warning("Azure OpenAI returned an error: %s", str(e))
|
logger.warning("Azure OpenAI returned an error: %s", str(e))
|
||||||
|
|||||||
@ -35,10 +35,14 @@ class GeminiClient(GenAIClient):
|
|||||||
for img in images
|
for img in images
|
||||||
] + [prompt]
|
] + [prompt]
|
||||||
try:
|
try:
|
||||||
|
# Merge runtime_options into generation_config if provided
|
||||||
|
generation_config_dict = {"candidate_count": 1}
|
||||||
|
generation_config_dict.update(self.genai_config.runtime_options)
|
||||||
|
|
||||||
response = self.provider.generate_content(
|
response = self.provider.generate_content(
|
||||||
data,
|
data,
|
||||||
generation_config=genai.types.GenerationConfig(
|
generation_config=genai.types.GenerationConfig(
|
||||||
candidate_count=1,
|
**generation_config_dict
|
||||||
),
|
),
|
||||||
request_options=genai.types.RequestOptions(
|
request_options=genai.types.RequestOptions(
|
||||||
timeout=self.timeout,
|
timeout=self.timeout,
|
||||||
|
|||||||
@ -58,11 +58,15 @@ class OllamaClient(GenAIClient):
|
|||||||
)
|
)
|
||||||
return None
|
return None
|
||||||
try:
|
try:
|
||||||
|
ollama_options = {
|
||||||
|
**self.provider_options,
|
||||||
|
**self.genai_config.runtime_options,
|
||||||
|
}
|
||||||
result = self.provider.generate(
|
result = self.provider.generate(
|
||||||
self.genai_config.model,
|
self.genai_config.model,
|
||||||
prompt,
|
prompt,
|
||||||
images=images if images else None,
|
images=images if images else None,
|
||||||
**self.provider_options,
|
**ollama_options,
|
||||||
)
|
)
|
||||||
logger.debug(
|
logger.debug(
|
||||||
f"Ollama tokens used: eval_count={result.get('eval_count')}, prompt_eval_count={result.get('prompt_eval_count')}"
|
f"Ollama tokens used: eval_count={result.get('eval_count')}, prompt_eval_count={result.get('prompt_eval_count')}"
|
||||||
|
|||||||
@ -61,6 +61,7 @@ class OpenAIClient(GenAIClient):
|
|||||||
},
|
},
|
||||||
],
|
],
|
||||||
timeout=self.timeout,
|
timeout=self.timeout,
|
||||||
|
**self.genai_config.runtime_options,
|
||||||
)
|
)
|
||||||
if (
|
if (
|
||||||
result is not None
|
result is not None
|
||||||
|
|||||||
@ -15,6 +15,9 @@
|
|||||||
},
|
},
|
||||||
"provider_options": {
|
"provider_options": {
|
||||||
"label": "GenAI Provider extra options."
|
"label": "GenAI Provider extra options."
|
||||||
|
},
|
||||||
|
"runtime_options": {
|
||||||
|
"label": "Options to pass during inference calls."
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user