This commit is contained in:
Nicolas Mowen 2025-11-19 18:26:25 -06:00 committed by GitHub
commit cf7442b2e2
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
4 changed files with 18 additions and 13 deletions

View File

@ -320,6 +320,12 @@ http {
add_header Cache-Control "public"; add_header Cache-Control "public";
} }
location /fonts/ {
access_log off;
expires 1y;
add_header Cache-Control "public";
}
location /locales/ { location /locales/ {
access_log off; access_log off;
add_header Cache-Control "public"; add_header Cache-Control "public";

View File

@ -70,7 +70,7 @@ You should have at least 8 GB of RAM available (or VRAM if running on GPU) to ru
genai: genai:
provider: ollama provider: ollama
base_url: http://localhost:11434 base_url: http://localhost:11434
model: llava:7b model: qwen3-vl:4b
``` ```
## Google Gemini ## Google Gemini

View File

@ -35,19 +35,18 @@ Each model is available in multiple parameter sizes (3b, 4b, 8b, etc.). Larger s
:::tip :::tip
If you are trying to use a single model for Frigate and HomeAssistant, it will need to support vision and tools calling. https://github.com/skye-harris/ollama-modelfiles contains optimized model configs for this task. If you are trying to use a single model for Frigate and HomeAssistant, it will need to support vision and tools calling. qwen3-VL supports vision and tools simultaneously in Ollama.
::: :::
The following models are recommended: The following models are recommended:
| Model | Notes | | Model | Notes |
| ----------------- | ----------------------------------------------------------- | | ----------------- | -------------------------------------------------------------------- |
| `qwen3-vl` | Strong visual and situational understanding | | `qwen3-vl` | Strong visual and situational understanding, higher vram requirement |
| `Intern3.5VL` | Relatively fast with good vision comprehension | | `Intern3.5VL` | Relatively fast with good vision comprehension |
| `gemma3` | Strong frame-to-frame understanding, slower inference times | | `gemma3` | Strong frame-to-frame understanding, slower inference times |
| `qwen2.5-vl` | Fast but capable model with good vision comprehension | | `qwen2.5-vl` | Fast but capable model with good vision comprehension |
| `llava-phi3` | Lightweight and fast model with vision comprehension |
:::note :::note

View File

@ -198,9 +198,9 @@ export default function TriggerView({
return axios return axios
.put("config/set", configBody) .put("config/set", configBody)
.then((configResponse) => { .then(async (configResponse) => {
if (configResponse.status === 200) { if (configResponse.status === 200) {
updateConfig(); await updateConfig();
const displayName = const displayName =
friendly_name && friendly_name !== "" friendly_name && friendly_name !== ""
? `${friendly_name} (${name})` ? `${friendly_name} (${name})`
@ -353,9 +353,9 @@ export default function TriggerView({
return axios return axios
.put("config/set", configBody) .put("config/set", configBody)
.then((configResponse) => { .then(async (configResponse) => {
if (configResponse.status === 200) { if (configResponse.status === 200) {
updateConfig(); await updateConfig();
const friendly = const friendly =
config?.cameras?.[selectedCamera]?.semantic_search config?.cameras?.[selectedCamera]?.semantic_search
?.triggers?.[name]?.friendly_name; ?.triggers?.[name]?.friendly_name;