mirror of
https://github.com/blakeblackshear/frigate.git
synced 2026-03-09 18:19:20 +03:00
Full UI configuration (#22151)
* use react-jsonschema-form for UI config * don't use properties wrapper when generating config i18n json * configure for full i18n support * section fields * add descriptions to all fields for i18n * motion i18n * fix nullable fields * sanitize internal fields * add switches widgets and use friendly names * fix nullable schema entries * ensure update_topic is added to api calls this needs further backend implementation to work correctly * add global sections, camera config overrides, and reset button * i18n * add reset logic to global config view * tweaks * fix sections and live validation * fix validation for schema objects that can be null * generic and custom per-field validation * improve generic error validation messages * remove show advanced fields switch * tweaks * use shadcn theme * fix array field template * i18n tweaks * remove collapsible around root section * deep merge schema for advanced fields * add array field item template and fix ffmpeg section * add missing i18n keys * tweaks * comment out api call for testing * add config groups as a separate i18n namespace * add descriptions to all pydantic fields * make titles more concise * new titles as i18n * update i18n config generation script to use json schema * tweaks * tweaks * rebase * clean up * form tweaks * add wildcards and fix object filter fields * add field template for additionalproperties schema objects * improve typing * add section description from schema and clarify global vs camera level descriptions * separate and consolidate global and camera i18n namespaces * clean up now obsolete namespaces * tweaks * refactor sections and overrides * add ability to render components before and after fields * fix titles * chore(sections): remove legacy single-section components replaced by template * refactor configs to use individual files with a template * fix review description * apply hidden fields after ui schema * move util * remove unused i18n * clean up error messages * fix fast refresh * add custom validation and use it for ffmpeg input roles * update nav tree * remove unused * re-add override and modified indicators * mark pending changes and add confirmation dialog for resets * fix red unsaved dot * tweaks * add docs links, readonly keys, and restart required per field * add special case and comments for global motion section * add section form special cases * combine review sections * tweaks * add audio labels endpoint * add audio label switches and input to filter list * fix type * remove key from config when resetting to default/global * don't show description for new key/val fields * tweaks * spacing tweaks * add activity indicator and scrollbar tweaks * add docs to filter fields * wording changes * fix global ffmpeg section * add review classification zones to review form * add backend endpoint and frontend widget for ffmpeg presets and manual args * improve wording * hide descriptions for additional properties arrays * add warning log about incorrectly nested model config * spacing and language tweaks * fix i18n keys * networking section docs and description * small wording tweaks * add layout grid field * refactor with shared utilities * field order * add individual detectors to schema add detector titles and descriptions (docstrings in pydantic are used for descriptions) and add i18n keys to globals * clean up detectors section and i18n * don't save model config back to yaml when saving detectors * add full detectors config to api model dump works around the way we use detector plugins so we can have the full detector config for the frontend * add restart button to toast when restart is required * add ui option to remove inner cards * fix buttons * section tweaks * don't zoom into text on mobile * make buttons sticky at bottom of sections * small tweaks * highlight label of changed fields * add null to enum list when unwrapping * refactor to shared utils and add save all button * add undo all button * add RJSF to dictionary * consolidate utils * preserve form data when changing cameras * add mono fonts * add popover to show what fields will be saved * fix mobile menu not re-rendering with unsaved dots * tweaks * fix logger and env vars config section saving use escaped periods in keys to retain them in the config file (eg "frigate.embeddings") * add timezone widget * role map field with validation * fix validation for model section * add another hidden field * add footer message for required restart * use rjsf for notifications view * fix config saving * add replace rules field * default column layout and add field sizing * clean up field template * refactor profile settings to match rjsf forms * tweaks * refactor frigate+ view and make tweaks to sections * show frigate+ model info in detection model settings when using a frigate+ model * update restartRequired for all fields * fix restart fields * tweaks and add ability enable disabled cameras more backend changes required * require restart when enabling camera that is disabled in config * disable save when form is invalid * refactor ffmpeg section for readability * change label * clean up camera inputs fields * misc tweaks to ffmpeg section - add raw paths endpoint to ensure credentials get saved - restart required tooltip * maintenance settings tweaks * don't mutate with lodash * fix description re-rendering for nullable object fields * hide reindex field * update rjsf * add frigate+ description to settings pane * disable save all when any section is invalid * show translated field name in validation error pane * clean up * remove unused * fix genai merge * fix genai
This commit is contained in:
parent
eeefbf2bb5
commit
e7250f24cb
@ -229,6 +229,7 @@ Reolink
|
||||
restream
|
||||
restreamed
|
||||
restreaming
|
||||
RJSF
|
||||
rkmpp
|
||||
rknn
|
||||
rkrga
|
||||
|
||||
@ -49,10 +49,12 @@ from frigate.types import JobStatusTypesEnum
|
||||
from frigate.util.builtin import (
|
||||
clean_camera_user_pass,
|
||||
flatten_config_data,
|
||||
load_labels,
|
||||
process_config_query_string,
|
||||
update_yaml_file_bulk,
|
||||
)
|
||||
from frigate.util.config import find_config_file
|
||||
from frigate.util.schema import get_config_schema
|
||||
from frigate.util.services import (
|
||||
get_nvidia_driver_info,
|
||||
process_logs,
|
||||
@ -77,9 +79,7 @@ def is_healthy():
|
||||
|
||||
@router.get("/config/schema.json", dependencies=[Depends(allow_public())])
|
||||
def config_schema(request: Request):
|
||||
return Response(
|
||||
content=request.app.frigate_config.schema_json(), media_type="application/json"
|
||||
)
|
||||
return JSONResponse(content=get_config_schema(FrigateConfig))
|
||||
|
||||
|
||||
@router.get(
|
||||
@ -125,6 +125,10 @@ def config(request: Request):
|
||||
config: dict[str, dict[str, Any]] = config_obj.model_dump(
|
||||
mode="json", warnings="none", exclude_none=True
|
||||
)
|
||||
config["detectors"] = {
|
||||
name: detector.model_dump(mode="json", warnings="none", exclude_none=True)
|
||||
for name, detector in config_obj.detectors.items()
|
||||
}
|
||||
|
||||
# remove the mqtt password
|
||||
config["mqtt"].pop("password", None)
|
||||
@ -195,6 +199,54 @@ def config(request: Request):
|
||||
return JSONResponse(content=config)
|
||||
|
||||
|
||||
@router.get("/ffmpeg/presets", dependencies=[Depends(allow_any_authenticated())])
|
||||
def ffmpeg_presets():
|
||||
"""Return available ffmpeg preset keys for config UI usage."""
|
||||
|
||||
# Whitelist based on documented presets in ffmpeg_presets.md
|
||||
hwaccel_presets = [
|
||||
"preset-rpi-64-h264",
|
||||
"preset-rpi-64-h265",
|
||||
"preset-vaapi",
|
||||
"preset-intel-qsv-h264",
|
||||
"preset-intel-qsv-h265",
|
||||
"preset-nvidia",
|
||||
"preset-jetson-h264",
|
||||
"preset-jetson-h265",
|
||||
"preset-rkmpp",
|
||||
]
|
||||
input_presets = [
|
||||
"preset-http-jpeg-generic",
|
||||
"preset-http-mjpeg-generic",
|
||||
"preset-http-reolink",
|
||||
"preset-rtmp-generic",
|
||||
"preset-rtsp-generic",
|
||||
"preset-rtsp-restream",
|
||||
"preset-rtsp-restream-low-latency",
|
||||
"preset-rtsp-udp",
|
||||
"preset-rtsp-blue-iris",
|
||||
]
|
||||
record_output_presets = [
|
||||
"preset-record-generic",
|
||||
"preset-record-generic-audio-copy",
|
||||
"preset-record-generic-audio-aac",
|
||||
"preset-record-mjpeg",
|
||||
"preset-record-jpeg",
|
||||
"preset-record-ubiquiti",
|
||||
]
|
||||
|
||||
return JSONResponse(
|
||||
content={
|
||||
"hwaccel_args": hwaccel_presets,
|
||||
"input_args": input_presets,
|
||||
"output_args": {
|
||||
"record": record_output_presets,
|
||||
"detect": [],
|
||||
},
|
||||
}
|
||||
)
|
||||
|
||||
|
||||
@router.get("/config/raw_paths", dependencies=[Depends(require_role(["admin"]))])
|
||||
def config_raw_paths(request: Request):
|
||||
"""Admin-only endpoint that returns camera paths and go2rtc streams without credential masking."""
|
||||
@ -755,6 +807,12 @@ def get_sub_labels(split_joined: Optional[int] = None):
|
||||
return JSONResponse(content=sub_labels)
|
||||
|
||||
|
||||
@router.get("/audio_labels", dependencies=[Depends(allow_any_authenticated())])
|
||||
def get_audio_labels():
|
||||
labels = load_labels("/audio-labelmap.txt", prefill=521)
|
||||
return JSONResponse(content=labels)
|
||||
|
||||
|
||||
@router.get("/plus/models", dependencies=[Depends(allow_any_authenticated())])
|
||||
def plusModels(request: Request, filterByCurrentModelDetector: bool = False):
|
||||
if not request.app.frigate_config.plus_api.is_active():
|
||||
|
||||
@ -8,39 +8,63 @@ __all__ = ["AuthConfig"]
|
||||
|
||||
|
||||
class AuthConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=True, title="Enable authentication")
|
||||
enabled: bool = Field(
|
||||
default=True,
|
||||
title="Enable authentication",
|
||||
description="Enable native authentication for the Frigate UI.",
|
||||
)
|
||||
reset_admin_password: bool = Field(
|
||||
default=False, title="Reset the admin password on startup"
|
||||
default=False,
|
||||
title="Reset admin password",
|
||||
description="If true, reset the admin user's password on startup and print the new password in logs.",
|
||||
)
|
||||
cookie_name: str = Field(
|
||||
default="frigate_token", title="Name for jwt token cookie", pattern=r"^[a-z_]+$"
|
||||
default="frigate_token",
|
||||
title="JWT cookie name",
|
||||
description="Name of the cookie used to store the JWT token for native authentication.",
|
||||
pattern=r"^[a-z_]+$",
|
||||
)
|
||||
cookie_secure: bool = Field(
|
||||
default=False,
|
||||
title="Secure cookie flag",
|
||||
description="Set the secure flag on the auth cookie; should be true when using TLS.",
|
||||
)
|
||||
cookie_secure: bool = Field(default=False, title="Set secure flag on cookie")
|
||||
session_length: int = Field(
|
||||
default=86400, title="Session length for jwt session tokens", ge=60
|
||||
default=86400,
|
||||
title="Session length",
|
||||
description="Session duration in seconds for JWT-based sessions.",
|
||||
ge=60,
|
||||
)
|
||||
refresh_time: int = Field(
|
||||
default=1800,
|
||||
title="Refresh the session if it is going to expire in this many seconds",
|
||||
title="Session refresh window",
|
||||
description="When a session is within this many seconds of expiring, refresh it back to full length.",
|
||||
ge=30,
|
||||
)
|
||||
failed_login_rate_limit: Optional[str] = Field(
|
||||
default=None,
|
||||
title="Rate limits for failed login attempts.",
|
||||
title="Failed login limits",
|
||||
description="Rate limiting rules for failed login attempts to reduce brute-force attacks.",
|
||||
)
|
||||
trusted_proxies: list[str] = Field(
|
||||
default=[],
|
||||
title="Trusted proxies for determining IP address to rate limit",
|
||||
title="Trusted proxies",
|
||||
description="List of trusted proxy IPs used when determining client IP for rate limiting.",
|
||||
)
|
||||
# As of Feb 2023, OWASP recommends 600000 iterations for PBKDF2-SHA256
|
||||
hash_iterations: int = Field(default=600000, title="Password hash iterations")
|
||||
hash_iterations: int = Field(
|
||||
default=600000,
|
||||
title="Hash iterations",
|
||||
description="Number of PBKDF2-SHA256 iterations to use when hashing user passwords.",
|
||||
)
|
||||
roles: Dict[str, List[str]] = Field(
|
||||
default_factory=dict,
|
||||
title="Role to camera mappings. Empty list grants access to all cameras.",
|
||||
title="Role mappings",
|
||||
description="Map roles to camera lists. An empty list grants access to all cameras for the role.",
|
||||
)
|
||||
admin_first_time_login: Optional[bool] = Field(
|
||||
default=False,
|
||||
title="Internal field to expose first-time admin login flag to the UI",
|
||||
title="First-time admin flag",
|
||||
description=(
|
||||
"When true the UI may show a help link on the login page informing users how to sign in after an admin password reset. "
|
||||
),
|
||||
|
||||
@ -17,25 +17,45 @@ class AudioFilterConfig(FrigateBaseModel):
|
||||
default=0.8,
|
||||
ge=AUDIO_MIN_CONFIDENCE,
|
||||
lt=1.0,
|
||||
title="Minimum detection confidence threshold for audio to be counted.",
|
||||
title="Minimum audio confidence",
|
||||
description="Minimum confidence threshold for the audio event to be counted.",
|
||||
)
|
||||
|
||||
|
||||
class AudioConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable audio events.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable audio detection",
|
||||
description="Enable or disable audio event detection for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
max_not_heard: int = Field(
|
||||
default=30, title="Seconds of not hearing the type of audio to end the event."
|
||||
default=30,
|
||||
title="End timeout",
|
||||
description="Amount of seconds without the configured audio type before the audio event is ended.",
|
||||
)
|
||||
min_volume: int = Field(
|
||||
default=500, title="Min volume required to run audio detection."
|
||||
default=500,
|
||||
title="Minimum volume",
|
||||
description="Minimum RMS volume threshold required to run audio detection; lower values increase sensitivity (e.g., 200 high, 500 medium, 1000 low).",
|
||||
)
|
||||
listen: list[str] = Field(
|
||||
default=DEFAULT_LISTEN_AUDIO, title="Audio to listen for."
|
||||
default=DEFAULT_LISTEN_AUDIO,
|
||||
title="Listen types",
|
||||
description="List of audio event types to detect (for example: bark, fire_alarm, scream, speech, yell).",
|
||||
)
|
||||
filters: Optional[dict[str, AudioFilterConfig]] = Field(
|
||||
None, title="Audio filters."
|
||||
None,
|
||||
title="Audio filters",
|
||||
description="Per-audio-type filter settings such as confidence thresholds used to reduce false positives.",
|
||||
)
|
||||
enabled_in_config: Optional[bool] = Field(
|
||||
None, title="Keep track of original state of audio detection."
|
||||
None,
|
||||
title="Original audio state",
|
||||
description="Indicates whether audio detection was originally enabled in the static config file.",
|
||||
)
|
||||
num_threads: int = Field(
|
||||
default=2,
|
||||
title="Detection threads",
|
||||
description="Number of threads to use for audio detection processing.",
|
||||
ge=1,
|
||||
)
|
||||
num_threads: int = Field(default=2, title="Number of detection threads", ge=1)
|
||||
|
||||
@ -29,45 +29,88 @@ class BirdseyeModeEnum(str, Enum):
|
||||
|
||||
class BirdseyeLayoutConfig(FrigateBaseModel):
|
||||
scaling_factor: float = Field(
|
||||
default=2.0, title="Birdseye Scaling Factor", ge=1.0, le=5.0
|
||||
default=2.0,
|
||||
title="Scaling factor",
|
||||
description="Scaling factor used by the layout calculator (range 1.0 to 5.0).",
|
||||
ge=1.0,
|
||||
le=5.0,
|
||||
)
|
||||
max_cameras: Optional[int] = Field(
|
||||
default=None,
|
||||
title="Max cameras",
|
||||
description="Maximum number of cameras to display at once in Birdseye; shows the most recent cameras.",
|
||||
)
|
||||
max_cameras: Optional[int] = Field(default=None, title="Max cameras")
|
||||
|
||||
|
||||
class BirdseyeConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=True, title="Enable birdseye view.")
|
||||
enabled: bool = Field(
|
||||
default=True,
|
||||
title="Enable Birdseye",
|
||||
description="Enable or disable the Birdseye view feature.",
|
||||
)
|
||||
mode: BirdseyeModeEnum = Field(
|
||||
default=BirdseyeModeEnum.objects, title="Tracking mode."
|
||||
default=BirdseyeModeEnum.objects,
|
||||
title="Tracking mode",
|
||||
description="Mode for including cameras in Birdseye: 'objects', 'motion', or 'continuous'.",
|
||||
)
|
||||
|
||||
restream: bool = Field(default=False, title="Restream birdseye via RTSP.")
|
||||
width: int = Field(default=1280, title="Birdseye width.")
|
||||
height: int = Field(default=720, title="Birdseye height.")
|
||||
restream: bool = Field(
|
||||
default=False,
|
||||
title="Restream RTSP",
|
||||
description="Re-stream the Birdseye output as an RTSP feed; enabling this will keep Birdseye running continuously.",
|
||||
)
|
||||
width: int = Field(
|
||||
default=1280,
|
||||
title="Width",
|
||||
description="Output width (pixels) of the composed Birdseye frame.",
|
||||
)
|
||||
height: int = Field(
|
||||
default=720,
|
||||
title="Height",
|
||||
description="Output height (pixels) of the composed Birdseye frame.",
|
||||
)
|
||||
quality: int = Field(
|
||||
default=8,
|
||||
title="Encoding quality.",
|
||||
title="Encoding quality",
|
||||
description="Encoding quality for the Birdseye mpeg1 feed (1 highest quality, 31 lowest).",
|
||||
ge=1,
|
||||
le=31,
|
||||
)
|
||||
inactivity_threshold: int = Field(
|
||||
default=30, title="Birdseye Inactivity Threshold", gt=0
|
||||
default=30,
|
||||
title="Inactivity threshold",
|
||||
description="Seconds of inactivity after which a camera will stop being shown in Birdseye.",
|
||||
gt=0,
|
||||
)
|
||||
layout: BirdseyeLayoutConfig = Field(
|
||||
default_factory=BirdseyeLayoutConfig, title="Birdseye Layout Config"
|
||||
default_factory=BirdseyeLayoutConfig,
|
||||
title="Layout",
|
||||
description="Layout options for the Birdseye composition.",
|
||||
)
|
||||
idle_heartbeat_fps: float = Field(
|
||||
default=0.0,
|
||||
ge=0.0,
|
||||
le=10.0,
|
||||
title="Idle heartbeat FPS (0 disables, max 10)",
|
||||
title="Idle heartbeat FPS",
|
||||
description="Frames-per-second to resend the last composed Birdseye frame when idle; set to 0 to disable.",
|
||||
)
|
||||
|
||||
|
||||
# uses BaseModel because some global attributes are not available at the camera level
|
||||
class BirdseyeCameraConfig(BaseModel):
|
||||
enabled: bool = Field(default=True, title="Enable birdseye view for camera.")
|
||||
enabled: bool = Field(
|
||||
default=True,
|
||||
title="Enable Birdseye",
|
||||
description="Enable or disable the Birdseye view feature.",
|
||||
)
|
||||
mode: BirdseyeModeEnum = Field(
|
||||
default=BirdseyeModeEnum.objects, title="Tracking mode for camera."
|
||||
default=BirdseyeModeEnum.objects,
|
||||
title="Tracking mode",
|
||||
description="Mode for including cameras in Birdseye: 'objects', 'motion', or 'continuous'.",
|
||||
)
|
||||
|
||||
order: int = Field(default=0, title="Position of the camera in the birdseye view.")
|
||||
order: int = Field(
|
||||
default=0,
|
||||
title="Position",
|
||||
description="Numeric position controlling the camera's ordering in the Birdseye layout.",
|
||||
)
|
||||
|
||||
@ -50,10 +50,17 @@ class CameraTypeEnum(str, Enum):
|
||||
|
||||
|
||||
class CameraConfig(FrigateBaseModel):
|
||||
name: Optional[str] = Field(None, title="Camera name.", pattern=REGEX_CAMERA_NAME)
|
||||
name: Optional[str] = Field(
|
||||
None,
|
||||
title="Camera name",
|
||||
description="Camera name is required",
|
||||
pattern=REGEX_CAMERA_NAME,
|
||||
)
|
||||
|
||||
friendly_name: Optional[str] = Field(
|
||||
None, title="Camera friendly name used in the Frigate UI."
|
||||
None,
|
||||
title="Friendly name",
|
||||
description="Camera friendly name used in the Frigate UI",
|
||||
)
|
||||
|
||||
@model_validator(mode="before")
|
||||
@ -63,80 +70,129 @@ class CameraConfig(FrigateBaseModel):
|
||||
pass
|
||||
return values
|
||||
|
||||
enabled: bool = Field(default=True, title="Enable camera.")
|
||||
enabled: bool = Field(default=True, title="Enabled", description="Enabled")
|
||||
|
||||
# Options with global fallback
|
||||
audio: AudioConfig = Field(
|
||||
default_factory=AudioConfig, title="Audio events configuration."
|
||||
default_factory=AudioConfig,
|
||||
title="Audio events",
|
||||
description="Settings for audio-based event detection for this camera.",
|
||||
)
|
||||
audio_transcription: CameraAudioTranscriptionConfig = Field(
|
||||
default_factory=CameraAudioTranscriptionConfig,
|
||||
title="Audio transcription config.",
|
||||
title="Audio transcription",
|
||||
description="Settings for live and speech audio transcription used for events and live captions.",
|
||||
)
|
||||
birdseye: BirdseyeCameraConfig = Field(
|
||||
default_factory=BirdseyeCameraConfig, title="Birdseye camera configuration."
|
||||
default_factory=BirdseyeCameraConfig,
|
||||
title="Birdseye",
|
||||
description="Settings for the Birdseye composite view that composes multiple camera feeds into a single layout.",
|
||||
)
|
||||
detect: DetectConfig = Field(
|
||||
default_factory=DetectConfig, title="Object detection configuration."
|
||||
default_factory=DetectConfig,
|
||||
title="Object Detection",
|
||||
description="Settings for the detection/detect role used to run object detection and initialize trackers.",
|
||||
)
|
||||
face_recognition: CameraFaceRecognitionConfig = Field(
|
||||
default_factory=CameraFaceRecognitionConfig, title="Face recognition config."
|
||||
default_factory=CameraFaceRecognitionConfig,
|
||||
title="Face recognition",
|
||||
description="Settings for face detection and recognition for this camera.",
|
||||
)
|
||||
ffmpeg: CameraFfmpegConfig = Field(
|
||||
title="FFmpeg",
|
||||
description="FFmpeg settings including binary path, args, hwaccel options, and per-role output args.",
|
||||
)
|
||||
ffmpeg: CameraFfmpegConfig = Field(title="FFmpeg configuration for the camera.")
|
||||
live: CameraLiveConfig = Field(
|
||||
default_factory=CameraLiveConfig, title="Live playback settings."
|
||||
default_factory=CameraLiveConfig,
|
||||
title="Live playback",
|
||||
description="Settings used by the Web UI to control live stream selection, resolution and quality.",
|
||||
)
|
||||
lpr: CameraLicensePlateRecognitionConfig = Field(
|
||||
default_factory=CameraLicensePlateRecognitionConfig, title="LPR config."
|
||||
default_factory=CameraLicensePlateRecognitionConfig,
|
||||
title="License Plate Recognition",
|
||||
description="License plate recognition settings including detection thresholds, formatting, and known plates.",
|
||||
)
|
||||
motion: MotionConfig = Field(
|
||||
None,
|
||||
title="Motion detection",
|
||||
description="Default motion detection settings for this camera.",
|
||||
)
|
||||
motion: MotionConfig = Field(None, title="Motion detection configuration.")
|
||||
objects: ObjectConfig = Field(
|
||||
default_factory=ObjectConfig, title="Object configuration."
|
||||
default_factory=ObjectConfig,
|
||||
title="Objects",
|
||||
description="Object tracking defaults including which labels to track and per-object filters.",
|
||||
)
|
||||
record: RecordConfig = Field(
|
||||
default_factory=RecordConfig, title="Record configuration."
|
||||
default_factory=RecordConfig,
|
||||
title="Recording",
|
||||
description="Recording and retention settings for this camera.",
|
||||
)
|
||||
review: ReviewConfig = Field(
|
||||
default_factory=ReviewConfig, title="Review configuration."
|
||||
default_factory=ReviewConfig,
|
||||
title="Review",
|
||||
description="Settings that control alerts, detections, and GenAI review summaries used by the UI and storage for this camera.",
|
||||
)
|
||||
semantic_search: CameraSemanticSearchConfig = Field(
|
||||
default_factory=CameraSemanticSearchConfig,
|
||||
title="Semantic search configuration.",
|
||||
title="Semantic Search",
|
||||
description="Settings for semantic search which builds and queries object embeddings to find similar items.",
|
||||
)
|
||||
snapshots: SnapshotsConfig = Field(
|
||||
default_factory=SnapshotsConfig, title="Snapshot configuration."
|
||||
default_factory=SnapshotsConfig,
|
||||
title="Snapshots",
|
||||
description="Settings for saved JPEG snapshots of tracked objects for this camera.",
|
||||
)
|
||||
timestamp_style: TimestampStyleConfig = Field(
|
||||
default_factory=TimestampStyleConfig, title="Timestamp style configuration."
|
||||
default_factory=TimestampStyleConfig,
|
||||
title="Timestamp style",
|
||||
description="Styling options for in-feed timestamps applied to recordings and snapshots.",
|
||||
)
|
||||
|
||||
# Options without global fallback
|
||||
best_image_timeout: int = Field(
|
||||
default=60,
|
||||
title="How long to wait for the image with the highest confidence score.",
|
||||
title="Best image timeout",
|
||||
description="How long to wait for the image with the highest confidence score.",
|
||||
)
|
||||
mqtt: CameraMqttConfig = Field(
|
||||
default_factory=CameraMqttConfig, title="MQTT configuration."
|
||||
default_factory=CameraMqttConfig,
|
||||
title="MQTT",
|
||||
description="MQTT image publishing settings.",
|
||||
)
|
||||
notifications: NotificationConfig = Field(
|
||||
default_factory=NotificationConfig, title="Notifications configuration."
|
||||
default_factory=NotificationConfig,
|
||||
title="Notifications",
|
||||
description="Settings to enable and control notifications for this camera.",
|
||||
)
|
||||
onvif: OnvifConfig = Field(
|
||||
default_factory=OnvifConfig, title="Camera Onvif Configuration."
|
||||
default_factory=OnvifConfig,
|
||||
title="ONVIF",
|
||||
description="ONVIF connection and PTZ autotracking settings for this camera.",
|
||||
)
|
||||
type: CameraTypeEnum = Field(
|
||||
default=CameraTypeEnum.generic,
|
||||
title="Camera type",
|
||||
description="Camera Type",
|
||||
)
|
||||
type: CameraTypeEnum = Field(default=CameraTypeEnum.generic, title="Camera Type")
|
||||
ui: CameraUiConfig = Field(
|
||||
default_factory=CameraUiConfig, title="Camera UI Modifications."
|
||||
default_factory=CameraUiConfig,
|
||||
title="Camera UI",
|
||||
description="Display ordering and visibility for this camera in the UI. Ordering affects the default dashboard. For more granular control, use camera groups.",
|
||||
)
|
||||
webui_url: Optional[str] = Field(
|
||||
None,
|
||||
title="URL to visit the camera directly from system page",
|
||||
title="Camera URL",
|
||||
description="URL to visit the camera directly from system page",
|
||||
)
|
||||
zones: dict[str, ZoneConfig] = Field(
|
||||
default_factory=dict, title="Zone configuration."
|
||||
default_factory=dict,
|
||||
title="Zones",
|
||||
description="Zones allow you to define a specific area of the frame so you can determine whether or not an object is within a particular area.",
|
||||
)
|
||||
enabled_in_config: Optional[bool] = Field(
|
||||
default=None, title="Keep track of original state of camera."
|
||||
default=None,
|
||||
title="Original camera state",
|
||||
description="Keep track of original state of camera.",
|
||||
)
|
||||
|
||||
_ffmpeg_cmds: list[dict[str, list[str]]] = PrivateAttr()
|
||||
|
||||
@ -8,56 +8,82 @@ __all__ = ["DetectConfig", "StationaryConfig", "StationaryMaxFramesConfig"]
|
||||
|
||||
|
||||
class StationaryMaxFramesConfig(FrigateBaseModel):
|
||||
default: Optional[int] = Field(default=None, title="Default max frames.", ge=1)
|
||||
default: Optional[int] = Field(
|
||||
default=None,
|
||||
title="Default max frames",
|
||||
description="Default maximum frames to track a stationary object before stopping.",
|
||||
ge=1,
|
||||
)
|
||||
objects: dict[str, int] = Field(
|
||||
default_factory=dict, title="Object specific max frames."
|
||||
default_factory=dict,
|
||||
title="Object max frames",
|
||||
description="Per-object overrides for maximum frames to track stationary objects.",
|
||||
)
|
||||
|
||||
|
||||
class StationaryConfig(FrigateBaseModel):
|
||||
interval: Optional[int] = Field(
|
||||
default=None,
|
||||
title="Frame interval for checking stationary objects.",
|
||||
title="Stationary interval",
|
||||
description="How often (in frames) to run a detection check to confirm a stationary object.",
|
||||
gt=0,
|
||||
)
|
||||
threshold: Optional[int] = Field(
|
||||
default=None,
|
||||
title="Number of frames without a position change for an object to be considered stationary",
|
||||
title="Stationary threshold",
|
||||
description="Number of frames with no position change required to mark an object as stationary.",
|
||||
ge=1,
|
||||
)
|
||||
max_frames: StationaryMaxFramesConfig = Field(
|
||||
default_factory=StationaryMaxFramesConfig,
|
||||
title="Max frames for stationary objects.",
|
||||
title="Max frames",
|
||||
description="Limits how long stationary objects are tracked before being discarded.",
|
||||
)
|
||||
classifier: bool = Field(
|
||||
default=True,
|
||||
title="Enable visual classifier for determing if objects with jittery bounding boxes are stationary.",
|
||||
title="Enable visual classifier",
|
||||
description="Use a visual classifier to detect truly stationary objects even when bounding boxes jitter.",
|
||||
)
|
||||
|
||||
|
||||
class DetectConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Detection Enabled.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Detection enabled",
|
||||
description="Enable or disable object detection for all cameras; can be overridden per-camera. Detection must be enabled for object tracking to run.",
|
||||
)
|
||||
height: Optional[int] = Field(
|
||||
default=None, title="Height of the stream for the detect role."
|
||||
default=None,
|
||||
title="Detect height",
|
||||
description="Height (pixels) of frames used for the detect stream; leave empty to use the native stream resolution.",
|
||||
)
|
||||
width: Optional[int] = Field(
|
||||
default=None, title="Width of the stream for the detect role."
|
||||
default=None,
|
||||
title="Detect width",
|
||||
description="Width (pixels) of frames used for the detect stream; leave empty to use the native stream resolution.",
|
||||
)
|
||||
fps: int = Field(
|
||||
default=5, title="Number of frames per second to process through detection."
|
||||
default=5,
|
||||
title="Detect FPS",
|
||||
description="Desired frames per second to run detection on; lower values reduce CPU usage (recommended value is 5, only set higher - at most 10 - if tracking extremely fast moving objects).",
|
||||
)
|
||||
min_initialized: Optional[int] = Field(
|
||||
default=None,
|
||||
title="Minimum number of consecutive hits for an object to be initialized by the tracker.",
|
||||
title="Minimum initialization frames",
|
||||
description="Number of consecutive detection hits required before creating a tracked object. Increase to reduce false initializations. Default value is fps divided by 2.",
|
||||
)
|
||||
max_disappeared: Optional[int] = Field(
|
||||
default=None,
|
||||
title="Maximum number of frames the object can disappear before detection ends.",
|
||||
title="Maximum disappeared frames",
|
||||
description="Number of frames without a detection before a tracked object is considered gone.",
|
||||
)
|
||||
stationary: StationaryConfig = Field(
|
||||
default_factory=StationaryConfig,
|
||||
title="Stationary objects config.",
|
||||
title="Stationary objects config",
|
||||
description="Settings to detect and manage objects that remain stationary for a period of time.",
|
||||
)
|
||||
annotation_offset: int = Field(
|
||||
default=0, title="Milliseconds to offset detect annotations by."
|
||||
default=0,
|
||||
title="Annotation offset",
|
||||
description="Milliseconds to shift detect annotations to better align timeline bounding boxes with recordings; can be positive or negative.",
|
||||
)
|
||||
|
||||
@ -35,39 +35,58 @@ DETECT_FFMPEG_OUTPUT_ARGS_DEFAULT = [
|
||||
class FfmpegOutputArgsConfig(FrigateBaseModel):
|
||||
detect: Union[str, list[str]] = Field(
|
||||
default=DETECT_FFMPEG_OUTPUT_ARGS_DEFAULT,
|
||||
title="Detect role FFmpeg output arguments.",
|
||||
title="Detect output arguments",
|
||||
description="Default output arguments for detect role streams.",
|
||||
)
|
||||
record: Union[str, list[str]] = Field(
|
||||
default=RECORD_FFMPEG_OUTPUT_ARGS_DEFAULT,
|
||||
title="Record role FFmpeg output arguments.",
|
||||
title="Record output arguments",
|
||||
description="Default output arguments for record role streams.",
|
||||
)
|
||||
|
||||
|
||||
class FfmpegConfig(FrigateBaseModel):
|
||||
path: str = Field(default="default", title="FFmpeg path")
|
||||
path: str = Field(
|
||||
default="default",
|
||||
title="FFmpeg path",
|
||||
description='Path to the FFmpeg binary to use or a version alias ("5.0" or "7.0").',
|
||||
)
|
||||
global_args: Union[str, list[str]] = Field(
|
||||
default=FFMPEG_GLOBAL_ARGS_DEFAULT, title="Global FFmpeg arguments."
|
||||
default=FFMPEG_GLOBAL_ARGS_DEFAULT,
|
||||
title="FFmpeg global arguments",
|
||||
description="Global arguments passed to FFmpeg processes.",
|
||||
)
|
||||
hwaccel_args: Union[str, list[str]] = Field(
|
||||
default="auto", title="FFmpeg hardware acceleration arguments."
|
||||
default="auto",
|
||||
title="Hardware acceleration arguments",
|
||||
description="Hardware acceleration arguments for FFmpeg. Provider-specific presets are recommended.",
|
||||
)
|
||||
input_args: Union[str, list[str]] = Field(
|
||||
default=FFMPEG_INPUT_ARGS_DEFAULT, title="FFmpeg input arguments."
|
||||
default=FFMPEG_INPUT_ARGS_DEFAULT,
|
||||
title="Input arguments",
|
||||
description="Input arguments applied to FFmpeg input streams.",
|
||||
)
|
||||
output_args: FfmpegOutputArgsConfig = Field(
|
||||
default_factory=FfmpegOutputArgsConfig,
|
||||
title="FFmpeg output arguments per role.",
|
||||
title="Output arguments",
|
||||
description="Default output arguments used for different FFmpeg roles such as detect and record.",
|
||||
)
|
||||
retry_interval: float = Field(
|
||||
default=10.0,
|
||||
title="Time in seconds to wait before FFmpeg retries connecting to the camera.",
|
||||
title="FFmpeg retry time",
|
||||
description="Seconds to wait before attempting to reconnect a camera stream after failure. Default is 10.",
|
||||
gt=0.0,
|
||||
)
|
||||
apple_compatibility: bool = Field(
|
||||
default=False,
|
||||
title="Set tag on HEVC (H.265) recording stream to improve compatibility with Apple players.",
|
||||
title="Apple compatibility",
|
||||
description="Enable HEVC tagging for better Apple player compatibility when recording H.265.",
|
||||
)
|
||||
gpu: int = Field(
|
||||
default=0,
|
||||
title="GPU index",
|
||||
description="Default GPU index used for hardware acceleration if available.",
|
||||
)
|
||||
gpu: int = Field(default=0, title="GPU index to use for hardware acceleration.")
|
||||
|
||||
@property
|
||||
def ffmpeg_path(self) -> str:
|
||||
@ -95,21 +114,36 @@ class CameraRoleEnum(str, Enum):
|
||||
|
||||
|
||||
class CameraInput(FrigateBaseModel):
|
||||
path: EnvString = Field(title="Camera input path.")
|
||||
roles: list[CameraRoleEnum] = Field(title="Roles assigned to this input.")
|
||||
path: EnvString = Field(
|
||||
title="Input path",
|
||||
description="Camera input stream URL or path.",
|
||||
)
|
||||
roles: list[CameraRoleEnum] = Field(
|
||||
title="Input roles",
|
||||
description="Roles for this input stream.",
|
||||
)
|
||||
global_args: Union[str, list[str]] = Field(
|
||||
default_factory=list, title="FFmpeg global arguments."
|
||||
default_factory=list,
|
||||
title="FFmpeg global arguments",
|
||||
description="FFmpeg global arguments for this input stream.",
|
||||
)
|
||||
hwaccel_args: Union[str, list[str]] = Field(
|
||||
default_factory=list, title="FFmpeg hardware acceleration arguments."
|
||||
default_factory=list,
|
||||
title="Hardware acceleration arguments",
|
||||
description="Hardware acceleration arguments for this input stream.",
|
||||
)
|
||||
input_args: Union[str, list[str]] = Field(
|
||||
default_factory=list, title="FFmpeg input arguments."
|
||||
default_factory=list,
|
||||
title="Input arguments",
|
||||
description="Input arguments specific to this stream.",
|
||||
)
|
||||
|
||||
|
||||
class CameraFfmpegConfig(FfmpegConfig):
|
||||
inputs: list[CameraInput] = Field(title="Camera inputs.")
|
||||
inputs: list[CameraInput] = Field(
|
||||
title="Camera inputs",
|
||||
description="List of input stream definitions (paths and roles) for this camera.",
|
||||
)
|
||||
|
||||
@field_validator("inputs")
|
||||
@classmethod
|
||||
|
||||
@ -67,6 +67,3 @@ class GenAIConfig(FrigateBaseModel):
|
||||
description="Runtime options passed to the provider for each inference call.",
|
||||
json_schema_extra={"additionalProperties": {"type": "string"}},
|
||||
)
|
||||
runtime_options: dict[str, Any] = Field(
|
||||
default={}, title="Options to pass during inference calls."
|
||||
)
|
||||
|
||||
@ -10,7 +10,18 @@ __all__ = ["CameraLiveConfig"]
|
||||
class CameraLiveConfig(FrigateBaseModel):
|
||||
streams: Dict[str, str] = Field(
|
||||
default_factory=list,
|
||||
title="Friendly names and restream names to use for live view.",
|
||||
title="Live stream names",
|
||||
description="Mapping of configured stream names to restream/go2rtc names used for live playback.",
|
||||
)
|
||||
height: int = Field(
|
||||
default=720,
|
||||
title="Live height",
|
||||
description="Height (pixels) to render the jsmpeg live stream in the Web UI; must be <= detect stream height.",
|
||||
)
|
||||
quality: int = Field(
|
||||
default=8,
|
||||
ge=1,
|
||||
le=31,
|
||||
title="Live quality",
|
||||
description="Encoding quality for the jsmpeg stream (1 highest, 31 lowest).",
|
||||
)
|
||||
height: int = Field(default=720, title="Live camera view height")
|
||||
quality: int = Field(default=8, ge=1, le=31, title="Live camera view quality")
|
||||
|
||||
@ -8,30 +8,64 @@ __all__ = ["MotionConfig"]
|
||||
|
||||
|
||||
class MotionConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=True, title="Enable motion on all cameras.")
|
||||
enabled: bool = Field(
|
||||
default=True,
|
||||
title="Enable motion detection",
|
||||
description="Enable or disable motion detection for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
threshold: int = Field(
|
||||
default=30,
|
||||
title="Motion detection threshold (1-255).",
|
||||
title="Motion threshold",
|
||||
description="Pixel difference threshold used by the motion detector; higher values reduce sensitivity (range 1-255).",
|
||||
ge=1,
|
||||
le=255,
|
||||
)
|
||||
lightning_threshold: float = Field(
|
||||
default=0.8, title="Lightning detection threshold (0.3-1.0).", ge=0.3, le=1.0
|
||||
default=0.8,
|
||||
title="Lightning threshold",
|
||||
description="Threshold to detect and ignore brief lighting spikes (lower is more sensitive, values between 0.3 and 1.0).",
|
||||
ge=0.3,
|
||||
le=1.0,
|
||||
)
|
||||
improve_contrast: bool = Field(
|
||||
default=True,
|
||||
title="Improve contrast",
|
||||
description="Apply contrast improvement to frames before motion analysis to help detection.",
|
||||
)
|
||||
contour_area: Optional[int] = Field(
|
||||
default=10,
|
||||
title="Contour area",
|
||||
description="Minimum contour area in pixels required for a motion contour to be counted.",
|
||||
)
|
||||
delta_alpha: float = Field(
|
||||
default=0.2,
|
||||
title="Delta alpha",
|
||||
description="Alpha blending factor used in frame differencing for motion calculation.",
|
||||
)
|
||||
frame_alpha: float = Field(
|
||||
default=0.01,
|
||||
title="Frame alpha",
|
||||
description="Alpha value used when blending frames for motion preprocessing.",
|
||||
)
|
||||
frame_height: Optional[int] = Field(
|
||||
default=100,
|
||||
title="Frame height",
|
||||
description="Height in pixels to scale frames to when computing motion.",
|
||||
)
|
||||
improve_contrast: bool = Field(default=True, title="Improve Contrast")
|
||||
contour_area: Optional[int] = Field(default=10, title="Contour Area")
|
||||
delta_alpha: float = Field(default=0.2, title="Delta Alpha")
|
||||
frame_alpha: float = Field(default=0.01, title="Frame Alpha")
|
||||
frame_height: Optional[int] = Field(default=100, title="Frame Height")
|
||||
mask: Union[str, list[str]] = Field(
|
||||
default="", title="Coordinates polygon for the motion mask."
|
||||
default="",
|
||||
title="Mask coordinates",
|
||||
description="Ordered x,y coordinates defining the motion mask polygon used to include/exclude areas.",
|
||||
)
|
||||
mqtt_off_delay: int = Field(
|
||||
default=30,
|
||||
title="Delay for updating MQTT with no motion detected.",
|
||||
title="MQTT off delay",
|
||||
description="Seconds to wait after last motion before publishing an MQTT 'off' state.",
|
||||
)
|
||||
enabled_in_config: Optional[bool] = Field(
|
||||
default=None, title="Keep track of original state of motion detection."
|
||||
default=None,
|
||||
title="Original motion state",
|
||||
description="Indicates whether motion detection was enabled in the original static configuration.",
|
||||
)
|
||||
raw_mask: Union[str, list[str]] = ""
|
||||
|
||||
|
||||
@ -6,18 +6,40 @@ __all__ = ["CameraMqttConfig"]
|
||||
|
||||
|
||||
class CameraMqttConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=True, title="Send image over MQTT.")
|
||||
timestamp: bool = Field(default=True, title="Add timestamp to MQTT image.")
|
||||
bounding_box: bool = Field(default=True, title="Add bounding box to MQTT image.")
|
||||
crop: bool = Field(default=True, title="Crop MQTT image to detected object.")
|
||||
height: int = Field(default=270, title="MQTT image height.")
|
||||
enabled: bool = Field(
|
||||
default=True,
|
||||
title="Send image",
|
||||
description="Enable publishing image snapshots for objects to MQTT topics for this camera.",
|
||||
)
|
||||
timestamp: bool = Field(
|
||||
default=True,
|
||||
title="Add timestamp",
|
||||
description="Overlay a timestamp on images published to MQTT.",
|
||||
)
|
||||
bounding_box: bool = Field(
|
||||
default=True,
|
||||
title="Add bounding box",
|
||||
description="Draw bounding boxes on images published over MQTT.",
|
||||
)
|
||||
crop: bool = Field(
|
||||
default=True,
|
||||
title="Crop image",
|
||||
description="Crop images published to MQTT to the detected object's bounding box.",
|
||||
)
|
||||
height: int = Field(
|
||||
default=270,
|
||||
title="Image height",
|
||||
description="Height (pixels) to resize images published over MQTT.",
|
||||
)
|
||||
required_zones: list[str] = Field(
|
||||
default_factory=list,
|
||||
title="List of required zones to be entered in order to send the image.",
|
||||
title="Required zones",
|
||||
description="Zones that an object must enter for an MQTT image to be published.",
|
||||
)
|
||||
quality: int = Field(
|
||||
default=70,
|
||||
title="Quality of the encoded jpeg (0-100).",
|
||||
title="JPEG quality",
|
||||
description="JPEG quality for images published to MQTT (0-100).",
|
||||
ge=0,
|
||||
le=100,
|
||||
)
|
||||
|
||||
@ -8,11 +8,24 @@ __all__ = ["NotificationConfig"]
|
||||
|
||||
|
||||
class NotificationConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable notifications")
|
||||
email: Optional[str] = Field(default=None, title="Email required for push.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable notifications",
|
||||
description="Enable or disable notifications for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
email: Optional[str] = Field(
|
||||
default=None,
|
||||
title="Notification email",
|
||||
description="Email address used for push notifications or required by certain notification providers.",
|
||||
)
|
||||
cooldown: int = Field(
|
||||
default=0, ge=0, title="Cooldown period for notifications (time in seconds)."
|
||||
default=0,
|
||||
ge=0,
|
||||
title="Cooldown period",
|
||||
description="Cooldown (seconds) between notifications to avoid spamming recipients.",
|
||||
)
|
||||
enabled_in_config: Optional[bool] = Field(
|
||||
default=None, title="Keep track of original state of notifications."
|
||||
default=None,
|
||||
title="Original notifications state",
|
||||
description="Indicates whether notifications were enabled in the original static configuration.",
|
||||
)
|
||||
|
||||
@ -13,30 +13,38 @@ DEFAULT_TRACKED_OBJECTS = ["person"]
|
||||
class FilterConfig(FrigateBaseModel):
|
||||
min_area: Union[int, float] = Field(
|
||||
default=0,
|
||||
title="Minimum area of bounding box for object to be counted. Can be pixels (int) or percentage (float between 0.000001 and 0.99).",
|
||||
title="Minimum object area",
|
||||
description="Minimum bounding box area (pixels or percentage) required for this object type. Can be pixels (int) or percentage (float between 0.000001 and 0.99).",
|
||||
)
|
||||
max_area: Union[int, float] = Field(
|
||||
default=24000000,
|
||||
title="Maximum area of bounding box for object to be counted. Can be pixels (int) or percentage (float between 0.000001 and 0.99).",
|
||||
title="Maximum object area",
|
||||
description="Maximum bounding box area (pixels or percentage) allowed for this object type. Can be pixels (int) or percentage (float between 0.000001 and 0.99).",
|
||||
)
|
||||
min_ratio: float = Field(
|
||||
default=0,
|
||||
title="Minimum ratio of bounding box's width/height for object to be counted.",
|
||||
title="Minimum aspect ratio",
|
||||
description="Minimum width/height ratio required for the bounding box to qualify.",
|
||||
)
|
||||
max_ratio: float = Field(
|
||||
default=24000000,
|
||||
title="Maximum ratio of bounding box's width/height for object to be counted.",
|
||||
title="Maximum aspect ratio",
|
||||
description="Maximum width/height ratio allowed for the bounding box to qualify.",
|
||||
)
|
||||
threshold: float = Field(
|
||||
default=0.7,
|
||||
title="Average detection confidence threshold for object to be counted.",
|
||||
title="Confidence threshold",
|
||||
description="Average detection confidence threshold required for the object to be considered a true positive.",
|
||||
)
|
||||
min_score: float = Field(
|
||||
default=0.5, title="Minimum detection confidence for object to be counted."
|
||||
default=0.5,
|
||||
title="Minimum confidence",
|
||||
description="Minimum single-frame detection confidence required for the object to be counted.",
|
||||
)
|
||||
mask: Optional[Union[str, list[str]]] = Field(
|
||||
default=None,
|
||||
title="Detection area polygon mask for this filter configuration.",
|
||||
title="Filter mask",
|
||||
description="Polygon coordinates defining where this filter applies within the frame.",
|
||||
)
|
||||
raw_mask: Union[str, list[str]] = ""
|
||||
|
||||
@ -51,46 +59,64 @@ class FilterConfig(FrigateBaseModel):
|
||||
|
||||
class GenAIObjectTriggerConfig(FrigateBaseModel):
|
||||
tracked_object_end: bool = Field(
|
||||
default=True, title="Send once the object is no longer tracked."
|
||||
default=True,
|
||||
title="Send on end",
|
||||
description="Send a request to GenAI when the tracked object ends.",
|
||||
)
|
||||
after_significant_updates: Optional[int] = Field(
|
||||
default=None,
|
||||
title="Send an early request to generative AI when X frames accumulated.",
|
||||
title="Early GenAI trigger",
|
||||
description="Send a request to GenAI after a specified number of significant updates for the tracked object.",
|
||||
ge=1,
|
||||
)
|
||||
|
||||
|
||||
class GenAIObjectConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable GenAI for camera.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable GenAI",
|
||||
description="Enable GenAI generation of descriptions for tracked objects by default.",
|
||||
)
|
||||
use_snapshot: bool = Field(
|
||||
default=False, title="Use snapshots for generating descriptions."
|
||||
default=False,
|
||||
title="Use snapshots",
|
||||
description="Use object snapshots instead of thumbnails for GenAI description generation.",
|
||||
)
|
||||
prompt: str = Field(
|
||||
default="Analyze the sequence of images containing the {label}. Focus on the likely intent or behavior of the {label} based on its actions and movement, rather than describing its appearance or the surroundings. Consider what the {label} is doing, why, and what it might do next.",
|
||||
title="Default caption prompt.",
|
||||
title="Caption prompt",
|
||||
description="Default prompt template used when generating descriptions with GenAI.",
|
||||
)
|
||||
object_prompts: dict[str, str] = Field(
|
||||
default_factory=dict, title="Object specific prompts."
|
||||
default_factory=dict,
|
||||
title="Object prompts",
|
||||
description="Per-object prompts to customize GenAI outputs for specific labels.",
|
||||
)
|
||||
|
||||
objects: Union[str, list[str]] = Field(
|
||||
default_factory=list,
|
||||
title="List of objects to run generative AI for.",
|
||||
title="GenAI objects",
|
||||
description="List of object labels to send to GenAI by default.",
|
||||
)
|
||||
required_zones: Union[str, list[str]] = Field(
|
||||
default_factory=list,
|
||||
title="List of required zones to be entered in order to run generative AI.",
|
||||
title="Required zones",
|
||||
description="Zones that must be entered for objects to qualify for GenAI description generation.",
|
||||
)
|
||||
debug_save_thumbnails: bool = Field(
|
||||
default=False,
|
||||
title="Save thumbnails sent to generative AI for debugging purposes.",
|
||||
title="Save thumbnails",
|
||||
description="Save thumbnails sent to GenAI for debugging and review.",
|
||||
)
|
||||
send_triggers: GenAIObjectTriggerConfig = Field(
|
||||
default_factory=GenAIObjectTriggerConfig,
|
||||
title="What triggers to use to send frames to generative AI for a tracked object.",
|
||||
title="GenAI triggers",
|
||||
description="Defines when frames should be sent to GenAI (on end, after updates, etc.).",
|
||||
)
|
||||
enabled_in_config: Optional[bool] = Field(
|
||||
default=None, title="Keep track of original state of generative AI."
|
||||
default=None,
|
||||
title="Original GenAI state",
|
||||
description="Indicates whether GenAI was enabled in the original static config.",
|
||||
)
|
||||
|
||||
@field_validator("required_zones", mode="before")
|
||||
@ -103,14 +129,25 @@ class GenAIObjectConfig(FrigateBaseModel):
|
||||
|
||||
|
||||
class ObjectConfig(FrigateBaseModel):
|
||||
track: list[str] = Field(default=DEFAULT_TRACKED_OBJECTS, title="Objects to track.")
|
||||
filters: dict[str, FilterConfig] = Field(
|
||||
default_factory=dict, title="Object filters."
|
||||
track: list[str] = Field(
|
||||
default=DEFAULT_TRACKED_OBJECTS,
|
||||
title="Objects to track",
|
||||
description="List of object labels to track for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
filters: dict[str, FilterConfig] = Field(
|
||||
default_factory=dict,
|
||||
title="Object filters",
|
||||
description="Filters applied to detected objects to reduce false positives (area, ratio, confidence).",
|
||||
)
|
||||
mask: Union[str, list[str]] = Field(
|
||||
default="",
|
||||
title="Object mask",
|
||||
description="Mask polygon used to prevent object detection in specified areas.",
|
||||
)
|
||||
mask: Union[str, list[str]] = Field(default="", title="Object mask.")
|
||||
genai: GenAIObjectConfig = Field(
|
||||
default_factory=GenAIObjectConfig,
|
||||
title="Config for using genai to analyze objects.",
|
||||
title="GenAI object config",
|
||||
description="GenAI options for describing tracked objects and sending frames for generation.",
|
||||
)
|
||||
_all_objects: list[str] = PrivateAttr()
|
||||
|
||||
|
||||
@ -17,37 +17,57 @@ class ZoomingModeEnum(str, Enum):
|
||||
|
||||
|
||||
class PtzAutotrackConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable PTZ object autotracking.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable Autotracking",
|
||||
description="Enable or disable automatic PTZ camera tracking of detected objects.",
|
||||
)
|
||||
calibrate_on_startup: bool = Field(
|
||||
default=False, title="Perform a camera calibration when Frigate starts."
|
||||
default=False,
|
||||
title="Calibrate on start",
|
||||
description="Measure PTZ motor speeds on startup to improve tracking accuracy. Frigate will update config with movement_weights after calibration.",
|
||||
)
|
||||
zooming: ZoomingModeEnum = Field(
|
||||
default=ZoomingModeEnum.disabled, title="Autotracker zooming mode."
|
||||
default=ZoomingModeEnum.disabled,
|
||||
title="Zoom mode",
|
||||
description="Control zoom behavior: disabled (pan/tilt only), absolute (most compatible), or relative (concurrent pan/tilt/zoom).",
|
||||
)
|
||||
zoom_factor: float = Field(
|
||||
default=0.3,
|
||||
title="Zooming factor (0.1-0.75).",
|
||||
title="Zoom factor",
|
||||
description="Control zoom level on tracked objects. Lower values keep more scene in view; higher values zoom in closer but may lose tracking. Values between 0.1 and 0.75.",
|
||||
ge=0.1,
|
||||
le=0.75,
|
||||
)
|
||||
track: list[str] = Field(default=DEFAULT_TRACKED_OBJECTS, title="Objects to track.")
|
||||
track: list[str] = Field(
|
||||
default=DEFAULT_TRACKED_OBJECTS,
|
||||
title="Tracked objects",
|
||||
description="List of object types that should trigger autotracking.",
|
||||
)
|
||||
required_zones: list[str] = Field(
|
||||
default_factory=list,
|
||||
title="List of required zones to be entered in order to begin autotracking.",
|
||||
title="Required zones",
|
||||
description="Objects must enter one of these zones before autotracking begins.",
|
||||
)
|
||||
return_preset: str = Field(
|
||||
default="home",
|
||||
title="Name of camera preset to return to when object tracking is over.",
|
||||
title="Return preset",
|
||||
description="ONVIF preset name configured in camera firmware to return to after tracking ends.",
|
||||
)
|
||||
timeout: int = Field(
|
||||
default=10, title="Seconds to delay before returning to preset."
|
||||
default=10,
|
||||
title="Return timeout",
|
||||
description="Wait this many seconds after losing tracking before returning camera to preset position.",
|
||||
)
|
||||
movement_weights: Optional[Union[str, list[str]]] = Field(
|
||||
default_factory=list,
|
||||
title="Internal value used for PTZ movements based on the speed of your camera's motor.",
|
||||
title="Movement weights",
|
||||
description="Calibration values automatically generated by camera calibration. Do not modify manually.",
|
||||
)
|
||||
enabled_in_config: Optional[bool] = Field(
|
||||
default=None, title="Keep track of original state of autotracking."
|
||||
default=None,
|
||||
title="Original autotrack state",
|
||||
description="Internal field to track whether autotracking was enabled in configuration.",
|
||||
)
|
||||
|
||||
@field_validator("movement_weights", mode="before")
|
||||
@ -72,16 +92,38 @@ class PtzAutotrackConfig(FrigateBaseModel):
|
||||
|
||||
|
||||
class OnvifConfig(FrigateBaseModel):
|
||||
host: str = Field(default="", title="Onvif Host")
|
||||
port: int = Field(default=8000, title="Onvif Port")
|
||||
user: Optional[EnvString] = Field(default=None, title="Onvif Username")
|
||||
password: Optional[EnvString] = Field(default=None, title="Onvif Password")
|
||||
tls_insecure: bool = Field(default=False, title="Onvif Disable TLS verification")
|
||||
host: str = Field(
|
||||
default="",
|
||||
title="ONVIF host",
|
||||
description="Host (and optional scheme) for the ONVIF service for this camera.",
|
||||
)
|
||||
port: int = Field(
|
||||
default=8000,
|
||||
title="ONVIF port",
|
||||
description="Port number for the ONVIF service.",
|
||||
)
|
||||
user: Optional[EnvString] = Field(
|
||||
default=None,
|
||||
title="ONVIF username",
|
||||
description="Username for ONVIF authentication; some devices require admin user for ONVIF.",
|
||||
)
|
||||
password: Optional[EnvString] = Field(
|
||||
default=None,
|
||||
title="ONVIF password",
|
||||
description="Password for ONVIF authentication.",
|
||||
)
|
||||
tls_insecure: bool = Field(
|
||||
default=False,
|
||||
title="Disable TLS verify",
|
||||
description="Skip TLS verification and disable digest auth for ONVIF (unsafe; use in safe networks only).",
|
||||
)
|
||||
autotracking: PtzAutotrackConfig = Field(
|
||||
default_factory=PtzAutotrackConfig,
|
||||
title="PTZ auto tracking config.",
|
||||
title="Autotracking",
|
||||
description="Automatically track moving objects and keep them centered in the frame using PTZ camera movements.",
|
||||
)
|
||||
ignore_time_mismatch: bool = Field(
|
||||
default=False,
|
||||
title="Onvif Ignore Time Synchronization Mismatch Between Camera and Server",
|
||||
title="Ignore time mismatch",
|
||||
description="Ignore time synchronization differences between camera and Frigate server for ONVIF communication.",
|
||||
)
|
||||
|
||||
@ -21,7 +21,12 @@ __all__ = [
|
||||
|
||||
|
||||
class RecordRetainConfig(FrigateBaseModel):
|
||||
days: float = Field(default=0, ge=0, title="Default retention period.")
|
||||
days: float = Field(
|
||||
default=0,
|
||||
ge=0,
|
||||
title="Retention days",
|
||||
description="Days to retain recordings.",
|
||||
)
|
||||
|
||||
|
||||
class RetainModeEnum(str, Enum):
|
||||
@ -31,22 +36,37 @@ class RetainModeEnum(str, Enum):
|
||||
|
||||
|
||||
class ReviewRetainConfig(FrigateBaseModel):
|
||||
days: float = Field(default=10, ge=0, title="Default retention period.")
|
||||
mode: RetainModeEnum = Field(default=RetainModeEnum.motion, title="Retain mode.")
|
||||
days: float = Field(
|
||||
default=10,
|
||||
ge=0,
|
||||
title="Retention days",
|
||||
description="Number of days to retain recordings of detection events.",
|
||||
)
|
||||
mode: RetainModeEnum = Field(
|
||||
default=RetainModeEnum.motion,
|
||||
title="Retention mode",
|
||||
description="Mode for retention: all (save all segments), motion (save segments with motion), or active_objects (save segments with active objects).",
|
||||
)
|
||||
|
||||
|
||||
class EventsConfig(FrigateBaseModel):
|
||||
pre_capture: int = Field(
|
||||
default=5,
|
||||
title="Seconds to retain before event starts.",
|
||||
title="Pre-capture seconds",
|
||||
description="Number of seconds before the detection event to include in the recording.",
|
||||
le=MAX_PRE_CAPTURE,
|
||||
ge=0,
|
||||
)
|
||||
post_capture: int = Field(
|
||||
default=5, ge=0, title="Seconds to retain after event ends."
|
||||
default=5,
|
||||
ge=0,
|
||||
title="Post-capture seconds",
|
||||
description="Number of seconds after the detection event to include in the recording.",
|
||||
)
|
||||
retain: ReviewRetainConfig = Field(
|
||||
default_factory=ReviewRetainConfig, title="Event retention settings."
|
||||
default_factory=ReviewRetainConfig,
|
||||
title="Event retention",
|
||||
description="Retention settings for recordings of detection events.",
|
||||
)
|
||||
|
||||
|
||||
@ -60,43 +80,65 @@ class RecordQualityEnum(str, Enum):
|
||||
|
||||
class RecordPreviewConfig(FrigateBaseModel):
|
||||
quality: RecordQualityEnum = Field(
|
||||
default=RecordQualityEnum.medium, title="Quality of recording preview."
|
||||
default=RecordQualityEnum.medium,
|
||||
title="Preview quality",
|
||||
description="Preview quality level (very_low, low, medium, high, very_high).",
|
||||
)
|
||||
|
||||
|
||||
class RecordExportConfig(FrigateBaseModel):
|
||||
hwaccel_args: Union[str, list[str]] = Field(
|
||||
default="auto", title="Export-specific FFmpeg hardware acceleration arguments."
|
||||
default="auto",
|
||||
title="Export hwaccel args",
|
||||
description="Hardware acceleration args to use for export/transcode operations.",
|
||||
)
|
||||
|
||||
|
||||
class RecordConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable record on all cameras.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable recording",
|
||||
description="Enable or disable recording for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
expire_interval: int = Field(
|
||||
default=60,
|
||||
title="Number of minutes to wait between cleanup runs.",
|
||||
title="Record cleanup interval",
|
||||
description="Minutes between cleanup passes that remove expired recording segments.",
|
||||
)
|
||||
continuous: RecordRetainConfig = Field(
|
||||
default_factory=RecordRetainConfig,
|
||||
title="Continuous recording retention settings.",
|
||||
title="Continuous retention",
|
||||
description="Number of days to retain recordings regardless of tracked objects or motion. Set to 0 if you only want to retain recordings of alerts and detections.",
|
||||
)
|
||||
motion: RecordRetainConfig = Field(
|
||||
default_factory=RecordRetainConfig, title="Motion recording retention settings."
|
||||
default_factory=RecordRetainConfig,
|
||||
title="Motion retention",
|
||||
description="Number of days to retain recordings triggered by motion regardless of tracked objects. Set to 0 if you only want to retain recordings of alerts and detections.",
|
||||
)
|
||||
detections: EventsConfig = Field(
|
||||
default_factory=EventsConfig, title="Detection specific retention settings."
|
||||
default_factory=EventsConfig,
|
||||
title="Detection retention",
|
||||
description="Recording retention settings for detection events including pre/post capture durations.",
|
||||
)
|
||||
alerts: EventsConfig = Field(
|
||||
default_factory=EventsConfig, title="Alert specific retention settings."
|
||||
default_factory=EventsConfig,
|
||||
title="Alert retention",
|
||||
description="Recording retention settings for alert events including pre/post capture durations.",
|
||||
)
|
||||
export: RecordExportConfig = Field(
|
||||
default_factory=RecordExportConfig, title="Recording Export Config"
|
||||
default_factory=RecordExportConfig,
|
||||
title="Export config",
|
||||
description="Settings used when exporting recordings such as timelapse and hardware acceleration.",
|
||||
)
|
||||
preview: RecordPreviewConfig = Field(
|
||||
default_factory=RecordPreviewConfig, title="Recording Preview Config"
|
||||
default_factory=RecordPreviewConfig,
|
||||
title="Preview config",
|
||||
description="Settings controlling the quality of recording previews shown in the UI.",
|
||||
)
|
||||
enabled_in_config: Optional[bool] = Field(
|
||||
default=None, title="Keep track of original state of recording."
|
||||
default=None,
|
||||
title="Original recording state",
|
||||
description="Indicates whether recording was enabled in the original static configuration.",
|
||||
)
|
||||
|
||||
@property
|
||||
|
||||
@ -21,22 +21,32 @@ DEFAULT_ALERT_OBJECTS = ["person", "car"]
|
||||
class AlertsConfig(FrigateBaseModel):
|
||||
"""Configure alerts"""
|
||||
|
||||
enabled: bool = Field(default=True, title="Enable alerts.")
|
||||
enabled: bool = Field(
|
||||
default=True,
|
||||
title="Enable alerts",
|
||||
description="Enable or disable alert generation for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
|
||||
labels: list[str] = Field(
|
||||
default=DEFAULT_ALERT_OBJECTS, title="Labels to create alerts for."
|
||||
default=DEFAULT_ALERT_OBJECTS,
|
||||
title="Alert labels",
|
||||
description="List of object labels that qualify as alerts (for example: car, person).",
|
||||
)
|
||||
required_zones: Union[str, list[str]] = Field(
|
||||
default_factory=list,
|
||||
title="List of required zones to be entered in order to save the event as an alert.",
|
||||
title="Required zones",
|
||||
description="Zones that an object must enter to be considered an alert; leave empty to allow any zone.",
|
||||
)
|
||||
|
||||
enabled_in_config: Optional[bool] = Field(
|
||||
default=None, title="Keep track of original state of alerts."
|
||||
default=None,
|
||||
title="Original alerts state",
|
||||
description="Tracks whether alerts were originally enabled in the static configuration.",
|
||||
)
|
||||
cutoff_time: int = Field(
|
||||
default=40,
|
||||
title="Time to cutoff alerts after no alert-causing activity has occurred.",
|
||||
title="Alerts cutoff time",
|
||||
description="Seconds to wait after no alert-causing activity before cutting off an alert.",
|
||||
)
|
||||
|
||||
@field_validator("required_zones", mode="before")
|
||||
@ -51,22 +61,32 @@ class AlertsConfig(FrigateBaseModel):
|
||||
class DetectionsConfig(FrigateBaseModel):
|
||||
"""Configure detections"""
|
||||
|
||||
enabled: bool = Field(default=True, title="Enable detections.")
|
||||
enabled: bool = Field(
|
||||
default=True,
|
||||
title="Enable detections",
|
||||
description="Enable or disable detection events for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
|
||||
labels: Optional[list[str]] = Field(
|
||||
default=None, title="Labels to create detections for."
|
||||
default=None,
|
||||
title="Detection labels",
|
||||
description="List of object labels that qualify as detection events.",
|
||||
)
|
||||
required_zones: Union[str, list[str]] = Field(
|
||||
default_factory=list,
|
||||
title="List of required zones to be entered in order to save the event as a detection.",
|
||||
title="Required zones",
|
||||
description="Zones that an object must enter to be considered a detection; leave empty to allow any zone.",
|
||||
)
|
||||
cutoff_time: int = Field(
|
||||
default=30,
|
||||
title="Time to cutoff detection after no detection-causing activity has occurred.",
|
||||
title="Detections cutoff time",
|
||||
description="Seconds to wait after no detection-causing activity before cutting off a detection.",
|
||||
)
|
||||
|
||||
enabled_in_config: Optional[bool] = Field(
|
||||
default=None, title="Keep track of original state of detections."
|
||||
default=None,
|
||||
title="Original detections state",
|
||||
description="Tracks whether detections were originally enabled in the static configuration.",
|
||||
)
|
||||
|
||||
@field_validator("required_zones", mode="before")
|
||||
@ -81,27 +101,42 @@ class DetectionsConfig(FrigateBaseModel):
|
||||
class GenAIReviewConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable GenAI descriptions for review items.",
|
||||
title="Enable GenAI descriptions",
|
||||
description="Enable or disable GenAI-generated descriptions and summaries for review items.",
|
||||
)
|
||||
alerts: bool = Field(
|
||||
default=True,
|
||||
title="Enable GenAI for alerts",
|
||||
description="Use GenAI to generate descriptions for alert items.",
|
||||
)
|
||||
detections: bool = Field(
|
||||
default=False,
|
||||
title="Enable GenAI for detections",
|
||||
description="Use GenAI to generate descriptions for detection items.",
|
||||
)
|
||||
alerts: bool = Field(default=True, title="Enable GenAI for alerts.")
|
||||
detections: bool = Field(default=False, title="Enable GenAI for detections.")
|
||||
image_source: ImageSourceEnum = Field(
|
||||
default=ImageSourceEnum.preview,
|
||||
title="Image source for review descriptions.",
|
||||
title="Review image source",
|
||||
description="Source of images sent to GenAI ('preview' or 'recordings'); 'recordings' uses higher quality frames but more tokens.",
|
||||
)
|
||||
additional_concerns: list[str] = Field(
|
||||
default=[],
|
||||
title="Additional concerns that GenAI should make note of on this camera.",
|
||||
title="Additional concerns",
|
||||
description="A list of additional concerns or notes the GenAI should consider when evaluating activity on this camera.",
|
||||
)
|
||||
debug_save_thumbnails: bool = Field(
|
||||
default=False,
|
||||
title="Save thumbnails sent to generative AI for debugging purposes.",
|
||||
title="Save thumbnails",
|
||||
description="Save thumbnails that are sent to the GenAI provider for debugging and review.",
|
||||
)
|
||||
enabled_in_config: Optional[bool] = Field(
|
||||
default=None, title="Keep track of original state of generative AI."
|
||||
default=None,
|
||||
title="Original GenAI state",
|
||||
description="Tracks whether GenAI review was originally enabled in the static configuration.",
|
||||
)
|
||||
preferred_language: str | None = Field(
|
||||
title="Preferred language for GenAI Response",
|
||||
title="Preferred language",
|
||||
description="Preferred language to request from the GenAI provider for generated responses.",
|
||||
default=None,
|
||||
)
|
||||
activity_context_prompt: str = Field(
|
||||
@ -139,19 +174,24 @@ Evaluate in this order:
|
||||
3. **Escalate to Level 2 if:** Weapons, break-in tools, forced entry in progress, violence, or active property damage visible (escalates from Level 0 or 1)
|
||||
|
||||
The mere presence of an unidentified person in private areas during late night hours is inherently suspicious and warrants human review, regardless of what activity they appear to be doing or how brief the sequence is.""",
|
||||
title="Custom activity context prompt defining normal and suspicious activity patterns for this property.",
|
||||
title="Activity context prompt",
|
||||
description="Custom prompt describing what is and is not suspicious activity to provide context for GenAI summaries.",
|
||||
)
|
||||
|
||||
|
||||
class ReviewConfig(FrigateBaseModel):
|
||||
"""Configure reviews"""
|
||||
|
||||
alerts: AlertsConfig = Field(
|
||||
default_factory=AlertsConfig, title="Review alerts config."
|
||||
default_factory=AlertsConfig,
|
||||
title="Alerts config",
|
||||
description="Settings for which tracked objects generate alerts and how alerts are retained.",
|
||||
)
|
||||
detections: DetectionsConfig = Field(
|
||||
default_factory=DetectionsConfig, title="Review detections config."
|
||||
default_factory=DetectionsConfig,
|
||||
title="Detections config",
|
||||
description="Settings for creating detection events (non-alert) and how long to keep them.",
|
||||
)
|
||||
genai: GenAIReviewConfig = Field(
|
||||
default_factory=GenAIReviewConfig, title="Review description genai config."
|
||||
default_factory=GenAIReviewConfig,
|
||||
title="GenAI config",
|
||||
description="Controls use of generative AI for producing descriptions and summaries of review items.",
|
||||
)
|
||||
|
||||
@ -9,36 +9,68 @@ __all__ = ["SnapshotsConfig", "RetainConfig"]
|
||||
|
||||
|
||||
class RetainConfig(FrigateBaseModel):
|
||||
default: float = Field(default=10, title="Default retention period.")
|
||||
mode: RetainModeEnum = Field(default=RetainModeEnum.motion, title="Retain mode.")
|
||||
default: float = Field(
|
||||
default=10,
|
||||
title="Default retention",
|
||||
description="Default number of days to retain snapshots.",
|
||||
)
|
||||
mode: RetainModeEnum = Field(
|
||||
default=RetainModeEnum.motion,
|
||||
title="Retention mode",
|
||||
description="Mode for retention: all (save all segments), motion (save segments with motion), or active_objects (save segments with active objects).",
|
||||
)
|
||||
objects: dict[str, float] = Field(
|
||||
default_factory=dict, title="Object retention period."
|
||||
default_factory=dict,
|
||||
title="Object retention",
|
||||
description="Per-object overrides for snapshot retention days.",
|
||||
)
|
||||
|
||||
|
||||
class SnapshotsConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Snapshots enabled.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Snapshots enabled",
|
||||
description="Enable or disable saving snapshots for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
clean_copy: bool = Field(
|
||||
default=True, title="Create a clean copy of the snapshot image."
|
||||
default=True,
|
||||
title="Save clean copy",
|
||||
description="Save an unannotated clean copy of snapshots in addition to annotated ones.",
|
||||
)
|
||||
timestamp: bool = Field(
|
||||
default=False, title="Add a timestamp overlay on the snapshot."
|
||||
default=False,
|
||||
title="Timestamp overlay",
|
||||
description="Overlay a timestamp on saved snapshots.",
|
||||
)
|
||||
bounding_box: bool = Field(
|
||||
default=True, title="Add a bounding box overlay on the snapshot."
|
||||
default=True,
|
||||
title="Bounding box overlay",
|
||||
description="Draw bounding boxes for tracked objects on saved snapshots.",
|
||||
)
|
||||
crop: bool = Field(
|
||||
default=False,
|
||||
title="Crop snapshot",
|
||||
description="Crop saved snapshots to the detected object's bounding box.",
|
||||
)
|
||||
crop: bool = Field(default=False, title="Crop the snapshot to the detected object.")
|
||||
required_zones: list[str] = Field(
|
||||
default_factory=list,
|
||||
title="List of required zones to be entered in order to save a snapshot.",
|
||||
title="Required zones",
|
||||
description="Zones an object must enter for a snapshot to be saved.",
|
||||
)
|
||||
height: Optional[int] = Field(
|
||||
default=None,
|
||||
title="Snapshot height",
|
||||
description="Height (pixels) to resize saved snapshots to; leave empty to preserve original size.",
|
||||
)
|
||||
height: Optional[int] = Field(default=None, title="Snapshot image height.")
|
||||
retain: RetainConfig = Field(
|
||||
default_factory=RetainConfig, title="Snapshot retention."
|
||||
default_factory=RetainConfig,
|
||||
title="Snapshot retention",
|
||||
description="Retention settings for saved snapshots including default days and per-object overrides.",
|
||||
)
|
||||
quality: int = Field(
|
||||
default=70,
|
||||
title="Quality of the encoded jpeg (0-100).",
|
||||
title="JPEG quality",
|
||||
description="JPEG encode quality for saved snapshots (0-100).",
|
||||
ge=0,
|
||||
le=100,
|
||||
)
|
||||
|
||||
@ -27,9 +27,27 @@ class TimestampPositionEnum(str, Enum):
|
||||
|
||||
|
||||
class ColorConfig(FrigateBaseModel):
|
||||
red: int = Field(default=255, ge=0, le=255, title="Red")
|
||||
green: int = Field(default=255, ge=0, le=255, title="Green")
|
||||
blue: int = Field(default=255, ge=0, le=255, title="Blue")
|
||||
red: int = Field(
|
||||
default=255,
|
||||
ge=0,
|
||||
le=255,
|
||||
title="Red",
|
||||
description="Red component (0-255) for timestamp color.",
|
||||
)
|
||||
green: int = Field(
|
||||
default=255,
|
||||
ge=0,
|
||||
le=255,
|
||||
title="Green",
|
||||
description="Green component (0-255) for timestamp color.",
|
||||
)
|
||||
blue: int = Field(
|
||||
default=255,
|
||||
ge=0,
|
||||
le=255,
|
||||
title="Blue",
|
||||
description="Blue component (0-255) for timestamp color.",
|
||||
)
|
||||
|
||||
|
||||
class TimestampEffectEnum(str, Enum):
|
||||
@ -39,11 +57,27 @@ class TimestampEffectEnum(str, Enum):
|
||||
|
||||
class TimestampStyleConfig(FrigateBaseModel):
|
||||
position: TimestampPositionEnum = Field(
|
||||
default=TimestampPositionEnum.tl, title="Timestamp position."
|
||||
default=TimestampPositionEnum.tl,
|
||||
title="Timestamp position",
|
||||
description="Position of the timestamp on the image (tl/tr/bl/br).",
|
||||
)
|
||||
format: str = Field(
|
||||
default=DEFAULT_TIME_FORMAT,
|
||||
title="Timestamp format",
|
||||
description="Datetime format string used for timestamps (Python datetime format codes).",
|
||||
)
|
||||
color: ColorConfig = Field(
|
||||
default_factory=ColorConfig,
|
||||
title="Timestamp color",
|
||||
description="RGB color values for the timestamp text (all values 0-255).",
|
||||
)
|
||||
thickness: int = Field(
|
||||
default=2,
|
||||
title="Timestamp thickness",
|
||||
description="Line thickness of the timestamp text.",
|
||||
)
|
||||
format: str = Field(default=DEFAULT_TIME_FORMAT, title="Timestamp format.")
|
||||
color: ColorConfig = Field(default_factory=ColorConfig, title="Timestamp color.")
|
||||
thickness: int = Field(default=2, title="Timestamp thickness.")
|
||||
effect: Optional[TimestampEffectEnum] = Field(
|
||||
default=None, title="Timestamp effect."
|
||||
default=None,
|
||||
title="Timestamp effect",
|
||||
description="Visual effect for the timestamp text (none, solid, shadow).",
|
||||
)
|
||||
|
||||
@ -6,7 +6,13 @@ __all__ = ["CameraUiConfig"]
|
||||
|
||||
|
||||
class CameraUiConfig(FrigateBaseModel):
|
||||
order: int = Field(default=0, title="Order of camera in UI.")
|
||||
dashboard: bool = Field(
|
||||
default=True, title="Show this camera in Frigate dashboard UI."
|
||||
order: int = Field(
|
||||
default=0,
|
||||
title="UI order",
|
||||
description="Numeric order used to sort the camera in the UI (default dashboard and lists); larger numbers appear later.",
|
||||
)
|
||||
dashboard: bool = Field(
|
||||
default=True,
|
||||
title="Show in UI",
|
||||
description="Toggle whether this camera is visible everywhere in the Frigate UI. Disabling this will require manually editing the config to view this camera in the UI again.",
|
||||
)
|
||||
|
||||
@ -14,36 +14,46 @@ logger = logging.getLogger(__name__)
|
||||
|
||||
class ZoneConfig(BaseModel):
|
||||
friendly_name: Optional[str] = Field(
|
||||
None, title="Zone friendly name used in the Frigate UI."
|
||||
None,
|
||||
title="Zone name",
|
||||
description="A user-friendly name for the zone, displayed in the Frigate UI. If not set, a formatted version of the zone name will be used.",
|
||||
)
|
||||
filters: dict[str, FilterConfig] = Field(
|
||||
default_factory=dict, title="Zone filters."
|
||||
default_factory=dict,
|
||||
title="Zone filters",
|
||||
description="Filters to apply to objects within this zone. Used to reduce false positives or restrict which objects are considered present in the zone.",
|
||||
)
|
||||
coordinates: Union[str, list[str]] = Field(
|
||||
title="Coordinates polygon for the defined zone."
|
||||
title="Coordinates",
|
||||
description="Polygon coordinates that define the zone area. Can be a comma-separated string or a list of coordinate strings. Coordinates should be relative (0-1) or absolute (legacy).",
|
||||
)
|
||||
distances: Optional[Union[str, list[str]]] = Field(
|
||||
default_factory=list,
|
||||
title="Real-world distances for the sides of quadrilateral for the defined zone.",
|
||||
title="Real-world distances",
|
||||
description="Optional real-world distances for each side of the zone quadrilateral, used for speed or distance calculations. Must have exactly 4 values if set.",
|
||||
)
|
||||
inertia: int = Field(
|
||||
default=3,
|
||||
title="Number of consecutive frames required for object to be considered present in the zone.",
|
||||
title="Inertia frames",
|
||||
gt=0,
|
||||
description="Number of consecutive frames an object must be detected in the zone before it is considered present. Helps filter out transient detections.",
|
||||
)
|
||||
loitering_time: int = Field(
|
||||
default=0,
|
||||
ge=0,
|
||||
title="Number of seconds that an object must loiter to be considered in the zone.",
|
||||
title="Loitering seconds",
|
||||
description="Number of seconds an object must remain in the zone to be considered as loitering. Set to 0 to disable loitering detection.",
|
||||
)
|
||||
speed_threshold: Optional[float] = Field(
|
||||
default=None,
|
||||
ge=0.1,
|
||||
title="Minimum speed value for an object to be considered in the zone.",
|
||||
title="Minimum speed",
|
||||
description="Minimum speed (in real-world units if distances are set) required for an object to be considered present in the zone. Used for speed-based zone triggers.",
|
||||
)
|
||||
objects: Union[str, list[str]] = Field(
|
||||
default_factory=list,
|
||||
title="List of objects that can trigger the zone.",
|
||||
title="Trigger objects",
|
||||
description="List of object types (from labelmap) that can trigger this zone. Can be a string or a list of strings. If empty, all objects are considered.",
|
||||
)
|
||||
_color: Optional[tuple[int, int, int]] = PrivateAttr()
|
||||
_contour: np.ndarray = PrivateAttr()
|
||||
|
||||
@ -8,13 +8,21 @@ __all__ = ["CameraGroupConfig"]
|
||||
|
||||
|
||||
class CameraGroupConfig(FrigateBaseModel):
|
||||
"""Represents a group of cameras."""
|
||||
|
||||
cameras: Union[str, list[str]] = Field(
|
||||
default_factory=list, title="List of cameras in this group."
|
||||
default_factory=list,
|
||||
title="Camera list",
|
||||
description="Array of camera names included in this group.",
|
||||
)
|
||||
icon: str = Field(
|
||||
default="generic",
|
||||
title="Group icon",
|
||||
description="Icon used to represent the camera group in the UI.",
|
||||
)
|
||||
order: int = Field(
|
||||
default=0,
|
||||
title="Sort order",
|
||||
description="Numeric order used to sort camera groups in the UI; larger numbers appear later.",
|
||||
)
|
||||
icon: str = Field(default="generic", title="Icon that represents camera group.")
|
||||
order: int = Field(default=0, title="Sort order for group.")
|
||||
|
||||
@field_validator("cameras", mode="before")
|
||||
@classmethod
|
||||
|
||||
@ -43,28 +43,43 @@ class ObjectClassificationType(str, Enum):
|
||||
|
||||
|
||||
class AudioTranscriptionConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable audio transcription.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable audio transcription",
|
||||
description="Enable or disable automatic audio transcription for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
language: str = Field(
|
||||
default="en",
|
||||
title="Language abbreviation to use for audio event transcription/translation.",
|
||||
title="Transcription language",
|
||||
description="Language code used for transcription/translation (for example 'en' for English). See https://whisper-api.com/docs/languages/ for supported language codes.",
|
||||
)
|
||||
device: Optional[EnrichmentsDeviceEnum] = Field(
|
||||
default=EnrichmentsDeviceEnum.CPU,
|
||||
title="The device used for audio transcription.",
|
||||
title="Transcription device",
|
||||
description="Device key (CPU/GPU) to run the transcription model on. Only NVIDIA CUDA GPUs are currently supported for transcription.",
|
||||
)
|
||||
model_size: str = Field(
|
||||
default="small", title="The size of the embeddings model used."
|
||||
default="small",
|
||||
title="Model size",
|
||||
description="Model size to use for offline audio event transcription.",
|
||||
)
|
||||
live_enabled: Optional[bool] = Field(
|
||||
default=False, title="Enable live transcriptions."
|
||||
default=False,
|
||||
title="Live transcription",
|
||||
description="Enable streaming live transcription for audio as it is received.",
|
||||
)
|
||||
|
||||
|
||||
class BirdClassificationConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable bird classification.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Bird classification",
|
||||
description="Enable or disable bird classification.",
|
||||
)
|
||||
threshold: float = Field(
|
||||
default=0.9,
|
||||
title="Minimum classification score required to be considered a match.",
|
||||
title="Minimum score",
|
||||
description="Minimum classification score required to accept a bird classification.",
|
||||
gt=0.0,
|
||||
le=1.0,
|
||||
)
|
||||
@ -72,42 +87,62 @@ class BirdClassificationConfig(FrigateBaseModel):
|
||||
|
||||
class CustomClassificationStateCameraConfig(FrigateBaseModel):
|
||||
crop: list[float, float, float, float] = Field(
|
||||
title="Crop of image frame on this camera to run classification on."
|
||||
title="Classification crop",
|
||||
description="Crop coordinates to use for running classification on this camera.",
|
||||
)
|
||||
|
||||
|
||||
class CustomClassificationStateConfig(FrigateBaseModel):
|
||||
cameras: Dict[str, CustomClassificationStateCameraConfig] = Field(
|
||||
title="Cameras to run classification on."
|
||||
title="Classification cameras",
|
||||
description="Per-camera crop and settings for running state classification.",
|
||||
)
|
||||
motion: bool = Field(
|
||||
default=False,
|
||||
title="If classification should be run when motion is detected in the crop.",
|
||||
title="Run on motion",
|
||||
description="If true, run classification when motion is detected within the specified crop.",
|
||||
)
|
||||
interval: int | None = Field(
|
||||
default=None,
|
||||
title="Interval to run classification on in seconds.",
|
||||
title="Classification interval",
|
||||
description="Interval (seconds) between periodic classification runs for state classification.",
|
||||
gt=0,
|
||||
)
|
||||
|
||||
|
||||
class CustomClassificationObjectConfig(FrigateBaseModel):
|
||||
objects: list[str] = Field(title="Object types to classify.")
|
||||
objects: list[str] = Field(
|
||||
default_factory=list,
|
||||
title="Classify objects",
|
||||
description="List of object types to run object classification on.",
|
||||
)
|
||||
classification_type: ObjectClassificationType = Field(
|
||||
default=ObjectClassificationType.sub_label,
|
||||
title="Type of classification that is applied.",
|
||||
title="Classification type",
|
||||
description="Classification type applied: 'sub_label' (adds sub_label) or other supported types.",
|
||||
)
|
||||
|
||||
|
||||
class CustomClassificationConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=True, title="Enable running the model.")
|
||||
name: str | None = Field(default=None, title="Name of classification model.")
|
||||
enabled: bool = Field(
|
||||
default=True,
|
||||
title="Enable model",
|
||||
description="Enable or disable the custom classification model.",
|
||||
)
|
||||
name: str | None = Field(
|
||||
default=None,
|
||||
title="Model name",
|
||||
description="Identifier for the custom classification model to use.",
|
||||
)
|
||||
threshold: float = Field(
|
||||
default=0.8, title="Classification score threshold to change the state."
|
||||
default=0.8,
|
||||
title="Score threshold",
|
||||
description="Score threshold used to change the classification state.",
|
||||
)
|
||||
save_attempts: int | None = Field(
|
||||
default=None,
|
||||
title="Number of classification attempts to save in the recent classifications tab. If not specified, defaults to 200 for object classification and 100 for state classification.",
|
||||
title="Save attempts",
|
||||
description="How many classification attempts to save for recent classifications UI.",
|
||||
ge=0,
|
||||
)
|
||||
object_config: CustomClassificationObjectConfig | None = Field(default=None)
|
||||
@ -116,47 +151,76 @@ class CustomClassificationConfig(FrigateBaseModel):
|
||||
|
||||
class ClassificationConfig(FrigateBaseModel):
|
||||
bird: BirdClassificationConfig = Field(
|
||||
default_factory=BirdClassificationConfig, title="Bird classification config."
|
||||
default_factory=BirdClassificationConfig,
|
||||
title="Bird classification config",
|
||||
description="Settings specific to bird classification models.",
|
||||
)
|
||||
custom: Dict[str, CustomClassificationConfig] = Field(
|
||||
default={}, title="Custom Classification Model Configs."
|
||||
default={},
|
||||
title="Custom Classification Models",
|
||||
description="Configuration for custom classification models used for objects or state detection.",
|
||||
)
|
||||
|
||||
|
||||
class SemanticSearchConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable semantic search.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable semantic search",
|
||||
description="Enable or disable the semantic search feature.",
|
||||
)
|
||||
reindex: Optional[bool] = Field(
|
||||
default=False, title="Reindex all tracked objects on startup."
|
||||
default=False,
|
||||
title="Reindex on startup",
|
||||
description="Trigger a full reindex of historical tracked objects into the embeddings database.",
|
||||
)
|
||||
model: Optional[SemanticSearchModelEnum] = Field(
|
||||
default=SemanticSearchModelEnum.jinav1,
|
||||
title="The CLIP model to use for semantic search.",
|
||||
title="Semantic search model",
|
||||
description="The embeddings model to use for semantic search (for example 'jinav1').",
|
||||
)
|
||||
model_size: str = Field(
|
||||
default="small", title="The size of the embeddings model used."
|
||||
default="small",
|
||||
title="Model size",
|
||||
description="Select model size; 'small' runs on CPU and 'large' typically requires GPU.",
|
||||
)
|
||||
device: Optional[str] = Field(
|
||||
default=None,
|
||||
title="The device key to use for semantic search.",
|
||||
title="Device",
|
||||
description="This is an override, to target a specific device. See https://onnxruntime.ai/docs/execution-providers/ for more information",
|
||||
)
|
||||
|
||||
|
||||
class TriggerConfig(FrigateBaseModel):
|
||||
friendly_name: Optional[str] = Field(
|
||||
None, title="Trigger friendly name used in the Frigate UI."
|
||||
None,
|
||||
title="Friendly name",
|
||||
description="Optional friendly name displayed in the UI for this trigger.",
|
||||
)
|
||||
enabled: bool = Field(
|
||||
default=True,
|
||||
title="Enable this trigger",
|
||||
description="Enable or disable this semantic search trigger.",
|
||||
)
|
||||
type: TriggerType = Field(
|
||||
default=TriggerType.DESCRIPTION,
|
||||
title="Trigger type",
|
||||
description="Type of trigger: 'thumbnail' (match against image) or 'description' (match against text).",
|
||||
)
|
||||
data: str = Field(
|
||||
title="Trigger content",
|
||||
description="Text phrase or thumbnail ID to match against tracked objects.",
|
||||
)
|
||||
enabled: bool = Field(default=True, title="Enable this trigger")
|
||||
type: TriggerType = Field(default=TriggerType.DESCRIPTION, title="Type of trigger")
|
||||
data: str = Field(title="Trigger content (text phrase or image ID)")
|
||||
threshold: float = Field(
|
||||
title="Confidence score required to run the trigger",
|
||||
title="Trigger threshold",
|
||||
description="Minimum similarity score (0-1) required to activate this trigger.",
|
||||
default=0.8,
|
||||
gt=0.0,
|
||||
le=1.0,
|
||||
)
|
||||
actions: List[TriggerAction] = Field(
|
||||
default=[], title="Actions to perform when trigger is matched"
|
||||
default=[],
|
||||
title="Trigger actions",
|
||||
description="List of actions to execute when trigger matches (notification, sub_label, attribute).",
|
||||
)
|
||||
|
||||
model_config = ConfigDict(extra="forbid", protected_namespaces=())
|
||||
@ -165,147 +229,191 @@ class TriggerConfig(FrigateBaseModel):
|
||||
class CameraSemanticSearchConfig(FrigateBaseModel):
|
||||
triggers: Dict[str, TriggerConfig] = Field(
|
||||
default={},
|
||||
title="Trigger actions on tracked objects that match existing thumbnails or descriptions",
|
||||
title="Triggers",
|
||||
description="Actions and matching criteria for camera-specific semantic search triggers.",
|
||||
)
|
||||
|
||||
model_config = ConfigDict(extra="forbid", protected_namespaces=())
|
||||
|
||||
|
||||
class FaceRecognitionConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable face recognition.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable face recognition",
|
||||
description="Enable or disable face recognition for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
model_size: str = Field(
|
||||
default="small", title="The size of the embeddings model used."
|
||||
default="small",
|
||||
title="Model size",
|
||||
description="Model size to use for face embeddings (small/large); larger may require GPU.",
|
||||
)
|
||||
unknown_score: float = Field(
|
||||
title="Minimum face distance score required to be marked as a potential match.",
|
||||
title="Unknown score threshold",
|
||||
description="Distance threshold below which a face is considered a potential match (higher = stricter).",
|
||||
default=0.8,
|
||||
gt=0.0,
|
||||
le=1.0,
|
||||
)
|
||||
detection_threshold: float = Field(
|
||||
default=0.7,
|
||||
title="Minimum face detection score required to be considered a face.",
|
||||
title="Detection threshold",
|
||||
description="Minimum detection confidence required to consider a face detection valid.",
|
||||
gt=0.0,
|
||||
le=1.0,
|
||||
)
|
||||
recognition_threshold: float = Field(
|
||||
default=0.9,
|
||||
title="Minimum face distance score required to be considered a match.",
|
||||
title="Recognition threshold",
|
||||
description="Face embedding distance threshold to consider two faces a match.",
|
||||
gt=0.0,
|
||||
le=1.0,
|
||||
)
|
||||
min_area: int = Field(
|
||||
default=750, title="Min area of face box to consider running face recognition."
|
||||
default=750,
|
||||
title="Minimum face area",
|
||||
description="Minimum area (pixels) of a detected face box required to attempt recognition.",
|
||||
)
|
||||
min_faces: int = Field(
|
||||
default=1,
|
||||
gt=0,
|
||||
le=6,
|
||||
title="Min face recognitions for the sub label to be applied to the person object.",
|
||||
title="Minimum faces",
|
||||
description="Minimum number of face recognitions required before applying a recognized sub-label to a person.",
|
||||
)
|
||||
save_attempts: int = Field(
|
||||
default=200,
|
||||
ge=0,
|
||||
title="Number of face attempts to save in the recent recognitions tab.",
|
||||
title="Save attempts",
|
||||
description="Number of face recognition attempts to retain for recent recognition UI.",
|
||||
)
|
||||
blur_confidence_filter: bool = Field(
|
||||
default=True, title="Apply blur quality filter to face confidence."
|
||||
default=True,
|
||||
title="Blur confidence filter",
|
||||
description="Adjust confidence scores based on image blur to reduce false positives for poor quality faces.",
|
||||
)
|
||||
device: Optional[str] = Field(
|
||||
default=None,
|
||||
title="The device key to use for face recognition.",
|
||||
title="Device",
|
||||
description="This is an override, to target a specific device. See https://onnxruntime.ai/docs/execution-providers/ for more information",
|
||||
)
|
||||
|
||||
|
||||
class CameraFaceRecognitionConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable face recognition.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable face recognition",
|
||||
description="Enable or disable face recognition.",
|
||||
)
|
||||
min_area: int = Field(
|
||||
default=750, title="Min area of face box to consider running face recognition."
|
||||
default=750,
|
||||
title="Minimum face area",
|
||||
description="Minimum area (pixels) of a detected face box required to attempt recognition.",
|
||||
)
|
||||
|
||||
model_config = ConfigDict(extra="forbid", protected_namespaces=())
|
||||
|
||||
|
||||
class ReplaceRule(FrigateBaseModel):
|
||||
pattern: str = Field(..., title="Regex pattern to match.")
|
||||
replacement: str = Field(
|
||||
..., title="Replacement string (supports backrefs like '\\1')."
|
||||
)
|
||||
pattern: str = Field(..., title="Regex pattern")
|
||||
replacement: str = Field(..., title="Replacement string")
|
||||
|
||||
|
||||
class LicensePlateRecognitionConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable license plate recognition.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable LPR",
|
||||
description="Enable or disable license plate recognition for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
model_size: str = Field(
|
||||
default="small", title="The size of the embeddings model used."
|
||||
default="small",
|
||||
title="Model size",
|
||||
description="Model size used for text detection/recognition. Most users should use 'small'.",
|
||||
)
|
||||
detection_threshold: float = Field(
|
||||
default=0.7,
|
||||
title="License plate object confidence score required to begin running recognition.",
|
||||
title="Detection threshold",
|
||||
description="Detection confidence threshold to begin running OCR on a suspected plate.",
|
||||
gt=0.0,
|
||||
le=1.0,
|
||||
)
|
||||
min_area: int = Field(
|
||||
default=1000,
|
||||
title="Minimum area of license plate to begin running recognition.",
|
||||
title="Minimum plate area",
|
||||
description="Minimum plate area (pixels) required to attempt recognition.",
|
||||
)
|
||||
recognition_threshold: float = Field(
|
||||
default=0.9,
|
||||
title="Recognition confidence score required to add the plate to the object as a sub label.",
|
||||
title="Recognition threshold",
|
||||
description="Confidence threshold required for recognized plate text to be attached as a sub-label.",
|
||||
gt=0.0,
|
||||
le=1.0,
|
||||
)
|
||||
min_plate_length: int = Field(
|
||||
default=4,
|
||||
title="Minimum number of characters a license plate must have to be added to the object as a sub label.",
|
||||
title="Min plate length",
|
||||
description="Minimum number of characters a recognized plate must contain to be considered valid.",
|
||||
)
|
||||
format: Optional[str] = Field(
|
||||
default=None,
|
||||
title="Regular expression for the expected format of license plate.",
|
||||
title="Plate format regex",
|
||||
description="Optional regex to validate recognized plate strings against an expected format.",
|
||||
)
|
||||
match_distance: int = Field(
|
||||
default=1,
|
||||
title="Allow this number of missing/incorrect characters to still cause a detected plate to match a known plate.",
|
||||
title="Match distance",
|
||||
description="Number of character mismatches allowed when comparing detected plates to known plates.",
|
||||
ge=0,
|
||||
)
|
||||
known_plates: Optional[Dict[str, List[str]]] = Field(
|
||||
default={}, title="Known plates to track (strings or regular expressions)."
|
||||
default={},
|
||||
title="Known plates",
|
||||
description="List of plates or regexes to specially track or alert on.",
|
||||
)
|
||||
enhancement: int = Field(
|
||||
default=0,
|
||||
title="Amount of contrast adjustment and denoising to apply to license plate images before recognition.",
|
||||
title="Enhancement level",
|
||||
description="Enhancement level (0-10) to apply to plate crops prior to OCR; higher values may not always improve results, levels above 5 may only work with night time plates and should be used with caution.",
|
||||
ge=0,
|
||||
le=10,
|
||||
)
|
||||
debug_save_plates: bool = Field(
|
||||
default=False,
|
||||
title="Save plates captured for LPR for debugging purposes.",
|
||||
title="Save debug plates",
|
||||
description="Save plate crop images for debugging LPR performance.",
|
||||
)
|
||||
device: Optional[str] = Field(
|
||||
default=None,
|
||||
title="The device key to use for LPR.",
|
||||
title="Device",
|
||||
description="This is an override, to target a specific device. See https://onnxruntime.ai/docs/execution-providers/ for more information",
|
||||
)
|
||||
replace_rules: List[ReplaceRule] = Field(
|
||||
default_factory=list,
|
||||
title="List of regex replacement rules for normalizing detected plates. Each rule has 'pattern' and 'replacement'.",
|
||||
title="Replacement rules",
|
||||
description="Regex replacement rules used to normalize detected plate strings before matching.",
|
||||
)
|
||||
|
||||
|
||||
class CameraLicensePlateRecognitionConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable license plate recognition.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable LPR",
|
||||
description="Enable or disable LPR on this camera.",
|
||||
)
|
||||
expire_time: int = Field(
|
||||
default=3,
|
||||
title="Expire plates not seen after number of seconds (for dedicated LPR cameras only).",
|
||||
title="Expire seconds",
|
||||
description="Time in seconds after which an unseen plate is expired from the tracker (for dedicated LPR cameras only).",
|
||||
gt=0,
|
||||
)
|
||||
min_area: int = Field(
|
||||
default=1000,
|
||||
title="Minimum area of license plate to begin running recognition.",
|
||||
title="Minimum plate area",
|
||||
description="Minimum plate area (pixels) required to attempt recognition.",
|
||||
)
|
||||
enhancement: int = Field(
|
||||
default=0,
|
||||
title="Amount of contrast adjustment and denoising to apply to license plate images before recognition.",
|
||||
title="Enhancement level",
|
||||
description="Enhancement level (0-10) to apply to plate crops prior to OCR; higher values may not always improve results, levels above 5 may only work with night time plates and should be used with caution.",
|
||||
ge=0,
|
||||
le=10,
|
||||
)
|
||||
@ -314,12 +422,18 @@ class CameraLicensePlateRecognitionConfig(FrigateBaseModel):
|
||||
|
||||
|
||||
class CameraAudioTranscriptionConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable audio transcription.")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable transcription",
|
||||
description="Enable or disable manually triggered audio event transcription.",
|
||||
)
|
||||
enabled_in_config: Optional[bool] = Field(
|
||||
default=None, title="Keep track of original state of audio transcription."
|
||||
default=None, title="Original transcription state"
|
||||
)
|
||||
live_enabled: Optional[bool] = Field(
|
||||
default=False, title="Enable live transcriptions."
|
||||
default=False,
|
||||
title="Live transcription",
|
||||
description="Enable streaming live transcription for audio as it is received.",
|
||||
)
|
||||
|
||||
model_config = ConfigDict(extra="forbid", protected_namespaces=())
|
||||
|
||||
@ -299,116 +299,189 @@ def verify_lpr_and_face(
|
||||
|
||||
|
||||
class FrigateConfig(FrigateBaseModel):
|
||||
version: Optional[str] = Field(default=None, title="Current config version.")
|
||||
version: Optional[str] = Field(
|
||||
default=None,
|
||||
title="Current config version",
|
||||
description="Numeric or string version of the active configuration to help detect migrations or format changes.",
|
||||
)
|
||||
safe_mode: bool = Field(
|
||||
default=False, title="If Frigate should be started in safe mode."
|
||||
default=False,
|
||||
title="Safe mode",
|
||||
description="When enabled, start Frigate in safe mode with reduced features for troubleshooting.",
|
||||
)
|
||||
|
||||
# Fields that install global state should be defined first, so that their validators run first.
|
||||
environment_vars: EnvVars = Field(
|
||||
default_factory=dict, title="Frigate environment variables."
|
||||
default_factory=dict,
|
||||
title="Environment variables",
|
||||
description="Key/value pairs of environment variables to set for the Frigate process in Home Assistant OS. Non-HAOS users must use Docker environment variable configuration instead.",
|
||||
)
|
||||
logger: LoggerConfig = Field(
|
||||
default_factory=LoggerConfig,
|
||||
title="Logging configuration.",
|
||||
title="Logging",
|
||||
description="Controls default log verbosity and per-component log level overrides.",
|
||||
validate_default=True,
|
||||
)
|
||||
|
||||
# Global config
|
||||
auth: AuthConfig = Field(default_factory=AuthConfig, title="Auth configuration.")
|
||||
auth: AuthConfig = Field(
|
||||
default_factory=AuthConfig,
|
||||
title="Authentication",
|
||||
description="Authentication and session-related settings including cookie and rate limit options.",
|
||||
)
|
||||
database: DatabaseConfig = Field(
|
||||
default_factory=DatabaseConfig, title="Database configuration."
|
||||
default_factory=DatabaseConfig,
|
||||
title="Database",
|
||||
description="Settings for the SQLite database used by Frigate to store tracked object and recording metadata.",
|
||||
)
|
||||
go2rtc: RestreamConfig = Field(
|
||||
default_factory=RestreamConfig, title="Global restream configuration."
|
||||
default_factory=RestreamConfig,
|
||||
title="go2rtc",
|
||||
description="Settings for the integrated go2rtc restreaming service used for live stream relaying and translation.",
|
||||
)
|
||||
mqtt: MqttConfig = Field(
|
||||
title="MQTT",
|
||||
description="Settings for connecting and publishing telemetry, snapshots, and event details to an MQTT broker.",
|
||||
)
|
||||
mqtt: MqttConfig = Field(title="MQTT configuration.")
|
||||
notifications: NotificationConfig = Field(
|
||||
default_factory=NotificationConfig, title="Global notification configuration."
|
||||
default_factory=NotificationConfig,
|
||||
title="Notifications",
|
||||
description="Settings to enable and control notifications for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
networking: NetworkingConfig = Field(
|
||||
default_factory=NetworkingConfig, title="Networking configuration"
|
||||
default_factory=NetworkingConfig,
|
||||
title="Networking",
|
||||
description="Network-related settings such as IPv6 enablement for Frigate endpoints.",
|
||||
)
|
||||
proxy: ProxyConfig = Field(
|
||||
default_factory=ProxyConfig, title="Proxy configuration."
|
||||
default_factory=ProxyConfig,
|
||||
title="Proxy",
|
||||
description="Settings for integrating Frigate behind a reverse proxy that passes authenticated user headers.",
|
||||
)
|
||||
telemetry: TelemetryConfig = Field(
|
||||
default_factory=TelemetryConfig, title="Telemetry configuration."
|
||||
default_factory=TelemetryConfig,
|
||||
title="Telemetry",
|
||||
description="System telemetry and stats options including GPU and network bandwidth monitoring.",
|
||||
)
|
||||
tls: TlsConfig = Field(
|
||||
default_factory=TlsConfig,
|
||||
title="TLS",
|
||||
description="TLS settings for Frigate's web endpoints (port 8971).",
|
||||
)
|
||||
ui: UIConfig = Field(
|
||||
default_factory=UIConfig,
|
||||
title="UI",
|
||||
description="User interface preferences such as timezone, time/date formatting, and units.",
|
||||
)
|
||||
tls: TlsConfig = Field(default_factory=TlsConfig, title="TLS configuration.")
|
||||
ui: UIConfig = Field(default_factory=UIConfig, title="UI configuration.")
|
||||
|
||||
# Detector config
|
||||
detectors: Dict[str, BaseDetectorConfig] = Field(
|
||||
default=DEFAULT_DETECTORS,
|
||||
title="Detector hardware configuration.",
|
||||
title="Detector hardware",
|
||||
description="Configuration for object detectors (CPU, GPU, ONNX backends) and any detector-specific model settings.",
|
||||
)
|
||||
model: ModelConfig = Field(
|
||||
default_factory=ModelConfig, title="Detection model configuration."
|
||||
default_factory=ModelConfig,
|
||||
title="Detection model",
|
||||
description="Settings to configure a custom object detection model and its input shape.",
|
||||
)
|
||||
|
||||
# GenAI config (named provider configs: name -> GenAIConfig)
|
||||
genai: Dict[str, GenAIConfig] = Field(
|
||||
default_factory=dict, title="Generative AI configuration (named providers)."
|
||||
default_factory=dict,
|
||||
title="Generative AI configuration (named providers).",
|
||||
description="Settings for integrated generative AI providers used to generate object descriptions and review summaries.",
|
||||
)
|
||||
|
||||
# Camera config
|
||||
cameras: Dict[str, CameraConfig] = Field(title="Camera configuration.")
|
||||
cameras: Dict[str, CameraConfig] = Field(title="Cameras", description="Cameras")
|
||||
audio: AudioConfig = Field(
|
||||
default_factory=AudioConfig, title="Global Audio events configuration."
|
||||
default_factory=AudioConfig,
|
||||
title="Audio events",
|
||||
description="Settings for audio-based event detection for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
birdseye: BirdseyeConfig = Field(
|
||||
default_factory=BirdseyeConfig, title="Birdseye configuration."
|
||||
default_factory=BirdseyeConfig,
|
||||
title="Birdseye",
|
||||
description="Settings for the Birdseye composite view that composes multiple camera feeds into a single layout.",
|
||||
)
|
||||
detect: DetectConfig = Field(
|
||||
default_factory=DetectConfig, title="Global object tracking configuration."
|
||||
default_factory=DetectConfig,
|
||||
title="Object Detection",
|
||||
description="Settings for the detection/detect role used to run object detection and initialize trackers.",
|
||||
)
|
||||
ffmpeg: FfmpegConfig = Field(
|
||||
default_factory=FfmpegConfig, title="Global FFmpeg configuration."
|
||||
default_factory=FfmpegConfig,
|
||||
title="FFmpeg",
|
||||
description="FFmpeg settings including binary path, args, hwaccel options, and per-role output args.",
|
||||
)
|
||||
live: CameraLiveConfig = Field(
|
||||
default_factory=CameraLiveConfig, title="Live playback settings."
|
||||
default_factory=CameraLiveConfig,
|
||||
title="Live playback",
|
||||
description="Settings used by the Web UI to control live stream resolution and quality.",
|
||||
)
|
||||
motion: Optional[MotionConfig] = Field(
|
||||
default=None, title="Global motion detection configuration."
|
||||
default=None,
|
||||
title="Motion detection",
|
||||
description="Default motion detection settings applied to cameras unless overridden per-camera.",
|
||||
)
|
||||
objects: ObjectConfig = Field(
|
||||
default_factory=ObjectConfig, title="Global object configuration."
|
||||
default_factory=ObjectConfig,
|
||||
title="Objects",
|
||||
description="Object tracking defaults including which labels to track and per-object filters.",
|
||||
)
|
||||
record: RecordConfig = Field(
|
||||
default_factory=RecordConfig, title="Global record configuration."
|
||||
default_factory=RecordConfig,
|
||||
title="Recording",
|
||||
description="Recording and retention settings applied to cameras unless overridden per-camera.",
|
||||
)
|
||||
review: ReviewConfig = Field(
|
||||
default_factory=ReviewConfig, title="Review configuration."
|
||||
default_factory=ReviewConfig,
|
||||
title="Review",
|
||||
description="Settings that control alerts, detections, and GenAI review summaries used by the UI and storage.",
|
||||
)
|
||||
snapshots: SnapshotsConfig = Field(
|
||||
default_factory=SnapshotsConfig, title="Global snapshots configuration."
|
||||
default_factory=SnapshotsConfig,
|
||||
title="Snapshots",
|
||||
description="Settings for saved JPEG snapshots of tracked objects for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
timestamp_style: TimestampStyleConfig = Field(
|
||||
default_factory=TimestampStyleConfig,
|
||||
title="Global timestamp style configuration.",
|
||||
title="Timestamp style",
|
||||
description="Styling options for in-feed timestamps applied to debug view and snapshots.",
|
||||
)
|
||||
|
||||
# Classification Config
|
||||
audio_transcription: AudioTranscriptionConfig = Field(
|
||||
default_factory=AudioTranscriptionConfig, title="Audio transcription config."
|
||||
default_factory=AudioTranscriptionConfig,
|
||||
title="Audio transcription",
|
||||
description="Settings for live and speech audio transcription used for events and live captions.",
|
||||
)
|
||||
classification: ClassificationConfig = Field(
|
||||
default_factory=ClassificationConfig, title="Object classification config."
|
||||
default_factory=ClassificationConfig,
|
||||
title="Object classification",
|
||||
description="Settings for classification models used to refine object labels or state classification.",
|
||||
)
|
||||
semantic_search: SemanticSearchConfig = Field(
|
||||
default_factory=SemanticSearchConfig, title="Semantic search configuration."
|
||||
default_factory=SemanticSearchConfig,
|
||||
title="Semantic Search",
|
||||
description="Settings for Semantic Search which builds and queries object embeddings to find similar items.",
|
||||
)
|
||||
face_recognition: FaceRecognitionConfig = Field(
|
||||
default_factory=FaceRecognitionConfig, title="Face recognition config."
|
||||
default_factory=FaceRecognitionConfig,
|
||||
title="Face recognition",
|
||||
description="Settings for face detection and recognition for all cameras; can be overridden per-camera.",
|
||||
)
|
||||
lpr: LicensePlateRecognitionConfig = Field(
|
||||
default_factory=LicensePlateRecognitionConfig,
|
||||
title="License Plate recognition config.",
|
||||
title="License Plate Recognition",
|
||||
description="License plate recognition settings including detection thresholds, formatting, and known plates.",
|
||||
)
|
||||
|
||||
camera_groups: Dict[str, CameraGroupConfig] = Field(
|
||||
default_factory=dict, title="Camera group configuration"
|
||||
default_factory=dict,
|
||||
title="Camera groups",
|
||||
description="Configuration for named camera groups used to organize cameras in the UI.",
|
||||
)
|
||||
|
||||
_plus_api: PlusApi
|
||||
@ -487,6 +560,9 @@ class FrigateConfig(FrigateBaseModel):
|
||||
|
||||
# users should not set model themselves
|
||||
if detector_config.model:
|
||||
logger.warning(
|
||||
"The model key should be specified at the root level of the config, not under detectors. The nested model key will be ignored."
|
||||
)
|
||||
detector_config.model = None
|
||||
|
||||
model_config = self.model.model_dump(exclude_unset=True, warnings="none")
|
||||
|
||||
@ -8,4 +8,8 @@ __all__ = ["DatabaseConfig"]
|
||||
|
||||
|
||||
class DatabaseConfig(FrigateBaseModel):
|
||||
path: str = Field(default=DEFAULT_DB_PATH, title="Database path.") # noqa: F821
|
||||
path: str = Field(
|
||||
default=DEFAULT_DB_PATH,
|
||||
title="Database path",
|
||||
description="Filesystem path where the Frigate SQLite database file will be stored.",
|
||||
) # noqa: F821
|
||||
|
||||
@ -9,9 +9,15 @@ __all__ = ["LoggerConfig"]
|
||||
|
||||
|
||||
class LoggerConfig(FrigateBaseModel):
|
||||
default: LogLevel = Field(default=LogLevel.info, title="Default logging level.")
|
||||
default: LogLevel = Field(
|
||||
default=LogLevel.info,
|
||||
title="Logging level",
|
||||
description="Default global log verbosity (debug, info, warning, error).",
|
||||
)
|
||||
logs: dict[str, LogLevel] = Field(
|
||||
default_factory=dict, title="Log level for specified processes."
|
||||
default_factory=dict,
|
||||
title="Per-process log level",
|
||||
description="Per-component log level overrides to increase or decrease verbosity for specific modules.",
|
||||
)
|
||||
|
||||
@model_validator(mode="after")
|
||||
|
||||
@ -12,25 +12,73 @@ __all__ = ["MqttConfig"]
|
||||
|
||||
|
||||
class MqttConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=True, title="Enable MQTT Communication.")
|
||||
host: str = Field(default="", title="MQTT Host")
|
||||
port: int = Field(default=1883, title="MQTT Port")
|
||||
topic_prefix: str = Field(default="frigate", title="MQTT Topic Prefix")
|
||||
client_id: str = Field(default="frigate", title="MQTT Client ID")
|
||||
enabled: bool = Field(
|
||||
default=True,
|
||||
title="Enable MQTT",
|
||||
description="Enable or disable MQTT integration for state, events, and snapshots.",
|
||||
)
|
||||
host: str = Field(
|
||||
default="",
|
||||
title="MQTT host",
|
||||
description="Hostname or IP address of the MQTT broker.",
|
||||
)
|
||||
port: int = Field(
|
||||
default=1883,
|
||||
title="MQTT port",
|
||||
description="Port of the MQTT broker (usually 1883 for plain MQTT).",
|
||||
)
|
||||
topic_prefix: str = Field(
|
||||
default="frigate",
|
||||
title="Topic prefix",
|
||||
description="MQTT topic prefix for all Frigate topics; must be unique if running multiple instances.",
|
||||
)
|
||||
client_id: str = Field(
|
||||
default="frigate",
|
||||
title="Client ID",
|
||||
description="Client identifier used when connecting to the MQTT broker; should be unique per instance.",
|
||||
)
|
||||
stats_interval: int = Field(
|
||||
default=60, ge=FREQUENCY_STATS_POINTS, title="MQTT Camera Stats Interval"
|
||||
default=60,
|
||||
ge=FREQUENCY_STATS_POINTS,
|
||||
title="Stats interval",
|
||||
description="Interval in seconds for publishing system and camera stats to MQTT.",
|
||||
)
|
||||
user: Optional[EnvString] = Field(
|
||||
default=None,
|
||||
title="MQTT username",
|
||||
description="Optional MQTT username; can be provided via environment variables or secrets.",
|
||||
)
|
||||
user: Optional[EnvString] = Field(default=None, title="MQTT Username")
|
||||
password: Optional[EnvString] = Field(
|
||||
default=None, title="MQTT Password", validate_default=True
|
||||
default=None,
|
||||
title="MQTT password",
|
||||
description="Optional MQTT password; can be provided via environment variables or secrets.",
|
||||
validate_default=True,
|
||||
)
|
||||
tls_ca_certs: Optional[str] = Field(
|
||||
default=None,
|
||||
title="TLS CA certs",
|
||||
description="Path to CA certificate for TLS connections to the broker (for self-signed certs).",
|
||||
)
|
||||
tls_ca_certs: Optional[str] = Field(default=None, title="MQTT TLS CA Certificates")
|
||||
tls_client_cert: Optional[str] = Field(
|
||||
default=None, title="MQTT TLS Client Certificate"
|
||||
default=None,
|
||||
title="Client cert",
|
||||
description="Client certificate path for TLS mutual authentication; do not set user/password when using client certs.",
|
||||
)
|
||||
tls_client_key: Optional[str] = Field(
|
||||
default=None,
|
||||
title="Client key",
|
||||
description="Private key path for the client certificate.",
|
||||
)
|
||||
tls_insecure: Optional[bool] = Field(
|
||||
default=None,
|
||||
title="TLS insecure",
|
||||
description="Allow insecure TLS connections by skipping hostname verification (not recommended).",
|
||||
)
|
||||
qos: int = Field(
|
||||
default=0,
|
||||
title="MQTT QoS",
|
||||
description="Quality of Service level for MQTT publishes/subscriptions (0, 1, or 2).",
|
||||
)
|
||||
tls_client_key: Optional[str] = Field(default=None, title="MQTT TLS Client Key")
|
||||
tls_insecure: Optional[bool] = Field(default=None, title="MQTT TLS Insecure")
|
||||
qos: int = Field(default=0, title="MQTT QoS")
|
||||
|
||||
@model_validator(mode="after")
|
||||
def user_requires_pass(self, info: ValidationInfo) -> Self:
|
||||
|
||||
@ -8,20 +8,34 @@ __all__ = ["IPv6Config", "ListenConfig", "NetworkingConfig"]
|
||||
|
||||
|
||||
class IPv6Config(FrigateBaseModel):
|
||||
enabled: bool = Field(default=False, title="Enable IPv6 for port 5000 and/or 8971")
|
||||
enabled: bool = Field(
|
||||
default=False,
|
||||
title="Enable IPv6",
|
||||
description="Enable IPv6 support for Frigate services (API and UI) where applicable.",
|
||||
)
|
||||
|
||||
|
||||
class ListenConfig(FrigateBaseModel):
|
||||
internal: Union[int, str] = Field(
|
||||
default=5000, title="Internal listening port for Frigate"
|
||||
default=5000,
|
||||
title="Internal port",
|
||||
description="Internal listening port for Frigate (default 5000).",
|
||||
)
|
||||
external: Union[int, str] = Field(
|
||||
default=8971, title="External listening port for Frigate"
|
||||
default=8971,
|
||||
title="External port",
|
||||
description="External listening port for Frigate (default 8971).",
|
||||
)
|
||||
|
||||
|
||||
class NetworkingConfig(FrigateBaseModel):
|
||||
ipv6: IPv6Config = Field(default_factory=IPv6Config, title="IPv6 configuration")
|
||||
listen: ListenConfig = Field(
|
||||
default_factory=ListenConfig, title="Listening ports configuration"
|
||||
ipv6: IPv6Config = Field(
|
||||
default_factory=IPv6Config,
|
||||
title="IPv6 configuration",
|
||||
description="IPv6-specific settings for Frigate network services.",
|
||||
)
|
||||
listen: ListenConfig = Field(
|
||||
default_factory=ListenConfig,
|
||||
title="Listening ports configuration",
|
||||
description="Configuration for internal and external listening ports. This is for advanced users. For the majority of use cases it's recommended to change the ports section of your Docker compose file.",
|
||||
)
|
||||
|
||||
@ -10,36 +10,47 @@ __all__ = ["ProxyConfig", "HeaderMappingConfig"]
|
||||
|
||||
class HeaderMappingConfig(FrigateBaseModel):
|
||||
user: str = Field(
|
||||
default=None, title="Header name from upstream proxy to identify user."
|
||||
default=None,
|
||||
title="User header",
|
||||
description="Header containing the authenticated username provided by the upstream proxy.",
|
||||
)
|
||||
role: str = Field(
|
||||
default=None,
|
||||
title="Header name from upstream proxy to identify user role.",
|
||||
title="Role header",
|
||||
description="Header containing the authenticated user's role or groups from the upstream proxy.",
|
||||
)
|
||||
role_map: Optional[dict[str, list[str]]] = Field(
|
||||
default_factory=dict,
|
||||
title=("Mapping of Frigate roles to upstream group values. "),
|
||||
title=("Role mapping"),
|
||||
description="Map upstream group values to Frigate roles (for example map admin groups to the admin role).",
|
||||
)
|
||||
|
||||
|
||||
class ProxyConfig(FrigateBaseModel):
|
||||
header_map: HeaderMappingConfig = Field(
|
||||
default_factory=HeaderMappingConfig,
|
||||
title="Header mapping definitions for proxy user passing.",
|
||||
title="Header mapping",
|
||||
description="Map incoming proxy headers to Frigate user and role fields for proxy-based auth.",
|
||||
)
|
||||
logout_url: Optional[str] = Field(
|
||||
default=None, title="Redirect url for logging out with proxy."
|
||||
default=None,
|
||||
title="Logout URL",
|
||||
description="URL to redirect users to when logging out via the proxy.",
|
||||
)
|
||||
auth_secret: Optional[EnvString] = Field(
|
||||
default=None,
|
||||
title="Secret value for proxy authentication.",
|
||||
title="Proxy secret",
|
||||
description="Optional secret checked against the X-Proxy-Secret header to verify trusted proxies.",
|
||||
)
|
||||
default_role: Optional[str] = Field(
|
||||
default="viewer", title="Default role for proxy users."
|
||||
default="viewer",
|
||||
title="Default role",
|
||||
description="Default role assigned to proxy-authenticated users when no role mapping applies (admin or viewer).",
|
||||
)
|
||||
separator: Optional[str] = Field(
|
||||
default=",",
|
||||
title="The character used to separate values in a mapped header.",
|
||||
title="Separator character",
|
||||
description="Character used to split multiple values provided in proxy headers.",
|
||||
)
|
||||
|
||||
@field_validator("separator", mode="before")
|
||||
|
||||
@ -8,22 +8,41 @@ __all__ = ["TelemetryConfig", "StatsConfig"]
|
||||
|
||||
|
||||
class StatsConfig(FrigateBaseModel):
|
||||
amd_gpu_stats: bool = Field(default=True, title="Enable AMD GPU stats.")
|
||||
intel_gpu_stats: bool = Field(default=True, title="Enable Intel GPU stats.")
|
||||
amd_gpu_stats: bool = Field(
|
||||
default=True,
|
||||
title="AMD GPU stats",
|
||||
description="Enable collection of AMD GPU statistics if an AMD GPU is present.",
|
||||
)
|
||||
intel_gpu_stats: bool = Field(
|
||||
default=True,
|
||||
title="Intel GPU stats",
|
||||
description="Enable collection of Intel GPU statistics if an Intel GPU is present.",
|
||||
)
|
||||
network_bandwidth: bool = Field(
|
||||
default=False, title="Enable network bandwidth for ffmpeg processes."
|
||||
default=False,
|
||||
title="Network bandwidth",
|
||||
description="Enable per-process network bandwidth monitoring for camera ffmpeg processes and detectors (requires capabilities).",
|
||||
)
|
||||
intel_gpu_device: Optional[str] = Field(
|
||||
default=None, title="Define the device to use when gathering SR-IOV stats."
|
||||
default=None,
|
||||
title="SR-IOV device",
|
||||
description="Device identifier used when treating Intel GPUs as SR-IOV to fix GPU stats.",
|
||||
)
|
||||
|
||||
|
||||
class TelemetryConfig(FrigateBaseModel):
|
||||
network_interfaces: list[str] = Field(
|
||||
default=[],
|
||||
title="Enabled network interfaces for bandwidth calculation.",
|
||||
title="Network interfaces",
|
||||
description="List of network interface name prefixes to monitor for bandwidth statistics.",
|
||||
)
|
||||
stats: StatsConfig = Field(
|
||||
default_factory=StatsConfig, title="System Stats Configuration"
|
||||
default_factory=StatsConfig,
|
||||
title="System stats",
|
||||
description="Options to enable/disable collection of various system and GPU statistics.",
|
||||
)
|
||||
version_check: bool = Field(
|
||||
default=True,
|
||||
title="Version check",
|
||||
description="Enable an outbound check to detect if a newer Frigate version is available.",
|
||||
)
|
||||
version_check: bool = Field(default=True, title="Enable latest version check.")
|
||||
|
||||
@ -6,4 +6,8 @@ __all__ = ["TlsConfig"]
|
||||
|
||||
|
||||
class TlsConfig(FrigateBaseModel):
|
||||
enabled: bool = Field(default=True, title="Enable TLS for port 8971")
|
||||
enabled: bool = Field(
|
||||
default=True,
|
||||
title="Enable TLS",
|
||||
description="Enable TLS for Frigate's web UI and API on the configured TLS port.",
|
||||
)
|
||||
|
||||
@ -27,16 +27,28 @@ class UnitSystemEnum(str, Enum):
|
||||
|
||||
|
||||
class UIConfig(FrigateBaseModel):
|
||||
timezone: Optional[str] = Field(default=None, title="Override UI timezone.")
|
||||
timezone: Optional[str] = Field(
|
||||
default=None,
|
||||
title="Timezone",
|
||||
description="Optional timezone to display across the UI (defaults to browser local time if unset).",
|
||||
)
|
||||
time_format: TimeFormatEnum = Field(
|
||||
default=TimeFormatEnum.browser, title="Override UI time format."
|
||||
default=TimeFormatEnum.browser,
|
||||
title="Time format",
|
||||
description="Time format to use in the UI (browser, 12hour, or 24hour).",
|
||||
)
|
||||
date_style: DateTimeStyleEnum = Field(
|
||||
default=DateTimeStyleEnum.short, title="Override UI dateStyle."
|
||||
default=DateTimeStyleEnum.short,
|
||||
title="Date style",
|
||||
description="Date style to use in the UI (full, long, medium, short).",
|
||||
)
|
||||
time_style: DateTimeStyleEnum = Field(
|
||||
default=DateTimeStyleEnum.medium, title="Override UI timeStyle."
|
||||
default=DateTimeStyleEnum.medium,
|
||||
title="Time style",
|
||||
description="Time style to use in the UI (full, long, medium, short).",
|
||||
)
|
||||
unit_system: UnitSystemEnum = Field(
|
||||
default=UnitSystemEnum.metric, title="The unit system to use for measurements."
|
||||
default=UnitSystemEnum.metric,
|
||||
title="Unit system",
|
||||
description="Unit system for display (metric or imperial) used in the UI and MQTT.",
|
||||
)
|
||||
|
||||
@ -45,30 +45,55 @@ class ModelTypeEnum(str, Enum):
|
||||
|
||||
|
||||
class ModelConfig(BaseModel):
|
||||
path: Optional[str] = Field(None, title="Custom Object detection model path.")
|
||||
labelmap_path: Optional[str] = Field(
|
||||
None, title="Label map for custom object detector."
|
||||
path: Optional[str] = Field(
|
||||
None,
|
||||
title="Custom Object detection model path",
|
||||
description="Path to a custom detection model file (or plus://<model_id> for Frigate+ models).",
|
||||
)
|
||||
labelmap_path: Optional[str] = Field(
|
||||
None,
|
||||
title="Label map for custom object detector",
|
||||
description="Path to a labelmap file that maps numeric classes to string labels for the detector.",
|
||||
)
|
||||
width: int = Field(
|
||||
default=320,
|
||||
title="Object detection model input width",
|
||||
description="Width of the model input tensor in pixels.",
|
||||
)
|
||||
height: int = Field(
|
||||
default=320,
|
||||
title="Object detection model input height",
|
||||
description="Height of the model input tensor in pixels.",
|
||||
)
|
||||
width: int = Field(default=320, title="Object detection model input width.")
|
||||
height: int = Field(default=320, title="Object detection model input height.")
|
||||
labelmap: Dict[int, str] = Field(
|
||||
default_factory=dict, title="Labelmap customization."
|
||||
default_factory=dict,
|
||||
title="Labelmap customization",
|
||||
description="Overrides or remapping entries to merge into the standard labelmap.",
|
||||
)
|
||||
attributes_map: Dict[str, list[str]] = Field(
|
||||
default=DEFAULT_ATTRIBUTE_LABEL_MAP,
|
||||
title="Map of object labels to their attribute labels.",
|
||||
title="Map of object labels to their attribute labels",
|
||||
description="Mapping from object labels to attribute labels used to attach metadata (for example 'car' -> ['license_plate']).",
|
||||
)
|
||||
input_tensor: InputTensorEnum = Field(
|
||||
default=InputTensorEnum.nhwc, title="Model Input Tensor Shape"
|
||||
default=InputTensorEnum.nhwc,
|
||||
title="Model Input Tensor Shape",
|
||||
description="Tensor format expected by the model: 'nhwc' or 'nchw'.",
|
||||
)
|
||||
input_pixel_format: PixelFormatEnum = Field(
|
||||
default=PixelFormatEnum.rgb, title="Model Input Pixel Color Format"
|
||||
default=PixelFormatEnum.rgb,
|
||||
title="Model Input Pixel Color Format",
|
||||
description="Pixel colorspace expected by the model: 'rgb', 'bgr', or 'yuv'.",
|
||||
)
|
||||
input_dtype: InputDTypeEnum = Field(
|
||||
default=InputDTypeEnum.int, title="Model Input D Type"
|
||||
default=InputDTypeEnum.int,
|
||||
title="Model Input D Type",
|
||||
description="Data type of the model input tensor (for example 'float32').",
|
||||
)
|
||||
model_type: ModelTypeEnum = Field(
|
||||
default=ModelTypeEnum.ssd, title="Object Detection Model Type"
|
||||
default=ModelTypeEnum.ssd,
|
||||
title="Object Detection Model Type",
|
||||
description="Detector model architecture type (ssd, yolox, yolonas) used by some detectors for optimization.",
|
||||
)
|
||||
_merged_labelmap: Optional[Dict[int, str]] = PrivateAttr()
|
||||
_colormap: Dict[int, Tuple[int, int, int]] = PrivateAttr()
|
||||
@ -210,12 +235,20 @@ class ModelConfig(BaseModel):
|
||||
|
||||
class BaseDetectorConfig(BaseModel):
|
||||
# the type field must be defined in all subclasses
|
||||
type: str = Field(default="cpu", title="Detector Type")
|
||||
type: str = Field(
|
||||
default="cpu",
|
||||
title="Detector Type",
|
||||
description="Type of detector to use for object detection (for example 'cpu', 'edgetpu', 'openvino').",
|
||||
)
|
||||
model: Optional[ModelConfig] = Field(
|
||||
default=None, title="Detector specific model configuration."
|
||||
default=None,
|
||||
title="Detector specific model configuration",
|
||||
description="Detector-specific model configuration options (path, input size, etc.).",
|
||||
)
|
||||
model_path: Optional[str] = Field(
|
||||
default=None, title="Detector specific model path."
|
||||
default=None,
|
||||
title="Detector specific model path",
|
||||
description="File path to the detector model binary if required by the chosen detector.",
|
||||
)
|
||||
model_config = ConfigDict(
|
||||
extra="allow", arbitrary_types_allowed=True, protected_namespaces=()
|
||||
|
||||
@ -1,6 +1,6 @@
|
||||
import logging
|
||||
|
||||
from pydantic import Field
|
||||
from pydantic import ConfigDict, Field
|
||||
from typing_extensions import Literal
|
||||
|
||||
from frigate.detectors.detection_api import DetectionApi
|
||||
@ -21,8 +21,18 @@ DETECTOR_KEY = "cpu"
|
||||
|
||||
|
||||
class CpuDetectorConfig(BaseDetectorConfig):
|
||||
"""CPU TFLite detector that runs TensorFlow Lite models on the host CPU without hardware acceleration. Not recommended."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="CPU",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
num_threads: int = Field(default=3, title="Number of detection threads")
|
||||
num_threads: int = Field(
|
||||
default=3,
|
||||
title="Number of detection threads",
|
||||
description="The number of threads used for CPU-based inference.",
|
||||
)
|
||||
|
||||
|
||||
class CpuTfl(DetectionApi):
|
||||
|
||||
@ -4,7 +4,7 @@ import logging
|
||||
import numpy as np
|
||||
import requests
|
||||
from PIL import Image
|
||||
from pydantic import Field
|
||||
from pydantic import ConfigDict, Field
|
||||
from typing_extensions import Literal
|
||||
|
||||
from frigate.detectors.detection_api import DetectionApi
|
||||
@ -16,12 +16,28 @@ DETECTOR_KEY = "deepstack"
|
||||
|
||||
|
||||
class DeepstackDetectorConfig(BaseDetectorConfig):
|
||||
"""DeepStack/CodeProject.AI detector that sends images to a remote DeepStack HTTP API for inference. Not recommended."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="DeepStack",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
api_url: str = Field(
|
||||
default="http://localhost:80/v1/vision/detection", title="DeepStack API URL"
|
||||
default="http://localhost:80/v1/vision/detection",
|
||||
title="DeepStack API URL",
|
||||
description="The URL of the DeepStack API.",
|
||||
)
|
||||
api_timeout: float = Field(
|
||||
default=0.1,
|
||||
title="DeepStack API timeout (in seconds)",
|
||||
description="Maximum time allowed for a DeepStack API request.",
|
||||
)
|
||||
api_key: str = Field(
|
||||
default="",
|
||||
title="DeepStack API key (if required)",
|
||||
description="Optional API key for authenticated DeepStack services.",
|
||||
)
|
||||
api_timeout: float = Field(default=0.1, title="DeepStack API timeout (in seconds)")
|
||||
api_key: str = Field(default="", title="DeepStack API key (if required)")
|
||||
|
||||
|
||||
class DeepStack(DetectionApi):
|
||||
|
||||
@ -2,7 +2,7 @@ import logging
|
||||
import queue
|
||||
|
||||
import numpy as np
|
||||
from pydantic import Field
|
||||
from pydantic import ConfigDict, Field
|
||||
from typing_extensions import Literal
|
||||
|
||||
from frigate.detectors.detection_api import DetectionApi
|
||||
@ -14,10 +14,28 @@ DETECTOR_KEY = "degirum"
|
||||
|
||||
### DETECTOR CONFIG ###
|
||||
class DGDetectorConfig(BaseDetectorConfig):
|
||||
"""DeGirum detector for running models via DeGirum cloud or local inference services."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="DeGirum",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
location: str = Field(default=None, title="Inference Location")
|
||||
zoo: str = Field(default=None, title="Model Zoo")
|
||||
token: str = Field(default=None, title="DeGirum Cloud Token")
|
||||
location: str = Field(
|
||||
default=None,
|
||||
title="Inference Location",
|
||||
description="Location of the DeGirim inference engine (e.g. '@cloud', '127.0.0.1').",
|
||||
)
|
||||
zoo: str = Field(
|
||||
default=None,
|
||||
title="Model Zoo",
|
||||
description="Path or URL to the DeGirum model zoo.",
|
||||
)
|
||||
token: str = Field(
|
||||
default=None,
|
||||
title="DeGirum Cloud Token",
|
||||
description="Token for DeGirum Cloud access.",
|
||||
)
|
||||
|
||||
|
||||
### ACTUAL DETECTOR ###
|
||||
|
||||
@ -4,7 +4,7 @@ import os
|
||||
|
||||
import cv2
|
||||
import numpy as np
|
||||
from pydantic import Field
|
||||
from pydantic import ConfigDict, Field
|
||||
from typing_extensions import Literal
|
||||
|
||||
from frigate.detectors.detection_api import DetectionApi
|
||||
@ -21,8 +21,18 @@ DETECTOR_KEY = "edgetpu"
|
||||
|
||||
|
||||
class EdgeTpuDetectorConfig(BaseDetectorConfig):
|
||||
"""EdgeTPU detector that runs TensorFlow Lite models compiled for Coral EdgeTPU using the EdgeTPU delegate."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="EdgeTPU",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
device: str = Field(default=None, title="Device Type")
|
||||
device: str = Field(
|
||||
default=None,
|
||||
title="Device Type",
|
||||
description="The device to use for EdgeTPU inference (e.g. 'usb', 'pci').",
|
||||
)
|
||||
|
||||
|
||||
class EdgeTpuTfl(DetectionApi):
|
||||
|
||||
@ -8,7 +8,7 @@ from typing import Dict, List, Optional, Tuple
|
||||
|
||||
import cv2
|
||||
import numpy as np
|
||||
from pydantic import Field
|
||||
from pydantic import ConfigDict, Field
|
||||
from typing_extensions import Literal
|
||||
|
||||
from frigate.const import MODEL_CACHE_DIR
|
||||
@ -410,5 +410,15 @@ class HailoDetector(DetectionApi):
|
||||
|
||||
# ----------------- HailoDetectorConfig Class ----------------- #
|
||||
class HailoDetectorConfig(BaseDetectorConfig):
|
||||
"""Hailo-8/Hailo-8L detector using HEF models and the HailoRT SDK for inference on Hailo hardware."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="Hailo-8/Hailo-8L",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
device: str = Field(default="PCIe", title="Device Type")
|
||||
device: str = Field(
|
||||
default="PCIe",
|
||||
title="Device Type",
|
||||
description="The device to use for Hailo inference (e.g. 'PCIe', 'M.2').",
|
||||
)
|
||||
|
||||
@ -8,7 +8,7 @@ from queue import Queue
|
||||
|
||||
import cv2
|
||||
import numpy as np
|
||||
from pydantic import BaseModel, Field
|
||||
from pydantic import BaseModel, ConfigDict, Field
|
||||
from typing_extensions import Literal
|
||||
|
||||
from frigate.detectors.detection_api import DetectionApi
|
||||
@ -30,8 +30,18 @@ class ModelConfig(BaseModel):
|
||||
|
||||
|
||||
class MemryXDetectorConfig(BaseDetectorConfig):
|
||||
"""MemryX MX3 detector that runs compiled DFP models on MemryX accelerators."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="MemryX",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
device: str = Field(default="PCIe", title="Device Path")
|
||||
device: str = Field(
|
||||
default="PCIe",
|
||||
title="Device Path",
|
||||
description="The device to use for MemryX inference (e.g. 'PCIe').",
|
||||
)
|
||||
|
||||
|
||||
class MemryXDetector(DetectionApi):
|
||||
|
||||
@ -1,7 +1,7 @@
|
||||
import logging
|
||||
|
||||
import numpy as np
|
||||
from pydantic import Field
|
||||
from pydantic import ConfigDict, Field
|
||||
from typing_extensions import Literal
|
||||
|
||||
from frigate.detectors.detection_api import DetectionApi
|
||||
@ -23,8 +23,18 @@ DETECTOR_KEY = "onnx"
|
||||
|
||||
|
||||
class ONNXDetectorConfig(BaseDetectorConfig):
|
||||
"""ONNX detector for running ONNX models; will use available acceleration backends (CUDA/ROCm/OpenVINO) when available."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="ONNX",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
device: str = Field(default="AUTO", title="Device Type")
|
||||
device: str = Field(
|
||||
default="AUTO",
|
||||
title="Device Type",
|
||||
description="The device to use for ONNX inference (e.g. 'AUTO', 'CPU', 'GPU').",
|
||||
)
|
||||
|
||||
|
||||
class ONNXDetector(DetectionApi):
|
||||
|
||||
@ -2,7 +2,7 @@ import logging
|
||||
|
||||
import numpy as np
|
||||
import openvino as ov
|
||||
from pydantic import Field
|
||||
from pydantic import ConfigDict, Field
|
||||
from typing_extensions import Literal
|
||||
|
||||
from frigate.detectors.detection_api import DetectionApi
|
||||
@ -20,8 +20,18 @@ DETECTOR_KEY = "openvino"
|
||||
|
||||
|
||||
class OvDetectorConfig(BaseDetectorConfig):
|
||||
"""OpenVINO detector for AMD and Intel CPUs, Intel GPUs and Intel VPU hardware."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="OpenVINO",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
device: str = Field(default=None, title="Device Type")
|
||||
device: str = Field(
|
||||
default=None,
|
||||
title="Device Type",
|
||||
description="The device to use for OpenVINO inference (e.g. 'CPU', 'GPU', 'NPU').",
|
||||
)
|
||||
|
||||
|
||||
class OvDetector(DetectionApi):
|
||||
|
||||
@ -6,7 +6,7 @@ from typing import Literal
|
||||
|
||||
import cv2
|
||||
import numpy as np
|
||||
from pydantic import Field
|
||||
from pydantic import ConfigDict, Field
|
||||
|
||||
from frigate.const import MODEL_CACHE_DIR, SUPPORTED_RK_SOCS
|
||||
from frigate.detectors.detection_api import DetectionApi
|
||||
@ -29,8 +29,20 @@ model_cache_dir = os.path.join(MODEL_CACHE_DIR, "rknn_cache/")
|
||||
|
||||
|
||||
class RknnDetectorConfig(BaseDetectorConfig):
|
||||
"""RKNN detector for Rockchip NPUs; runs compiled RKNN models on Rockchip hardware."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="RKNN",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
num_cores: int = Field(default=0, ge=0, le=3, title="Number of NPU cores to use.")
|
||||
num_cores: int = Field(
|
||||
default=0,
|
||||
ge=0,
|
||||
le=3,
|
||||
title="Number of NPU cores to use.",
|
||||
description="The number of NPU cores to use (0 for auto).",
|
||||
)
|
||||
|
||||
|
||||
class Rknn(DetectionApi):
|
||||
|
||||
@ -2,6 +2,7 @@ import logging
|
||||
import os
|
||||
|
||||
import numpy as np
|
||||
from pydantic import ConfigDict
|
||||
from typing_extensions import Literal
|
||||
|
||||
from frigate.detectors.detection_api import DetectionApi
|
||||
@ -27,6 +28,12 @@ DETECTOR_KEY = "synaptics"
|
||||
|
||||
|
||||
class SynapDetectorConfig(BaseDetectorConfig):
|
||||
"""Synaptics NPU detector for models in .synap format using the Synap SDK on Synaptics hardware."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="Synaptics",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
|
||||
|
||||
|
||||
@ -1,5 +1,6 @@
|
||||
import logging
|
||||
|
||||
from pydantic import ConfigDict
|
||||
from typing_extensions import Literal
|
||||
|
||||
from frigate.detectors.detection_api import DetectionApi
|
||||
@ -18,6 +19,12 @@ DETECTOR_KEY = "teflon_tfl"
|
||||
|
||||
|
||||
class TeflonDetectorConfig(BaseDetectorConfig):
|
||||
"""Teflon delegate detector for TFLite using Mesa Teflon delegate library to accelerate inference on supported GPUs."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="Teflon",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
|
||||
|
||||
|
||||
@ -14,7 +14,7 @@ try:
|
||||
except ModuleNotFoundError:
|
||||
TRT_SUPPORT = False
|
||||
|
||||
from pydantic import Field
|
||||
from pydantic import ConfigDict, Field
|
||||
from typing_extensions import Literal
|
||||
|
||||
from frigate.detectors.detection_api import DetectionApi
|
||||
@ -46,8 +46,16 @@ if TRT_SUPPORT:
|
||||
|
||||
|
||||
class TensorRTDetectorConfig(BaseDetectorConfig):
|
||||
"""TensorRT detector for Nvidia Jetson devices using serialized TensorRT engines for accelerated inference."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="TensorRT",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
device: int = Field(default=0, title="GPU Device Index")
|
||||
device: int = Field(
|
||||
default=0, title="GPU Device Index", description="The GPU device index to use."
|
||||
)
|
||||
|
||||
|
||||
class HostDeviceMem(object):
|
||||
|
||||
@ -5,7 +5,7 @@ from typing import Any, List
|
||||
|
||||
import numpy as np
|
||||
import zmq
|
||||
from pydantic import Field
|
||||
from pydantic import ConfigDict, Field
|
||||
from typing_extensions import Literal
|
||||
|
||||
from frigate.detectors.detection_api import DetectionApi
|
||||
@ -17,14 +17,28 @@ DETECTOR_KEY = "zmq"
|
||||
|
||||
|
||||
class ZmqDetectorConfig(BaseDetectorConfig):
|
||||
"""ZMQ IPC detector that offloads inference to an external process via a ZeroMQ IPC endpoint."""
|
||||
|
||||
model_config = ConfigDict(
|
||||
title="ZMQ IPC",
|
||||
)
|
||||
|
||||
type: Literal[DETECTOR_KEY]
|
||||
endpoint: str = Field(
|
||||
default="ipc:///tmp/cache/zmq_detector", title="ZMQ IPC endpoint"
|
||||
default="ipc:///tmp/cache/zmq_detector",
|
||||
title="ZMQ IPC endpoint",
|
||||
description="The ZMQ endpoint to connect to.",
|
||||
)
|
||||
request_timeout_ms: int = Field(
|
||||
default=200, title="ZMQ request timeout in milliseconds"
|
||||
default=200,
|
||||
title="ZMQ request timeout in milliseconds",
|
||||
description="Timeout for ZMQ requests in milliseconds.",
|
||||
)
|
||||
linger_ms: int = Field(
|
||||
default=0,
|
||||
title="ZMQ socket linger in milliseconds",
|
||||
description="Socket linger period in milliseconds.",
|
||||
)
|
||||
linger_ms: int = Field(default=0, title="ZMQ socket linger in milliseconds")
|
||||
|
||||
|
||||
class ZmqIpcDetector(DetectionApi):
|
||||
|
||||
@ -195,7 +195,8 @@ def flatten_config_data(
|
||||
) -> Dict[str, Any]:
|
||||
items = []
|
||||
for key, value in config_data.items():
|
||||
new_key = f"{parent_key}.{key}" if parent_key else key
|
||||
escaped_key = escape_config_key_segment(str(key))
|
||||
new_key = f"{parent_key}.{escaped_key}" if parent_key else escaped_key
|
||||
if isinstance(value, dict):
|
||||
items.extend(flatten_config_data(value, new_key).items())
|
||||
else:
|
||||
@ -203,6 +204,41 @@ def flatten_config_data(
|
||||
return dict(items)
|
||||
|
||||
|
||||
def escape_config_key_segment(segment: str) -> str:
|
||||
"""Escape dots and backslashes so they can be treated as literal key chars."""
|
||||
return segment.replace("\\", "\\\\").replace(".", "\\.")
|
||||
|
||||
|
||||
def split_config_key_path(key_path_str: str) -> list[str]:
|
||||
"""Split a dotted config path, honoring \\. as a literal dot in a key."""
|
||||
parts: list[str] = []
|
||||
current: list[str] = []
|
||||
escaped = False
|
||||
|
||||
for char in key_path_str:
|
||||
if escaped:
|
||||
current.append(char)
|
||||
escaped = False
|
||||
continue
|
||||
|
||||
if char == "\\":
|
||||
escaped = True
|
||||
continue
|
||||
|
||||
if char == ".":
|
||||
parts.append("".join(current))
|
||||
current = []
|
||||
continue
|
||||
|
||||
current.append(char)
|
||||
|
||||
if escaped:
|
||||
current.append("\\")
|
||||
|
||||
parts.append("".join(current))
|
||||
return parts
|
||||
|
||||
|
||||
def update_yaml_file_bulk(file_path: str, updates: Dict[str, Any]):
|
||||
yaml = YAML()
|
||||
yaml.indent(mapping=2, sequence=4, offset=2)
|
||||
@ -218,7 +254,7 @@ def update_yaml_file_bulk(file_path: str, updates: Dict[str, Any]):
|
||||
|
||||
# Apply all updates
|
||||
for key_path_str, new_value in updates.items():
|
||||
key_path = key_path_str.split(".")
|
||||
key_path = split_config_key_path(key_path_str)
|
||||
for i in range(len(key_path)):
|
||||
try:
|
||||
index = int(key_path[i])
|
||||
|
||||
46
frigate/util/schema.py
Normal file
46
frigate/util/schema.py
Normal file
@ -0,0 +1,46 @@
|
||||
"""JSON schema utilities for Frigate."""
|
||||
|
||||
from typing import Any, Dict, Type
|
||||
|
||||
from pydantic import BaseModel, TypeAdapter
|
||||
|
||||
|
||||
def get_config_schema(config_class: Type[BaseModel]) -> Dict[str, Any]:
|
||||
"""
|
||||
Returns the JSON schema for FrigateConfig with polymorphic detectors.
|
||||
|
||||
This utility patches the FrigateConfig schema to include the full polymorphic
|
||||
definitions for detectors. By default, Pydantic's schema for Dict[str, BaseDetectorConfig]
|
||||
only includes the base class fields. This function replaces it with a reference
|
||||
to the DetectorConfig union, which includes all available detector subclasses.
|
||||
"""
|
||||
# Import here to ensure all detector plugins are loaded through the detectors module
|
||||
from frigate.detectors import DetectorConfig
|
||||
|
||||
# Get the base schema for FrigateConfig
|
||||
schema = config_class.model_json_schema()
|
||||
|
||||
# Get the schema for the polymorphic DetectorConfig union
|
||||
detector_adapter: TypeAdapter = TypeAdapter(DetectorConfig)
|
||||
detector_schema = detector_adapter.json_schema()
|
||||
|
||||
# Ensure $defs exists in FrigateConfig schema
|
||||
if "$defs" not in schema:
|
||||
schema["$defs"] = {}
|
||||
|
||||
# Merge $defs from DetectorConfig into FrigateConfig schema
|
||||
# This includes the specific schemas for each detector plugin (OvDetectorConfig, etc.)
|
||||
if "$defs" in detector_schema:
|
||||
schema["$defs"].update(detector_schema["$defs"])
|
||||
|
||||
# Extract the union schema (oneOf/discriminator) and add it as a definition
|
||||
detector_union_schema = {k: v for k, v in detector_schema.items() if k != "$defs"}
|
||||
schema["$defs"]["DetectorConfig"] = detector_union_schema
|
||||
|
||||
# Update the 'detectors' property to use the polymorphic DetectorConfig definition
|
||||
if "detectors" in schema.get("properties", {}):
|
||||
schema["properties"]["detectors"]["additionalProperties"] = {
|
||||
"$ref": "#/$defs/DetectorConfig"
|
||||
}
|
||||
|
||||
return schema
|
||||
@ -8,20 +8,18 @@ and generates JSON translation files with titles and descriptions for the web UI
|
||||
|
||||
import json
|
||||
import logging
|
||||
import shutil
|
||||
import sys
|
||||
from pathlib import Path
|
||||
from typing import Any, Dict, Optional, get_args, get_origin
|
||||
|
||||
from pydantic import BaseModel
|
||||
from pydantic.fields import FieldInfo
|
||||
from typing import Any, Dict, get_args, get_origin
|
||||
|
||||
from frigate.config.config import FrigateConfig
|
||||
from frigate.util.schema import get_config_schema
|
||||
|
||||
logging.basicConfig(level=logging.INFO)
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def get_field_translations(field_info: FieldInfo) -> Dict[str, str]:
|
||||
def get_field_translations(field_info) -> Dict[str, str]:
|
||||
"""Extract title and description from a Pydantic field."""
|
||||
translations = {}
|
||||
|
||||
@ -34,50 +32,147 @@ def get_field_translations(field_info: FieldInfo) -> Dict[str, str]:
|
||||
return translations
|
||||
|
||||
|
||||
def process_model_fields(model: type[BaseModel]) -> Dict[str, Any]:
|
||||
def extract_translations_from_schema(
|
||||
schema: Dict[str, Any], defs: Dict[str, Any] = None
|
||||
) -> Dict[str, Any]:
|
||||
"""
|
||||
Recursively process a Pydantic model to extract translations.
|
||||
Recursively extract translations (titles and descriptions) from a JSON schema.
|
||||
|
||||
Returns a nested dictionary structure matching the config schema,
|
||||
with title and description for each field.
|
||||
Returns a dictionary structure with label and description for each field,
|
||||
and nested fields directly under their parent keys.
|
||||
"""
|
||||
if defs is None:
|
||||
defs = schema.get("$defs", {})
|
||||
|
||||
translations = {}
|
||||
|
||||
model_fields = model.model_fields
|
||||
# Add top-level title and description if present
|
||||
if "title" in schema:
|
||||
translations["label"] = schema["title"]
|
||||
if "description" in schema:
|
||||
translations["description"] = schema["description"]
|
||||
|
||||
for field_name, field_info in model_fields.items():
|
||||
field_translations = get_field_translations(field_info)
|
||||
# Process nested properties
|
||||
properties = schema.get("properties", {})
|
||||
for field_name, field_schema in properties.items():
|
||||
field_translations = {}
|
||||
|
||||
# Get the field's type annotation
|
||||
field_type = field_info.annotation
|
||||
# Handle $ref references
|
||||
if "$ref" in field_schema:
|
||||
ref_path = field_schema["$ref"]
|
||||
if ref_path.startswith("#/$defs/"):
|
||||
ref_name = ref_path.split("/")[-1]
|
||||
if ref_name in defs:
|
||||
ref_schema = defs[ref_name]
|
||||
# Extract from the referenced schema
|
||||
ref_translations = extract_translations_from_schema(
|
||||
ref_schema, defs=defs
|
||||
)
|
||||
# Use the $ref field's own title/description if present
|
||||
if "title" in field_schema:
|
||||
field_translations["label"] = field_schema["title"]
|
||||
elif "label" in ref_translations:
|
||||
field_translations["label"] = ref_translations["label"]
|
||||
if "description" in field_schema:
|
||||
field_translations["description"] = field_schema["description"]
|
||||
elif "description" in ref_translations:
|
||||
field_translations["description"] = ref_translations[
|
||||
"description"
|
||||
]
|
||||
# Add nested properties from referenced schema
|
||||
nested_without_root = {
|
||||
k: v
|
||||
for k, v in ref_translations.items()
|
||||
if k not in ("label", "description")
|
||||
}
|
||||
field_translations.update(nested_without_root)
|
||||
# Handle additionalProperties with $ref (for dict types)
|
||||
elif "additionalProperties" in field_schema:
|
||||
additional_props = field_schema["additionalProperties"]
|
||||
# Extract title and description from the field itself
|
||||
if "title" in field_schema:
|
||||
field_translations["label"] = field_schema["title"]
|
||||
if "description" in field_schema:
|
||||
field_translations["description"] = field_schema["description"]
|
||||
|
||||
# Handle Optional types
|
||||
origin = get_origin(field_type)
|
||||
# If additionalProperties contains a $ref, extract nested translations
|
||||
if "$ref" in additional_props:
|
||||
ref_path = additional_props["$ref"]
|
||||
if ref_path.startswith("#/$defs/"):
|
||||
ref_name = ref_path.split("/")[-1]
|
||||
if ref_name in defs:
|
||||
ref_schema = defs[ref_name]
|
||||
nested = extract_translations_from_schema(ref_schema, defs=defs)
|
||||
nested_without_root = {
|
||||
k: v
|
||||
for k, v in nested.items()
|
||||
if k not in ("label", "description")
|
||||
}
|
||||
field_translations.update(nested_without_root)
|
||||
# Handle items with $ref (for array types)
|
||||
elif "items" in field_schema:
|
||||
items = field_schema["items"]
|
||||
# Extract title and description from the field itself
|
||||
if "title" in field_schema:
|
||||
field_translations["label"] = field_schema["title"]
|
||||
if "description" in field_schema:
|
||||
field_translations["description"] = field_schema["description"]
|
||||
|
||||
if origin is Optional or (
|
||||
hasattr(origin, "__name__") and origin.__name__ == "UnionType"
|
||||
):
|
||||
args = get_args(field_type)
|
||||
field_type = next(
|
||||
(arg for arg in args if arg is not type(None)), field_type
|
||||
)
|
||||
# If items contains a $ref, extract nested translations
|
||||
if "$ref" in items:
|
||||
ref_path = items["$ref"]
|
||||
if ref_path.startswith("#/$defs/"):
|
||||
ref_name = ref_path.split("/")[-1]
|
||||
if ref_name in defs:
|
||||
ref_schema = defs[ref_name]
|
||||
nested = extract_translations_from_schema(ref_schema, defs=defs)
|
||||
nested_without_root = {
|
||||
k: v
|
||||
for k, v in nested.items()
|
||||
if k not in ("label", "description")
|
||||
}
|
||||
field_translations.update(nested_without_root)
|
||||
else:
|
||||
# Extract title and description
|
||||
if "title" in field_schema:
|
||||
field_translations["label"] = field_schema["title"]
|
||||
if "description" in field_schema:
|
||||
field_translations["description"] = field_schema["description"]
|
||||
|
||||
# Handle Dict types (like Dict[str, CameraConfig])
|
||||
if get_origin(field_type) is dict:
|
||||
dict_args = get_args(field_type)
|
||||
|
||||
if len(dict_args) >= 2:
|
||||
value_type = dict_args[1]
|
||||
|
||||
if isinstance(value_type, type) and issubclass(value_type, BaseModel):
|
||||
nested_translations = process_model_fields(value_type)
|
||||
|
||||
if nested_translations:
|
||||
field_translations["properties"] = nested_translations
|
||||
elif isinstance(field_type, type) and issubclass(field_type, BaseModel):
|
||||
nested_translations = process_model_fields(field_type)
|
||||
if nested_translations:
|
||||
field_translations["properties"] = nested_translations
|
||||
# Recursively process nested properties
|
||||
if "properties" in field_schema:
|
||||
nested = extract_translations_from_schema(field_schema, defs=defs)
|
||||
# Merge nested translations
|
||||
nested_without_root = {
|
||||
k: v for k, v in nested.items() if k not in ("label", "description")
|
||||
}
|
||||
field_translations.update(nested_without_root)
|
||||
# Handle anyOf cases
|
||||
elif "anyOf" in field_schema:
|
||||
for item in field_schema["anyOf"]:
|
||||
if "properties" in item:
|
||||
nested = extract_translations_from_schema(item, defs=defs)
|
||||
nested_without_root = {
|
||||
k: v
|
||||
for k, v in nested.items()
|
||||
if k not in ("label", "description")
|
||||
}
|
||||
field_translations.update(nested_without_root)
|
||||
elif "$ref" in item:
|
||||
ref_path = item["$ref"]
|
||||
if ref_path.startswith("#/$defs/"):
|
||||
ref_name = ref_path.split("/")[-1]
|
||||
if ref_name in defs:
|
||||
ref_schema = defs[ref_name]
|
||||
nested = extract_translations_from_schema(
|
||||
ref_schema, defs=defs
|
||||
)
|
||||
nested_without_root = {
|
||||
k: v
|
||||
for k, v in nested.items()
|
||||
if k not in ("label", "description")
|
||||
}
|
||||
field_translations.update(nested_without_root)
|
||||
|
||||
if field_translations:
|
||||
translations[field_name] = field_translations
|
||||
@ -85,76 +180,350 @@ def process_model_fields(model: type[BaseModel]) -> Dict[str, Any]:
|
||||
return translations
|
||||
|
||||
|
||||
def generate_section_translation(
|
||||
section_name: str, field_info: FieldInfo
|
||||
) -> Dict[str, Any]:
|
||||
def generate_section_translation(config_class: type) -> Dict[str, Any]:
|
||||
"""
|
||||
Generate translation structure for a top-level config section.
|
||||
Generate translation structure for a config section using its JSON schema.
|
||||
"""
|
||||
section_translations = get_field_translations(field_info)
|
||||
field_type = field_info.annotation
|
||||
origin = get_origin(field_type)
|
||||
schema = config_class.model_json_schema()
|
||||
return extract_translations_from_schema(schema)
|
||||
|
||||
if origin is Optional or (
|
||||
hasattr(origin, "__name__") and origin.__name__ == "UnionType"
|
||||
):
|
||||
args = get_args(field_type)
|
||||
field_type = next((arg for arg in args if arg is not type(None)), field_type)
|
||||
|
||||
# Handle Dict types (like detectors, cameras, camera_groups)
|
||||
if get_origin(field_type) is dict:
|
||||
dict_args = get_args(field_type)
|
||||
if len(dict_args) >= 2:
|
||||
value_type = dict_args[1]
|
||||
if isinstance(value_type, type) and issubclass(value_type, BaseModel):
|
||||
nested = process_model_fields(value_type)
|
||||
if nested:
|
||||
section_translations["properties"] = nested
|
||||
def get_detector_translations(
|
||||
config_schema: Dict[str, Any],
|
||||
) -> tuple[Dict[str, Any], set[str]]:
|
||||
"""Build detector type translations with nested fields based on schema definitions."""
|
||||
defs = config_schema.get("$defs", {})
|
||||
detector_schema = defs.get("DetectorConfig", {})
|
||||
discriminator = detector_schema.get("discriminator", {})
|
||||
mapping = discriminator.get("mapping", {})
|
||||
|
||||
# If the field itself is a BaseModel, process it
|
||||
elif isinstance(field_type, type) and issubclass(field_type, BaseModel):
|
||||
nested = process_model_fields(field_type)
|
||||
if nested:
|
||||
section_translations["properties"] = nested
|
||||
type_translations: Dict[str, Any] = {}
|
||||
nested_field_keys: set[str] = set()
|
||||
for detector_type, ref in mapping.items():
|
||||
if not isinstance(ref, str):
|
||||
continue
|
||||
|
||||
return section_translations
|
||||
if not ref.startswith("#/$defs/"):
|
||||
continue
|
||||
|
||||
ref_name = ref.split("/")[-1]
|
||||
ref_schema = defs.get(ref_name, {})
|
||||
if not ref_schema:
|
||||
continue
|
||||
|
||||
type_entry: Dict[str, str] = {}
|
||||
title = ref_schema.get("title")
|
||||
description = ref_schema.get("description")
|
||||
if title:
|
||||
type_entry["label"] = title
|
||||
if description:
|
||||
type_entry["description"] = description
|
||||
|
||||
nested = extract_translations_from_schema(ref_schema, defs=defs)
|
||||
nested_without_root = {
|
||||
k: v for k, v in nested.items() if k not in ("label", "description")
|
||||
}
|
||||
if nested_without_root:
|
||||
type_entry.update(nested_without_root)
|
||||
nested_field_keys.update(nested_without_root.keys())
|
||||
|
||||
if type_entry:
|
||||
type_translations[detector_type] = type_entry
|
||||
|
||||
return type_translations, nested_field_keys
|
||||
|
||||
|
||||
def main():
|
||||
"""Main function to generate config translations."""
|
||||
|
||||
# Define output directory
|
||||
output_dir = Path(__file__).parent / "web" / "public" / "locales" / "en" / "config"
|
||||
if len(sys.argv) > 1:
|
||||
output_dir = Path(sys.argv[1])
|
||||
else:
|
||||
output_dir = (
|
||||
Path(__file__).parent / "web" / "public" / "locales" / "en" / "config"
|
||||
)
|
||||
|
||||
logger.info(f"Output directory: {output_dir}")
|
||||
|
||||
# Clean and recreate the output directory
|
||||
if output_dir.exists():
|
||||
logger.info(f"Removing existing directory: {output_dir}")
|
||||
shutil.rmtree(output_dir)
|
||||
|
||||
logger.info(f"Creating directory: {output_dir}")
|
||||
# Ensure the output directory exists; do not delete existing files.
|
||||
output_dir.mkdir(parents=True, exist_ok=True)
|
||||
logger.info(
|
||||
f"Using output directory (existing files will be overwritten): {output_dir}"
|
||||
)
|
||||
|
||||
config_fields = FrigateConfig.model_fields
|
||||
config_schema = get_config_schema(FrigateConfig)
|
||||
logger.info(f"Found {len(config_fields)} top-level config sections")
|
||||
|
||||
global_translations = {}
|
||||
|
||||
for field_name, field_info in config_fields.items():
|
||||
if field_name.startswith("_"):
|
||||
continue
|
||||
|
||||
logger.info(f"Processing section: {field_name}")
|
||||
section_data = generate_section_translation(field_name, field_info)
|
||||
|
||||
# Get the field's type
|
||||
field_type = field_info.annotation
|
||||
from typing import Optional, Union
|
||||
|
||||
origin = get_origin(field_type)
|
||||
if (
|
||||
origin is Optional
|
||||
or origin is Union
|
||||
or (
|
||||
hasattr(origin, "__name__")
|
||||
and origin.__name__ in ("UnionType", "Union")
|
||||
)
|
||||
):
|
||||
args = get_args(field_type)
|
||||
field_type = next(
|
||||
(arg for arg in args if arg is not type(None)), field_type
|
||||
)
|
||||
|
||||
# Handle Dict[str, SomeModel] - extract the value type
|
||||
if origin is dict:
|
||||
args = get_args(field_type)
|
||||
if args and len(args) > 1:
|
||||
field_type = args[1] # Get value type from Dict[key, value]
|
||||
|
||||
# Start with field's top-level metadata (label, description)
|
||||
section_data = get_field_translations(field_info)
|
||||
|
||||
# Generate nested translations from the field type's schema
|
||||
if hasattr(field_type, "model_json_schema"):
|
||||
schema = field_type.model_json_schema()
|
||||
# Extract nested properties from schema
|
||||
nested = extract_translations_from_schema(schema)
|
||||
# Remove top-level label/description from nested since we got those from field_info
|
||||
nested_without_root = {
|
||||
k: v for k, v in nested.items() if k not in ("label", "description")
|
||||
}
|
||||
section_data.update(nested_without_root)
|
||||
|
||||
if field_name == "detectors":
|
||||
detector_types, detector_field_keys = get_detector_translations(
|
||||
config_schema
|
||||
)
|
||||
section_data.update(detector_types)
|
||||
for key in detector_field_keys:
|
||||
if key == "type":
|
||||
continue
|
||||
section_data.pop(key, None)
|
||||
|
||||
if not section_data:
|
||||
logger.warning(f"No translations found for section: {field_name}")
|
||||
continue
|
||||
|
||||
output_file = output_dir / f"{field_name}.json"
|
||||
with open(output_file, "w", encoding="utf-8") as f:
|
||||
json.dump(section_data, f, indent=2, ensure_ascii=False)
|
||||
# Add camera-level fields to global config documentation if applicable
|
||||
CAMERA_LEVEL_FIELDS = {
|
||||
"birdseye": (
|
||||
"frigate.config.camera.birdseye",
|
||||
"BirdseyeCameraConfig",
|
||||
["order"],
|
||||
),
|
||||
"ffmpeg": (
|
||||
"frigate.config.camera.ffmpeg",
|
||||
"CameraFfmpegConfig",
|
||||
["inputs"],
|
||||
),
|
||||
"lpr": (
|
||||
"frigate.config.classification",
|
||||
"CameraLicensePlateRecognitionConfig",
|
||||
["expire_time"],
|
||||
),
|
||||
"semantic_search": (
|
||||
"frigate.config.classification",
|
||||
"CameraSemanticSearchConfig",
|
||||
["triggers"],
|
||||
),
|
||||
}
|
||||
|
||||
logger.info(f"Generated: {output_file}")
|
||||
if field_name in CAMERA_LEVEL_FIELDS:
|
||||
module_path, class_name, field_names = CAMERA_LEVEL_FIELDS[field_name]
|
||||
try:
|
||||
import importlib
|
||||
|
||||
module = importlib.import_module(module_path)
|
||||
camera_class = getattr(module, class_name)
|
||||
schema = camera_class.model_json_schema()
|
||||
camera_fields = schema.get("properties", {})
|
||||
defs = schema.get("$defs", {})
|
||||
|
||||
for fname in field_names:
|
||||
if fname in camera_fields:
|
||||
field_schema = camera_fields[fname]
|
||||
field_trans = {}
|
||||
if "title" in field_schema:
|
||||
field_trans["label"] = field_schema["title"]
|
||||
if "description" in field_schema:
|
||||
field_trans["description"] = field_schema["description"]
|
||||
|
||||
# Extract nested properties based on schema type
|
||||
nested_to_extract = None
|
||||
|
||||
# Handle direct $ref
|
||||
if "$ref" in field_schema:
|
||||
ref_path = field_schema["$ref"]
|
||||
if ref_path.startswith("#/$defs/"):
|
||||
ref_name = ref_path.split("/")[-1]
|
||||
if ref_name in defs:
|
||||
nested_to_extract = defs[ref_name]
|
||||
|
||||
# Handle additionalProperties with $ref (for dict types)
|
||||
elif "additionalProperties" in field_schema:
|
||||
additional_props = field_schema["additionalProperties"]
|
||||
if "$ref" in additional_props:
|
||||
ref_path = additional_props["$ref"]
|
||||
if ref_path.startswith("#/$defs/"):
|
||||
ref_name = ref_path.split("/")[-1]
|
||||
if ref_name in defs:
|
||||
nested_to_extract = defs[ref_name]
|
||||
|
||||
# Handle items with $ref (for array types)
|
||||
elif "items" in field_schema:
|
||||
items = field_schema["items"]
|
||||
if "$ref" in items:
|
||||
ref_path = items["$ref"]
|
||||
if ref_path.startswith("#/$defs/"):
|
||||
ref_name = ref_path.split("/")[-1]
|
||||
if ref_name in defs:
|
||||
nested_to_extract = defs[ref_name]
|
||||
|
||||
# Extract nested properties if we found a schema to use
|
||||
if nested_to_extract:
|
||||
nested = extract_translations_from_schema(
|
||||
nested_to_extract, defs=defs
|
||||
)
|
||||
nested_without_root = {
|
||||
k: v
|
||||
for k, v in nested.items()
|
||||
if k not in ("label", "description")
|
||||
}
|
||||
field_trans.update(nested_without_root)
|
||||
|
||||
if field_trans:
|
||||
section_data[fname] = field_trans
|
||||
except Exception as e:
|
||||
logger.warning(
|
||||
f"Could not add camera-level fields for {field_name}: {e}"
|
||||
)
|
||||
|
||||
# Add to global translations instead of writing separate files
|
||||
global_translations[field_name] = section_data
|
||||
|
||||
logger.info(f"Added section to global translations: {field_name}")
|
||||
|
||||
# Handle camera-level configs that aren't top-level FrigateConfig fields
|
||||
# These are defined as fields in CameraConfig, so we extract title/description from there
|
||||
camera_level_configs = {
|
||||
"camera_mqtt": ("frigate.config.camera.mqtt", "CameraMqttConfig", "mqtt"),
|
||||
"camera_ui": ("frigate.config.camera.ui", "CameraUiConfig", "ui"),
|
||||
"onvif": ("frigate.config.camera.onvif", "OnvifConfig", "onvif"),
|
||||
}
|
||||
|
||||
# Import CameraConfig to extract field metadata
|
||||
from frigate.config.camera.camera import CameraConfig
|
||||
|
||||
camera_config_schema = CameraConfig.model_json_schema()
|
||||
camera_properties = camera_config_schema.get("properties", {})
|
||||
|
||||
for config_name, (
|
||||
module_path,
|
||||
class_name,
|
||||
camera_field_name,
|
||||
) in camera_level_configs.items():
|
||||
try:
|
||||
logger.info(f"Processing camera-level section: {config_name}")
|
||||
import importlib
|
||||
|
||||
module = importlib.import_module(module_path)
|
||||
config_class = getattr(module, class_name)
|
||||
|
||||
section_data = {}
|
||||
|
||||
# Extract top-level label and description from CameraConfig field definition
|
||||
if camera_field_name in camera_properties:
|
||||
field_schema = camera_properties[camera_field_name]
|
||||
if "title" in field_schema:
|
||||
section_data["label"] = field_schema["title"]
|
||||
if "description" in field_schema:
|
||||
section_data["description"] = field_schema["description"]
|
||||
|
||||
# Process model fields from schema
|
||||
schema = config_class.model_json_schema()
|
||||
nested = extract_translations_from_schema(schema)
|
||||
# Remove top-level label/description since we got those from CameraConfig
|
||||
nested_without_root = {
|
||||
k: v for k, v in nested.items() if k not in ("label", "description")
|
||||
}
|
||||
section_data.update(nested_without_root)
|
||||
|
||||
# Add camera-level section into global translations (do not write separate file)
|
||||
global_translations[config_name] = section_data
|
||||
logger.info(
|
||||
f"Added camera-level section to global translations: {config_name}"
|
||||
)
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to generate {config_name}: {e}")
|
||||
|
||||
# Remove top-level 'cameras' field if present so it remains a separate file
|
||||
if "cameras" in global_translations:
|
||||
logger.info(
|
||||
"Removing top-level 'cameras' from global translations to keep it as a separate cameras.json"
|
||||
)
|
||||
del global_translations["cameras"]
|
||||
|
||||
# Write consolidated global.json with per-section keys
|
||||
global_file = output_dir / "global.json"
|
||||
with open(global_file, "w", encoding="utf-8") as f:
|
||||
json.dump(global_translations, f, indent=2, ensure_ascii=False)
|
||||
f.write("\n")
|
||||
|
||||
logger.info(f"Generated consolidated translations: {global_file}")
|
||||
|
||||
if not global_translations:
|
||||
logger.warning("No global translations were generated!")
|
||||
else:
|
||||
logger.info(f"Global contains {len(global_translations)} sections")
|
||||
|
||||
# Generate cameras.json from CameraConfig schema
|
||||
cameras_file = output_dir / "cameras.json"
|
||||
logger.info(f"Generating cameras.json: {cameras_file}")
|
||||
try:
|
||||
if "camera_config_schema" in locals():
|
||||
camera_schema = camera_config_schema
|
||||
else:
|
||||
from frigate.config.camera.camera import CameraConfig
|
||||
|
||||
camera_schema = CameraConfig.model_json_schema()
|
||||
|
||||
camera_translations = extract_translations_from_schema(camera_schema)
|
||||
|
||||
# Change descriptions to use 'for this camera' for fields that are global
|
||||
def sanitize_camera_descriptions(obj):
|
||||
if isinstance(obj, dict):
|
||||
for k, v in list(obj.items()):
|
||||
if k == "description" and isinstance(v, str):
|
||||
obj[k] = v.replace(
|
||||
"for all cameras; can be overridden per-camera",
|
||||
"for this camera",
|
||||
)
|
||||
else:
|
||||
sanitize_camera_descriptions(v)
|
||||
elif isinstance(obj, list):
|
||||
for item in obj:
|
||||
sanitize_camera_descriptions(item)
|
||||
|
||||
sanitize_camera_descriptions(camera_translations)
|
||||
|
||||
with open(cameras_file, "w", encoding="utf-8") as f:
|
||||
json.dump(camera_translations, f, indent=2, ensure_ascii=False)
|
||||
f.write("\n")
|
||||
logger.info(f"Generated cameras.json: {cameras_file}")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to generate cameras.json: {e}")
|
||||
|
||||
logger.info("Translation generation complete!")
|
||||
|
||||
|
||||
1415
web/package-lock.json
generated
1415
web/package-lock.json
generated
File diff suppressed because it is too large
Load Diff
@ -38,6 +38,10 @@
|
||||
"@radix-ui/react-toggle": "^1.1.2",
|
||||
"@radix-ui/react-toggle-group": "^1.1.2",
|
||||
"@radix-ui/react-tooltip": "^1.2.8",
|
||||
"@rjsf/core": "^6.3.1",
|
||||
"@rjsf/shadcn": "^6.3.1",
|
||||
"@rjsf/utils": "^6.3.1",
|
||||
"@rjsf/validator-ajv8": "^6.3.1",
|
||||
"apexcharts": "^3.52.0",
|
||||
"axios": "^1.7.7",
|
||||
"class-variance-authority": "^0.7.1",
|
||||
|
||||
@ -115,8 +115,10 @@
|
||||
"internalID": "The Internal ID Frigate uses in the configuration and database"
|
||||
},
|
||||
"button": {
|
||||
"add": "Add",
|
||||
"apply": "Apply",
|
||||
"reset": "Reset",
|
||||
"undo": "Undo",
|
||||
"done": "Done",
|
||||
"enabled": "Enabled",
|
||||
"enable": "Enable",
|
||||
@ -150,7 +152,14 @@
|
||||
"export": "Export",
|
||||
"deleteNow": "Delete Now",
|
||||
"next": "Next",
|
||||
"continue": "Continue"
|
||||
"continue": "Continue",
|
||||
"modified": "Modified",
|
||||
"overridden": "Overridden",
|
||||
"resetToGlobal": "Reset to Global",
|
||||
"resetToDefault": "Reset to Default",
|
||||
"saveAll": "Save All",
|
||||
"savingAll": "Saving All…",
|
||||
"undoAll": "Undo All"
|
||||
},
|
||||
"menu": {
|
||||
"system": "System",
|
||||
|
||||
@ -1,26 +0,0 @@
|
||||
{
|
||||
"label": "Global Audio events configuration.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable audio events."
|
||||
},
|
||||
"max_not_heard": {
|
||||
"label": "Seconds of not hearing the type of audio to end the event."
|
||||
},
|
||||
"min_volume": {
|
||||
"label": "Min volume required to run audio detection."
|
||||
},
|
||||
"listen": {
|
||||
"label": "Audio to listen for."
|
||||
},
|
||||
"filters": {
|
||||
"label": "Audio filters."
|
||||
},
|
||||
"enabled_in_config": {
|
||||
"label": "Keep track of original state of audio detection."
|
||||
},
|
||||
"num_threads": {
|
||||
"label": "Number of detection threads"
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,23 +0,0 @@
|
||||
{
|
||||
"label": "Audio transcription config.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable audio transcription."
|
||||
},
|
||||
"language": {
|
||||
"label": "Language abbreviation to use for audio event transcription/translation."
|
||||
},
|
||||
"device": {
|
||||
"label": "The device used for license plate recognition."
|
||||
},
|
||||
"model_size": {
|
||||
"label": "The size of the embeddings model used."
|
||||
},
|
||||
"enabled_in_config": {
|
||||
"label": "Keep track of original state of camera."
|
||||
},
|
||||
"live_enabled": {
|
||||
"label": "Enable live transcriptions."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,35 +0,0 @@
|
||||
{
|
||||
"label": "Auth configuration.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable authentication"
|
||||
},
|
||||
"reset_admin_password": {
|
||||
"label": "Reset the admin password on startup"
|
||||
},
|
||||
"cookie_name": {
|
||||
"label": "Name for jwt token cookie"
|
||||
},
|
||||
"cookie_secure": {
|
||||
"label": "Set secure flag on cookie"
|
||||
},
|
||||
"session_length": {
|
||||
"label": "Session length for jwt session tokens"
|
||||
},
|
||||
"refresh_time": {
|
||||
"label": "Refresh the session if it is going to expire in this many seconds"
|
||||
},
|
||||
"failed_login_rate_limit": {
|
||||
"label": "Rate limits for failed login attempts."
|
||||
},
|
||||
"trusted_proxies": {
|
||||
"label": "Trusted proxies for determining IP address to rate limit"
|
||||
},
|
||||
"hash_iterations": {
|
||||
"label": "Password hash iterations"
|
||||
},
|
||||
"roles": {
|
||||
"label": "Role to camera mappings. Empty list grants access to all cameras."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,37 +0,0 @@
|
||||
{
|
||||
"label": "Birdseye configuration.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable birdseye view."
|
||||
},
|
||||
"mode": {
|
||||
"label": "Tracking mode."
|
||||
},
|
||||
"restream": {
|
||||
"label": "Restream birdseye via RTSP."
|
||||
},
|
||||
"width": {
|
||||
"label": "Birdseye width."
|
||||
},
|
||||
"height": {
|
||||
"label": "Birdseye height."
|
||||
},
|
||||
"quality": {
|
||||
"label": "Encoding quality."
|
||||
},
|
||||
"inactivity_threshold": {
|
||||
"label": "Birdseye Inactivity Threshold"
|
||||
},
|
||||
"layout": {
|
||||
"label": "Birdseye Layout Config",
|
||||
"properties": {
|
||||
"scaling_factor": {
|
||||
"label": "Birdseye Scaling Factor"
|
||||
},
|
||||
"max_cameras": {
|
||||
"label": "Max cameras"
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,14 +0,0 @@
|
||||
{
|
||||
"label": "Camera group configuration",
|
||||
"properties": {
|
||||
"cameras": {
|
||||
"label": "List of cameras in this group."
|
||||
},
|
||||
"icon": {
|
||||
"label": "Icon that represents camera group."
|
||||
},
|
||||
"order": {
|
||||
"label": "Sort order for group."
|
||||
}
|
||||
}
|
||||
}
|
||||
File diff suppressed because it is too large
Load Diff
@ -1,58 +0,0 @@
|
||||
{
|
||||
"label": "Object classification config.",
|
||||
"properties": {
|
||||
"bird": {
|
||||
"label": "Bird classification config.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable bird classification."
|
||||
},
|
||||
"threshold": {
|
||||
"label": "Minimum classification score required to be considered a match."
|
||||
}
|
||||
}
|
||||
},
|
||||
"custom": {
|
||||
"label": "Custom Classification Model Configs.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable running the model."
|
||||
},
|
||||
"name": {
|
||||
"label": "Name of classification model."
|
||||
},
|
||||
"threshold": {
|
||||
"label": "Classification score threshold to change the state."
|
||||
},
|
||||
"object_config": {
|
||||
"properties": {
|
||||
"objects": {
|
||||
"label": "Object types to classify."
|
||||
},
|
||||
"classification_type": {
|
||||
"label": "Type of classification that is applied."
|
||||
}
|
||||
}
|
||||
},
|
||||
"state_config": {
|
||||
"properties": {
|
||||
"cameras": {
|
||||
"label": "Cameras to run classification on.",
|
||||
"properties": {
|
||||
"crop": {
|
||||
"label": "Crop of image frame on this camera to run classification on."
|
||||
}
|
||||
}
|
||||
},
|
||||
"motion": {
|
||||
"label": "If classification should be run when motion is detected in the crop."
|
||||
},
|
||||
"interval": {
|
||||
"label": "Interval to run classification on in seconds."
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,8 +0,0 @@
|
||||
{
|
||||
"label": "Database configuration.",
|
||||
"properties": {
|
||||
"path": {
|
||||
"label": "Database path."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,51 +0,0 @@
|
||||
{
|
||||
"label": "Global object tracking configuration.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Detection Enabled."
|
||||
},
|
||||
"height": {
|
||||
"label": "Height of the stream for the detect role."
|
||||
},
|
||||
"width": {
|
||||
"label": "Width of the stream for the detect role."
|
||||
},
|
||||
"fps": {
|
||||
"label": "Number of frames per second to process through detection."
|
||||
},
|
||||
"min_initialized": {
|
||||
"label": "Minimum number of consecutive hits for an object to be initialized by the tracker."
|
||||
},
|
||||
"max_disappeared": {
|
||||
"label": "Maximum number of frames the object can disappear before detection ends."
|
||||
},
|
||||
"stationary": {
|
||||
"label": "Stationary objects config.",
|
||||
"properties": {
|
||||
"interval": {
|
||||
"label": "Frame interval for checking stationary objects."
|
||||
},
|
||||
"threshold": {
|
||||
"label": "Number of frames without a position change for an object to be considered stationary"
|
||||
},
|
||||
"max_frames": {
|
||||
"label": "Max frames for stationary objects.",
|
||||
"properties": {
|
||||
"default": {
|
||||
"label": "Default max frames."
|
||||
},
|
||||
"objects": {
|
||||
"label": "Object specific max frames."
|
||||
}
|
||||
}
|
||||
},
|
||||
"classifier": {
|
||||
"label": "Enable visual classifier for determing if objects with jittery bounding boxes are stationary."
|
||||
}
|
||||
}
|
||||
},
|
||||
"annotation_offset": {
|
||||
"label": "Milliseconds to offset detect annotations by."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,14 +0,0 @@
|
||||
{
|
||||
"label": "Detector hardware configuration.",
|
||||
"properties": {
|
||||
"type": {
|
||||
"label": "Detector Type"
|
||||
},
|
||||
"model": {
|
||||
"label": "Detector specific model configuration."
|
||||
},
|
||||
"model_path": {
|
||||
"label": "Detector specific model path."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,3 +0,0 @@
|
||||
{
|
||||
"label": "Frigate environment variables."
|
||||
}
|
||||
@ -1,36 +0,0 @@
|
||||
{
|
||||
"label": "Face recognition config.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable face recognition."
|
||||
},
|
||||
"model_size": {
|
||||
"label": "The size of the embeddings model used."
|
||||
},
|
||||
"unknown_score": {
|
||||
"label": "Minimum face distance score required to be marked as a potential match."
|
||||
},
|
||||
"detection_threshold": {
|
||||
"label": "Minimum face detection score required to be considered a face."
|
||||
},
|
||||
"recognition_threshold": {
|
||||
"label": "Minimum face distance score required to be considered a match."
|
||||
},
|
||||
"min_area": {
|
||||
"label": "Min area of face box to consider running face recognition."
|
||||
},
|
||||
"min_faces": {
|
||||
"label": "Min face recognitions for the sub label to be applied to the person object."
|
||||
},
|
||||
"save_attempts": {
|
||||
"label": "Number of face attempts to save in the recent recognitions tab."
|
||||
},
|
||||
"blur_confidence_filter": {
|
||||
"label": "Apply blur quality filter to face confidence."
|
||||
},
|
||||
"device": {
|
||||
"label": "The device key to use for face recognition.",
|
||||
"description": "This is an override, to target a specific device. See https://onnxruntime.ai/docs/execution-providers/ for more information"
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,34 +0,0 @@
|
||||
{
|
||||
"label": "Global FFmpeg configuration.",
|
||||
"properties": {
|
||||
"path": {
|
||||
"label": "FFmpeg path"
|
||||
},
|
||||
"global_args": {
|
||||
"label": "Global FFmpeg arguments."
|
||||
},
|
||||
"hwaccel_args": {
|
||||
"label": "FFmpeg hardware acceleration arguments."
|
||||
},
|
||||
"input_args": {
|
||||
"label": "FFmpeg input arguments."
|
||||
},
|
||||
"output_args": {
|
||||
"label": "FFmpeg output arguments per role.",
|
||||
"properties": {
|
||||
"detect": {
|
||||
"label": "Detect role FFmpeg output arguments."
|
||||
},
|
||||
"record": {
|
||||
"label": "Record role FFmpeg output arguments."
|
||||
}
|
||||
}
|
||||
},
|
||||
"retry_interval": {
|
||||
"label": "Time in seconds to wait before FFmpeg retries connecting to the camera."
|
||||
},
|
||||
"apple_compatibility": {
|
||||
"label": "Set tag on HEVC (H.265) recording stream to improve compatibility with Apple players."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,23 +0,0 @@
|
||||
{
|
||||
"label": "Generative AI configuration.",
|
||||
"properties": {
|
||||
"api_key": {
|
||||
"label": "Provider API key."
|
||||
},
|
||||
"base_url": {
|
||||
"label": "Provider base url."
|
||||
},
|
||||
"model": {
|
||||
"label": "GenAI model."
|
||||
},
|
||||
"provider": {
|
||||
"label": "GenAI provider."
|
||||
},
|
||||
"provider_options": {
|
||||
"label": "GenAI Provider extra options."
|
||||
},
|
||||
"runtime_options": {
|
||||
"label": "Options to pass during inference calls."
|
||||
}
|
||||
}
|
||||
}
|
||||
2185
web/public/locales/en/config/global.json
Normal file
2185
web/public/locales/en/config/global.json
Normal file
File diff suppressed because it is too large
Load Diff
@ -1,3 +0,0 @@
|
||||
{
|
||||
"label": "Global restream configuration."
|
||||
}
|
||||
73
web/public/locales/en/config/groups.json
Normal file
73
web/public/locales/en/config/groups.json
Normal file
@ -0,0 +1,73 @@
|
||||
{
|
||||
"audio": {
|
||||
"global": {
|
||||
"detection": "Global Detection",
|
||||
"sensitivity": "Global Sensitivity"
|
||||
},
|
||||
"cameras": {
|
||||
"detection": "Detection",
|
||||
"sensitivity": "Sensitivity"
|
||||
}
|
||||
},
|
||||
"timestamp_style": {
|
||||
"global": {
|
||||
"appearance": "Global Appearance"
|
||||
},
|
||||
"cameras": {
|
||||
"appearance": "Appearance"
|
||||
}
|
||||
},
|
||||
"motion": {
|
||||
"global": {
|
||||
"sensitivity": "Global Sensitivity",
|
||||
"algorithm": "Global Algorithm"
|
||||
},
|
||||
"cameras": {
|
||||
"sensitivity": "Sensitivity",
|
||||
"algorithm": "Algorithm"
|
||||
}
|
||||
},
|
||||
"snapshots": {
|
||||
"global": {
|
||||
"display": "Global Display"
|
||||
},
|
||||
"cameras": {
|
||||
"display": "Display"
|
||||
}
|
||||
},
|
||||
"detect": {
|
||||
"global": {
|
||||
"resolution": "Global Resolution",
|
||||
"tracking": "Global Tracking"
|
||||
},
|
||||
"cameras": {
|
||||
"resolution": "Resolution",
|
||||
"tracking": "Tracking"
|
||||
}
|
||||
},
|
||||
"objects": {
|
||||
"global": {
|
||||
"tracking": "Global Tracking",
|
||||
"filtering": "Global Filtering"
|
||||
},
|
||||
"cameras": {
|
||||
"tracking": "Tracking",
|
||||
"filtering": "Filtering"
|
||||
}
|
||||
},
|
||||
"record": {
|
||||
"global": {
|
||||
"retention": "Global Retention",
|
||||
"events": "Global Events"
|
||||
},
|
||||
"cameras": {
|
||||
"retention": "Retention",
|
||||
"events": "Events"
|
||||
}
|
||||
},
|
||||
"ffmpeg": {
|
||||
"cameras": {
|
||||
"cameraFfmpeg": "Camera-specific FFmpeg arguments"
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,14 +0,0 @@
|
||||
{
|
||||
"label": "Live playback settings.",
|
||||
"properties": {
|
||||
"streams": {
|
||||
"label": "Friendly names and restream names to use for live view."
|
||||
},
|
||||
"height": {
|
||||
"label": "Live camera view height"
|
||||
},
|
||||
"quality": {
|
||||
"label": "Live camera view quality"
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,11 +0,0 @@
|
||||
{
|
||||
"label": "Logging configuration.",
|
||||
"properties": {
|
||||
"default": {
|
||||
"label": "Default logging level."
|
||||
},
|
||||
"logs": {
|
||||
"label": "Log level for specified processes."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,45 +0,0 @@
|
||||
{
|
||||
"label": "License Plate recognition config.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable license plate recognition."
|
||||
},
|
||||
"model_size": {
|
||||
"label": "The size of the embeddings model used."
|
||||
},
|
||||
"detection_threshold": {
|
||||
"label": "License plate object confidence score required to begin running recognition."
|
||||
},
|
||||
"min_area": {
|
||||
"label": "Minimum area of license plate to begin running recognition."
|
||||
},
|
||||
"recognition_threshold": {
|
||||
"label": "Recognition confidence score required to add the plate to the object as a sub label."
|
||||
},
|
||||
"min_plate_length": {
|
||||
"label": "Minimum number of characters a license plate must have to be added to the object as a sub label."
|
||||
},
|
||||
"format": {
|
||||
"label": "Regular expression for the expected format of license plate."
|
||||
},
|
||||
"match_distance": {
|
||||
"label": "Allow this number of missing/incorrect characters to still cause a detected plate to match a known plate."
|
||||
},
|
||||
"known_plates": {
|
||||
"label": "Known plates to track (strings or regular expressions)."
|
||||
},
|
||||
"enhancement": {
|
||||
"label": "Amount of contrast adjustment and denoising to apply to license plate images before recognition."
|
||||
},
|
||||
"debug_save_plates": {
|
||||
"label": "Save plates captured for LPR for debugging purposes."
|
||||
},
|
||||
"device": {
|
||||
"label": "The device key to use for LPR.",
|
||||
"description": "This is an override, to target a specific device. See https://onnxruntime.ai/docs/execution-providers/ for more information"
|
||||
},
|
||||
"replace_rules": {
|
||||
"label": "List of regex replacement rules for normalizing detected plates. Each rule has 'pattern' and 'replacement'."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,35 +0,0 @@
|
||||
{
|
||||
"label": "Detection model configuration.",
|
||||
"properties": {
|
||||
"path": {
|
||||
"label": "Custom Object detection model path."
|
||||
},
|
||||
"labelmap_path": {
|
||||
"label": "Label map for custom object detector."
|
||||
},
|
||||
"width": {
|
||||
"label": "Object detection model input width."
|
||||
},
|
||||
"height": {
|
||||
"label": "Object detection model input height."
|
||||
},
|
||||
"labelmap": {
|
||||
"label": "Labelmap customization."
|
||||
},
|
||||
"attributes_map": {
|
||||
"label": "Map of object labels to their attribute labels."
|
||||
},
|
||||
"input_tensor": {
|
||||
"label": "Model Input Tensor Shape"
|
||||
},
|
||||
"input_pixel_format": {
|
||||
"label": "Model Input Pixel Color Format"
|
||||
},
|
||||
"input_dtype": {
|
||||
"label": "Model Input D Type"
|
||||
},
|
||||
"model_type": {
|
||||
"label": "Object Detection Model Type"
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,3 +0,0 @@
|
||||
{
|
||||
"label": "Global motion detection configuration."
|
||||
}
|
||||
@ -1,44 +0,0 @@
|
||||
{
|
||||
"label": "MQTT configuration.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable MQTT Communication."
|
||||
},
|
||||
"host": {
|
||||
"label": "MQTT Host"
|
||||
},
|
||||
"port": {
|
||||
"label": "MQTT Port"
|
||||
},
|
||||
"topic_prefix": {
|
||||
"label": "MQTT Topic Prefix"
|
||||
},
|
||||
"client_id": {
|
||||
"label": "MQTT Client ID"
|
||||
},
|
||||
"stats_interval": {
|
||||
"label": "MQTT Camera Stats Interval"
|
||||
},
|
||||
"user": {
|
||||
"label": "MQTT Username"
|
||||
},
|
||||
"password": {
|
||||
"label": "MQTT Password"
|
||||
},
|
||||
"tls_ca_certs": {
|
||||
"label": "MQTT TLS CA Certificates"
|
||||
},
|
||||
"tls_client_cert": {
|
||||
"label": "MQTT TLS Client Certificate"
|
||||
},
|
||||
"tls_client_key": {
|
||||
"label": "MQTT TLS Client Key"
|
||||
},
|
||||
"tls_insecure": {
|
||||
"label": "MQTT TLS Insecure"
|
||||
},
|
||||
"qos": {
|
||||
"label": "MQTT QoS"
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,24 +0,0 @@
|
||||
{
|
||||
"label": "Networking configuration",
|
||||
"properties": {
|
||||
"ipv6": {
|
||||
"label": "IPv6 configuration",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable IPv6 for port 5000 and/or 8971"
|
||||
}
|
||||
}
|
||||
},
|
||||
"listen": {
|
||||
"label": "Listening ports configuration",
|
||||
"properties": {
|
||||
"internal": {
|
||||
"label": "Internal listening port for Frigate"
|
||||
},
|
||||
"external": {
|
||||
"label": "External listening port for Frigate"
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,17 +0,0 @@
|
||||
{
|
||||
"label": "Global notification configuration.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable notifications"
|
||||
},
|
||||
"email": {
|
||||
"label": "Email required for push."
|
||||
},
|
||||
"cooldown": {
|
||||
"label": "Cooldown period for notifications (time in seconds)."
|
||||
},
|
||||
"enabled_in_config": {
|
||||
"label": "Keep track of original state of notifications."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,77 +0,0 @@
|
||||
{
|
||||
"label": "Global object configuration.",
|
||||
"properties": {
|
||||
"track": {
|
||||
"label": "Objects to track."
|
||||
},
|
||||
"filters": {
|
||||
"label": "Object filters.",
|
||||
"properties": {
|
||||
"min_area": {
|
||||
"label": "Minimum area of bounding box for object to be counted. Can be pixels (int) or percentage (float between 0.000001 and 0.99)."
|
||||
},
|
||||
"max_area": {
|
||||
"label": "Maximum area of bounding box for object to be counted. Can be pixels (int) or percentage (float between 0.000001 and 0.99)."
|
||||
},
|
||||
"min_ratio": {
|
||||
"label": "Minimum ratio of bounding box's width/height for object to be counted."
|
||||
},
|
||||
"max_ratio": {
|
||||
"label": "Maximum ratio of bounding box's width/height for object to be counted."
|
||||
},
|
||||
"threshold": {
|
||||
"label": "Average detection confidence threshold for object to be counted."
|
||||
},
|
||||
"min_score": {
|
||||
"label": "Minimum detection confidence for object to be counted."
|
||||
},
|
||||
"mask": {
|
||||
"label": "Detection area polygon mask for this filter configuration."
|
||||
}
|
||||
}
|
||||
},
|
||||
"mask": {
|
||||
"label": "Object mask."
|
||||
},
|
||||
"genai": {
|
||||
"label": "Config for using genai to analyze objects.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable GenAI for camera."
|
||||
},
|
||||
"use_snapshot": {
|
||||
"label": "Use snapshots for generating descriptions."
|
||||
},
|
||||
"prompt": {
|
||||
"label": "Default caption prompt."
|
||||
},
|
||||
"object_prompts": {
|
||||
"label": "Object specific prompts."
|
||||
},
|
||||
"objects": {
|
||||
"label": "List of objects to run generative AI for."
|
||||
},
|
||||
"required_zones": {
|
||||
"label": "List of required zones to be entered in order to run generative AI."
|
||||
},
|
||||
"debug_save_thumbnails": {
|
||||
"label": "Save thumbnails sent to generative AI for debugging purposes."
|
||||
},
|
||||
"send_triggers": {
|
||||
"label": "What triggers to use to send frames to generative AI for a tracked object.",
|
||||
"properties": {
|
||||
"tracked_object_end": {
|
||||
"label": "Send once the object is no longer tracked."
|
||||
},
|
||||
"after_significant_updates": {
|
||||
"label": "Send an early request to generative AI when X frames accumulated."
|
||||
}
|
||||
}
|
||||
},
|
||||
"enabled_in_config": {
|
||||
"label": "Keep track of original state of generative AI."
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,31 +0,0 @@
|
||||
{
|
||||
"label": "Proxy configuration.",
|
||||
"properties": {
|
||||
"header_map": {
|
||||
"label": "Header mapping definitions for proxy user passing.",
|
||||
"properties": {
|
||||
"user": {
|
||||
"label": "Header name from upstream proxy to identify user."
|
||||
},
|
||||
"role": {
|
||||
"label": "Header name from upstream proxy to identify user role."
|
||||
},
|
||||
"role_map": {
|
||||
"label": "Mapping of Frigate roles to upstream group values. "
|
||||
}
|
||||
}
|
||||
},
|
||||
"logout_url": {
|
||||
"label": "Redirect url for logging out with proxy."
|
||||
},
|
||||
"auth_secret": {
|
||||
"label": "Secret value for proxy authentication."
|
||||
},
|
||||
"default_role": {
|
||||
"label": "Default role for proxy users."
|
||||
},
|
||||
"separator": {
|
||||
"label": "The character used to separate values in a mapped header."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,90 +0,0 @@
|
||||
{
|
||||
"label": "Global record configuration.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable record on all cameras."
|
||||
},
|
||||
"expire_interval": {
|
||||
"label": "Number of minutes to wait between cleanup runs."
|
||||
},
|
||||
"continuous": {
|
||||
"label": "Continuous recording retention settings.",
|
||||
"properties": {
|
||||
"days": {
|
||||
"label": "Default retention period."
|
||||
}
|
||||
}
|
||||
},
|
||||
"motion": {
|
||||
"label": "Motion recording retention settings.",
|
||||
"properties": {
|
||||
"days": {
|
||||
"label": "Default retention period."
|
||||
}
|
||||
}
|
||||
},
|
||||
"detections": {
|
||||
"label": "Detection specific retention settings.",
|
||||
"properties": {
|
||||
"pre_capture": {
|
||||
"label": "Seconds to retain before event starts."
|
||||
},
|
||||
"post_capture": {
|
||||
"label": "Seconds to retain after event ends."
|
||||
},
|
||||
"retain": {
|
||||
"label": "Event retention settings.",
|
||||
"properties": {
|
||||
"days": {
|
||||
"label": "Default retention period."
|
||||
},
|
||||
"mode": {
|
||||
"label": "Retain mode."
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
},
|
||||
"alerts": {
|
||||
"label": "Alert specific retention settings.",
|
||||
"properties": {
|
||||
"pre_capture": {
|
||||
"label": "Seconds to retain before event starts."
|
||||
},
|
||||
"post_capture": {
|
||||
"label": "Seconds to retain after event ends."
|
||||
},
|
||||
"retain": {
|
||||
"label": "Event retention settings.",
|
||||
"properties": {
|
||||
"days": {
|
||||
"label": "Default retention period."
|
||||
},
|
||||
"mode": {
|
||||
"label": "Retain mode."
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
},
|
||||
"export": {
|
||||
"label": "Recording Export Config",
|
||||
"properties": {
|
||||
"timelapse_args": {
|
||||
"label": "Timelapse Args"
|
||||
}
|
||||
}
|
||||
},
|
||||
"preview": {
|
||||
"label": "Recording Preview Config",
|
||||
"properties": {
|
||||
"quality": {
|
||||
"label": "Quality of recording preview."
|
||||
}
|
||||
}
|
||||
},
|
||||
"enabled_in_config": {
|
||||
"label": "Keep track of original state of recording."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,74 +0,0 @@
|
||||
{
|
||||
"label": "Review configuration.",
|
||||
"properties": {
|
||||
"alerts": {
|
||||
"label": "Review alerts config.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable alerts."
|
||||
},
|
||||
"labels": {
|
||||
"label": "Labels to create alerts for."
|
||||
},
|
||||
"required_zones": {
|
||||
"label": "List of required zones to be entered in order to save the event as an alert."
|
||||
},
|
||||
"enabled_in_config": {
|
||||
"label": "Keep track of original state of alerts."
|
||||
},
|
||||
"cutoff_time": {
|
||||
"label": "Time to cutoff alerts after no alert-causing activity has occurred."
|
||||
}
|
||||
}
|
||||
},
|
||||
"detections": {
|
||||
"label": "Review detections config.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable detections."
|
||||
},
|
||||
"labels": {
|
||||
"label": "Labels to create detections for."
|
||||
},
|
||||
"required_zones": {
|
||||
"label": "List of required zones to be entered in order to save the event as a detection."
|
||||
},
|
||||
"cutoff_time": {
|
||||
"label": "Time to cutoff detection after no detection-causing activity has occurred."
|
||||
},
|
||||
"enabled_in_config": {
|
||||
"label": "Keep track of original state of detections."
|
||||
}
|
||||
}
|
||||
},
|
||||
"genai": {
|
||||
"label": "Review description genai config.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable GenAI descriptions for review items."
|
||||
},
|
||||
"alerts": {
|
||||
"label": "Enable GenAI for alerts."
|
||||
},
|
||||
"detections": {
|
||||
"label": "Enable GenAI for detections."
|
||||
},
|
||||
"additional_concerns": {
|
||||
"label": "Additional concerns that GenAI should make note of on this camera."
|
||||
},
|
||||
"debug_save_thumbnails": {
|
||||
"label": "Save thumbnails sent to generative AI for debugging purposes."
|
||||
},
|
||||
"enabled_in_config": {
|
||||
"label": "Keep track of original state of generative AI."
|
||||
},
|
||||
"preferred_language": {
|
||||
"label": "Preferred language for GenAI Response"
|
||||
},
|
||||
"activity_context_prompt": {
|
||||
"label": "Custom activity context prompt defining normal activity patterns for this property."
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,3 +0,0 @@
|
||||
{
|
||||
"label": "If Frigate should be started in safe mode."
|
||||
}
|
||||
@ -1,21 +0,0 @@
|
||||
{
|
||||
"label": "Semantic search configuration.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable semantic search."
|
||||
},
|
||||
"reindex": {
|
||||
"label": "Reindex all tracked objects on startup."
|
||||
},
|
||||
"model": {
|
||||
"label": "The CLIP model to use for semantic search."
|
||||
},
|
||||
"model_size": {
|
||||
"label": "The size of the embeddings model used."
|
||||
},
|
||||
"device": {
|
||||
"label": "The device key to use for semantic search.",
|
||||
"description": "This is an override, to target a specific device. See https://onnxruntime.ai/docs/execution-providers/ for more information"
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,43 +0,0 @@
|
||||
{
|
||||
"label": "Global snapshots configuration.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Snapshots enabled."
|
||||
},
|
||||
"clean_copy": {
|
||||
"label": "Create a clean copy of the snapshot image."
|
||||
},
|
||||
"timestamp": {
|
||||
"label": "Add a timestamp overlay on the snapshot."
|
||||
},
|
||||
"bounding_box": {
|
||||
"label": "Add a bounding box overlay on the snapshot."
|
||||
},
|
||||
"crop": {
|
||||
"label": "Crop the snapshot to the detected object."
|
||||
},
|
||||
"required_zones": {
|
||||
"label": "List of required zones to be entered in order to save a snapshot."
|
||||
},
|
||||
"height": {
|
||||
"label": "Snapshot image height."
|
||||
},
|
||||
"retain": {
|
||||
"label": "Snapshot retention.",
|
||||
"properties": {
|
||||
"default": {
|
||||
"label": "Default retention period."
|
||||
},
|
||||
"mode": {
|
||||
"label": "Retain mode."
|
||||
},
|
||||
"objects": {
|
||||
"label": "Object retention period."
|
||||
}
|
||||
}
|
||||
},
|
||||
"quality": {
|
||||
"label": "Quality of the encoded jpeg (0-100)."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,28 +0,0 @@
|
||||
{
|
||||
"label": "Telemetry configuration.",
|
||||
"properties": {
|
||||
"network_interfaces": {
|
||||
"label": "Enabled network interfaces for bandwidth calculation."
|
||||
},
|
||||
"stats": {
|
||||
"label": "System Stats Configuration",
|
||||
"properties": {
|
||||
"amd_gpu_stats": {
|
||||
"label": "Enable AMD GPU stats."
|
||||
},
|
||||
"intel_gpu_stats": {
|
||||
"label": "Enable Intel GPU stats."
|
||||
},
|
||||
"network_bandwidth": {
|
||||
"label": "Enable network bandwidth for ffmpeg processes."
|
||||
},
|
||||
"intel_gpu_device": {
|
||||
"label": "Define the device to use when gathering SR-IOV stats."
|
||||
}
|
||||
}
|
||||
},
|
||||
"version_check": {
|
||||
"label": "Enable latest version check."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,31 +0,0 @@
|
||||
{
|
||||
"label": "Global timestamp style configuration.",
|
||||
"properties": {
|
||||
"position": {
|
||||
"label": "Timestamp position."
|
||||
},
|
||||
"format": {
|
||||
"label": "Timestamp format."
|
||||
},
|
||||
"color": {
|
||||
"label": "Timestamp color.",
|
||||
"properties": {
|
||||
"red": {
|
||||
"label": "Red"
|
||||
},
|
||||
"green": {
|
||||
"label": "Green"
|
||||
},
|
||||
"blue": {
|
||||
"label": "Blue"
|
||||
}
|
||||
}
|
||||
},
|
||||
"thickness": {
|
||||
"label": "Timestamp thickness."
|
||||
},
|
||||
"effect": {
|
||||
"label": "Timestamp effect."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,8 +0,0 @@
|
||||
{
|
||||
"label": "TLS configuration.",
|
||||
"properties": {
|
||||
"enabled": {
|
||||
"label": "Enable TLS for port 8971"
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,20 +0,0 @@
|
||||
{
|
||||
"label": "UI configuration.",
|
||||
"properties": {
|
||||
"timezone": {
|
||||
"label": "Override UI timezone."
|
||||
},
|
||||
"time_format": {
|
||||
"label": "Override UI time format."
|
||||
},
|
||||
"date_style": {
|
||||
"label": "Override UI dateStyle."
|
||||
},
|
||||
"time_style": {
|
||||
"label": "Override UI timeStyle."
|
||||
},
|
||||
"unit_system": {
|
||||
"label": "The unit system to use for measurements."
|
||||
}
|
||||
}
|
||||
}
|
||||
32
web/public/locales/en/config/validation.json
Normal file
32
web/public/locales/en/config/validation.json
Normal file
@ -0,0 +1,32 @@
|
||||
{
|
||||
"minimum": "Must be at least {{limit}}",
|
||||
"maximum": "Must be at most {{limit}}",
|
||||
"exclusiveMinimum": "Must be greater than {{limit}}",
|
||||
"exclusiveMaximum": "Must be less than {{limit}}",
|
||||
"minLength": "Must be at least {{limit}} character(s)",
|
||||
"maxLength": "Must be at most {{limit}} character(s)",
|
||||
"minItems": "Must have at least {{limit}} items",
|
||||
"maxItems": "Must have at most {{limit}} items",
|
||||
"pattern": "Invalid format",
|
||||
"required": "This field is required",
|
||||
"type": "Invalid value type",
|
||||
"enum": "Must be one of the allowed values",
|
||||
"const": "Value does not match expected constant",
|
||||
"uniqueItems": "All items must be unique",
|
||||
"format": "Invalid format",
|
||||
"additionalProperties": "Unknown property is not allowed",
|
||||
"oneOf": "Must match exactly one of the allowed schemas",
|
||||
"anyOf": "Must match at least one of the allowed schemas",
|
||||
"proxy": {
|
||||
"header_map": {
|
||||
"roleHeaderRequired": "Role header is required when role mappings are configured."
|
||||
}
|
||||
},
|
||||
"ffmpeg": {
|
||||
"inputs": {
|
||||
"rolesUnique": "Each role can only be assigned to one input stream.",
|
||||
"detectRequired": "At least one input stream must be assigned the 'detect' role.",
|
||||
"hwaccelDetectOnly": "Only the input stream with the detect role can define hardware acceleration arguments."
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1,3 +0,0 @@
|
||||
{
|
||||
"label": "Current config version."
|
||||
}
|
||||
@ -8,23 +8,82 @@
|
||||
"masksAndZones": "Mask and Zone Editor - Frigate",
|
||||
"motionTuner": "Motion Tuner - Frigate",
|
||||
"object": "Debug - Frigate",
|
||||
"general": "UI Settings - Frigate",
|
||||
"general": "Profile Settings - Frigate",
|
||||
"globalConfig": "Global Configuration - Frigate",
|
||||
"cameraConfig": "Camera Configuration - Frigate",
|
||||
"frigatePlus": "Frigate+ Settings - Frigate",
|
||||
"notifications": "Notification Settings - Frigate"
|
||||
"notifications": "Notification Settings - Frigate",
|
||||
"maintenance": "Maintenance - Frigate"
|
||||
},
|
||||
"menu": {
|
||||
"general": "General",
|
||||
"globalConfig": "Global configuration",
|
||||
"system": "System",
|
||||
"integrations": "Integrations",
|
||||
"cameras": "Camera configuration",
|
||||
"ui": "UI",
|
||||
"enrichments": "Enrichments",
|
||||
"profileSettings": "Profile settings",
|
||||
"globalDetect": "Object detection",
|
||||
"globalRecording": "Recording",
|
||||
"globalSnapshots": "Snapshots",
|
||||
"globalFfmpeg": "FFmpeg",
|
||||
"globalMotion": "Motion detection",
|
||||
"globalObjects": "Objects",
|
||||
"globalReview": "Review",
|
||||
"globalAudioEvents": "Audio events",
|
||||
"globalLivePlayback": "Live playback",
|
||||
"globalTimestampStyle": "Timestamp style",
|
||||
"systemDatabase": "Database",
|
||||
"systemTls": "TLS",
|
||||
"systemAuthentication": "Authentication",
|
||||
"systemNetworking": "Networking",
|
||||
"systemProxy": "Proxy",
|
||||
"systemUi": "UI",
|
||||
"systemLogging": "Logging",
|
||||
"systemEnvironmentVariables": "Environment variables",
|
||||
"systemTelemetry": "Telemetry",
|
||||
"systemBirdseye": "Birdseye",
|
||||
"systemFfmpeg": "FFmpeg",
|
||||
"systemDetectorHardware": "Detector hardware",
|
||||
"systemDetectionModel": "Detection model",
|
||||
"systemMqtt": "MQTT",
|
||||
"integrationSemanticSearch": "Semantic search",
|
||||
"integrationGenerativeAi": "Generative AI",
|
||||
"integrationFaceRecognition": "Face recognition",
|
||||
"integrationLpr": "License plate recognition",
|
||||
"integrationObjectClassification": "Object classification",
|
||||
"integrationAudioTranscription": "Audio transcription",
|
||||
"cameraDetect": "Object detection",
|
||||
"cameraFfmpeg": "FFmpeg",
|
||||
"cameraRecording": "Recording",
|
||||
"cameraSnapshots": "Snapshots",
|
||||
"cameraMotion": "Motion detection",
|
||||
"cameraObjects": "Objects",
|
||||
"cameraConfigReview": "Review",
|
||||
"cameraAudioEvents": "Audio events",
|
||||
"cameraAudioTranscription": "Audio transcription",
|
||||
"cameraNotifications": "Notifications",
|
||||
"cameraLivePlayback": "Live playback",
|
||||
"cameraBirdseye": "Birdseye",
|
||||
"cameraFaceRecognition": "Face recognition",
|
||||
"cameraLpr": "License plate recognition",
|
||||
"cameraMqttConfig": "MQTT",
|
||||
"cameraOnvif": "ONVIF",
|
||||
"cameraUi": "Camera UI",
|
||||
"cameraTimestampStyle": "Timestamp style",
|
||||
"cameraMqtt": "Camera MQTT",
|
||||
"cameraManagement": "Management",
|
||||
"cameraReview": "Review",
|
||||
"masksAndZones": "Masks / Zones",
|
||||
"motionTuner": "Motion Tuner",
|
||||
"triggers": "Triggers",
|
||||
"debug": "Debug",
|
||||
"motionTuner": "Motion tuner",
|
||||
"enrichments": "Enrichments",
|
||||
"users": "Users",
|
||||
"roles": "Roles",
|
||||
"notifications": "Notifications",
|
||||
"frigateplus": "Frigate+"
|
||||
"triggers": "Triggers",
|
||||
"debug": "Debug",
|
||||
"frigateplus": "Frigate+",
|
||||
"maintenance": "Maintenance"
|
||||
},
|
||||
"dialog": {
|
||||
"unsavedChanges": {
|
||||
@ -32,12 +91,29 @@
|
||||
"desc": "Do you want to save your changes before continuing?"
|
||||
}
|
||||
},
|
||||
"saveAllPreview": {
|
||||
"title": "Changes to be saved",
|
||||
"triggerLabel": "Review pending changes",
|
||||
"empty": "No pending changes.",
|
||||
"scope": {
|
||||
"label": "Scope",
|
||||
"global": "Global",
|
||||
"camera": "Camera: {{cameraName}}"
|
||||
},
|
||||
"field": {
|
||||
"label": "Field"
|
||||
},
|
||||
"value": {
|
||||
"label": "New value",
|
||||
"reset": "Reset"
|
||||
}
|
||||
},
|
||||
"cameraSetting": {
|
||||
"camera": "Camera",
|
||||
"noCamera": "No Camera"
|
||||
},
|
||||
"general": {
|
||||
"title": "UI Settings",
|
||||
"title": "Profile Settings",
|
||||
"liveDashboard": {
|
||||
"title": "Live Dashboard",
|
||||
"automaticLiveView": {
|
||||
@ -106,7 +182,7 @@
|
||||
"desc": "Semantic Search in Frigate allows you to find tracked objects within your review items using either the image itself, a user-defined text description, or an automatically generated one.",
|
||||
"reindexNow": {
|
||||
"label": "Reindex Now",
|
||||
"desc": "Reindexing will regenerate embeddings for all tracked object. This process runs in the background and may max out your CPU and take a fair amount of time depending on the number of tracked objects you have.",
|
||||
"desc": "Reindexing will regenerate embeddings for all tracked objects. This process runs in the background and may max out your CPU and take a fair amount of time depending on the number of tracked objects you have.",
|
||||
"confirmTitle": "Confirm Reindexing",
|
||||
"confirmDesc": "Are you sure you want to reindex all tracked object embeddings? This process will run in the background but it may max out your CPU and take a fair amount of time. You can watch the progress on the Explore page.",
|
||||
"confirmButton": "Reindex",
|
||||
@ -350,7 +426,11 @@
|
||||
"backToSettings": "Back to Camera Settings",
|
||||
"streams": {
|
||||
"title": "Enable / Disable Cameras",
|
||||
"desc": "Temporarily disable a camera until Frigate restarts. Disabling a camera completely stops Frigate's processing of this camera's streams. Detection, recording, and debugging will be unavailable.<br /> <em>Note: This does not disable go2rtc restreams.</em>"
|
||||
"enableLabel": "Enabled cameras",
|
||||
"enableDesc": "Temporarily disable an enabled camera until Frigate restarts. Disabling a camera completely stops Frigate's processing of this camera's streams. Detection, recording, and debugging will be unavailable.<br /> <em>Note: This does not disable go2rtc restreams.</em>",
|
||||
"disableLabel": "Disabled cameras",
|
||||
"disableDesc": "Enable a camera that is currently not visible in the UI and disabled in the configuration. A restart of Frigate is required after enabling.",
|
||||
"enableSuccess": "Enabled {{cameraName}} in configuration. Restart Frigate to apply the changes."
|
||||
},
|
||||
"cameraConfig": {
|
||||
"add": "Add Camera",
|
||||
@ -906,6 +986,13 @@
|
||||
},
|
||||
"frigatePlus": {
|
||||
"title": "Frigate+ Settings",
|
||||
"description": "Frigate+ is a subscription service that provides access to additional features and capabilities for your Frigate instance, including the ability to use custom object detection models trained on your own data. You can manage your Frigate+ model settings here.",
|
||||
"cardTitles": {
|
||||
"api": "API",
|
||||
"currentModel": "Current Model",
|
||||
"otherModels": "Other Models",
|
||||
"configuration": "Configuration"
|
||||
},
|
||||
"apiKey": {
|
||||
"title": "Frigate+ API Key",
|
||||
"validated": "Frigate+ API key is detected and validated",
|
||||
@ -947,6 +1034,15 @@
|
||||
"error": "Failed to save config changes: {{errorMessage}}"
|
||||
}
|
||||
},
|
||||
"detectionModel": {
|
||||
"plusActive": {
|
||||
"title": "Frigate+ model management",
|
||||
"label": "Current model source",
|
||||
"description": "This instance is running a Frigate+ model. Select or change your model in Frigate+ settings.",
|
||||
"goToFrigatePlus": "Go to Frigate+ settings",
|
||||
"showModelForm": "Manually configure a model"
|
||||
}
|
||||
},
|
||||
"triggers": {
|
||||
"documentTitle": "Triggers",
|
||||
"semanticSearch": {
|
||||
@ -1115,5 +1211,178 @@
|
||||
"exports": "Exports",
|
||||
"recordings": "Recordings"
|
||||
}
|
||||
}
|
||||
},
|
||||
"configForm": {
|
||||
"global": {
|
||||
"title": "Global Settings",
|
||||
"description": "These settings apply to all cameras unless overridden in the camera-specific settings."
|
||||
},
|
||||
"camera": {
|
||||
"title": "Camera Settings",
|
||||
"description": "These settings apply only to this camera and override the global settings."
|
||||
},
|
||||
"advancedSettingsCount": "Advanced Settings ({{count}})",
|
||||
"advancedCount": "Advanced ({{count}})",
|
||||
"showAdvanced": "Show Advanced Settings",
|
||||
"tabs": {
|
||||
"sharedDefaults": "Shared Defaults",
|
||||
"system": "System",
|
||||
"integrations": "Integrations"
|
||||
},
|
||||
"additionalProperties": {
|
||||
"keyLabel": "Key",
|
||||
"valueLabel": "Value",
|
||||
"keyPlaceholder": "New key",
|
||||
"remove": "Remove"
|
||||
},
|
||||
"timezone": {
|
||||
"defaultOption": "Use browser timezone"
|
||||
},
|
||||
"roleMap": {
|
||||
"empty": "No role mappings",
|
||||
"roleLabel": "Role",
|
||||
"groupsLabel": "Groups",
|
||||
"addMapping": "Add role mapping",
|
||||
"remove": "Remove"
|
||||
},
|
||||
"ffmpegArgs": {
|
||||
"preset": "Preset",
|
||||
"manual": "Manual arguments",
|
||||
"inherit": "Inherit from camera setting",
|
||||
"selectPreset": "Select preset",
|
||||
"manualPlaceholder": "Enter FFmpeg arguments"
|
||||
},
|
||||
"cameraInputs": {
|
||||
"itemTitle": "Stream {{index}}"
|
||||
},
|
||||
"restartRequiredField": "Restart required",
|
||||
"restartRequiredFooter": "Configuration changed - Restart required",
|
||||
"sections": {
|
||||
"detect": "Detection",
|
||||
"record": "Recording",
|
||||
"snapshots": "Snapshots",
|
||||
"motion": "Motion",
|
||||
"objects": "Objects",
|
||||
"review": "Review",
|
||||
"audio": "Audio",
|
||||
"notifications": "Notifications",
|
||||
"live": "Live View",
|
||||
"timestamp_style": "Timestamps",
|
||||
"mqtt": "MQTT",
|
||||
"database": "Database",
|
||||
"telemetry": "Telemetry",
|
||||
"auth": "Authentication",
|
||||
"tls": "TLS",
|
||||
"proxy": "Proxy",
|
||||
"go2rtc": "go2rtc",
|
||||
"ffmpeg": "FFmpeg",
|
||||
"detectors": "Detectors",
|
||||
"model": "Model",
|
||||
"semantic_search": "Semantic Search",
|
||||
"genai": "GenAI",
|
||||
"face_recognition": "Face Recognition",
|
||||
"lpr": "License Plate Recognition",
|
||||
"birdseye": "Birdseye"
|
||||
},
|
||||
"detect": {
|
||||
"title": "Detection Settings"
|
||||
},
|
||||
"detectors": {
|
||||
"title": "Detector Settings",
|
||||
"singleType": "Only one {{type}} detector is allowed.",
|
||||
"keyRequired": "Detector name is required.",
|
||||
"keyDuplicate": "Detector name already exists.",
|
||||
"noSchema": "No detector schemas are available.",
|
||||
"none": "No detector instances configured.",
|
||||
"add": "Add detector"
|
||||
},
|
||||
"record": {
|
||||
"title": "Recording Settings"
|
||||
},
|
||||
"snapshots": {
|
||||
"title": "Snapshot Settings"
|
||||
},
|
||||
"motion": {
|
||||
"title": "Motion Settings"
|
||||
},
|
||||
"objects": {
|
||||
"title": "Object Settings"
|
||||
},
|
||||
"audioLabels": {
|
||||
"summary": "{{count}} audio labels selected",
|
||||
"empty": "No audio labels available"
|
||||
},
|
||||
"objectLabels": {
|
||||
"summary": "{{count}} object types selected",
|
||||
"empty": "No object labels available"
|
||||
},
|
||||
"filters": {
|
||||
"objectFieldLabel": "{{field}} for {{label}}"
|
||||
},
|
||||
"zoneNames": {
|
||||
"summary": "{{count}} selected",
|
||||
"empty": "No zones available"
|
||||
},
|
||||
"inputRoles": {
|
||||
"summary": "{{count}} roles selected",
|
||||
"empty": "No roles available",
|
||||
"options": {
|
||||
"detect": "Detect",
|
||||
"record": "Record",
|
||||
"audio": "Audio"
|
||||
}
|
||||
},
|
||||
"review": {
|
||||
"title": "Review Settings"
|
||||
},
|
||||
"audio": {
|
||||
"title": "Audio Settings"
|
||||
},
|
||||
"notifications": {
|
||||
"title": "Notification Settings"
|
||||
},
|
||||
"live": {
|
||||
"title": "Live View Settings"
|
||||
},
|
||||
"timestamp_style": {
|
||||
"title": "Timestamp Settings"
|
||||
},
|
||||
"searchPlaceholder": "Search..."
|
||||
},
|
||||
"globalConfig": {
|
||||
"title": "Global Configuration",
|
||||
"description": "Configure global settings that apply to all cameras unless overridden.",
|
||||
"toast": {
|
||||
"success": "Global settings saved successfully",
|
||||
"error": "Failed to save global settings",
|
||||
"validationError": "Validation failed"
|
||||
}
|
||||
},
|
||||
"cameraConfig": {
|
||||
"title": "Camera Configuration",
|
||||
"description": "Configure settings for individual cameras. Settings override global defaults.",
|
||||
"overriddenBadge": "Overridden",
|
||||
"resetToGlobal": "Reset to Global",
|
||||
"toast": {
|
||||
"success": "Camera settings saved successfully",
|
||||
"error": "Failed to save camera settings"
|
||||
}
|
||||
},
|
||||
"toast": {
|
||||
"success": "Settings saved successfully",
|
||||
"successRestartRequired": "Settings saved successfully. Restart Frigate to apply your changes.",
|
||||
"error": "Failed to save settings",
|
||||
"validationError": "Validation failed: {{message}}",
|
||||
"resetSuccess": "Reset to global defaults",
|
||||
"resetError": "Failed to reset settings",
|
||||
"saveAllSuccess_one": "Saved {{count}} section successfully.",
|
||||
"saveAllSuccess_other": "All {{count}} sections saved successfully.",
|
||||
"saveAllPartial_one": "{{successCount}} of {{totalCount}} section saved. {{failCount}} failed.",
|
||||
"saveAllPartial_other": "{{successCount}} of {{totalCount}} sections saved. {{failCount}} failed.",
|
||||
"saveAllFailure": "Failed to save all sections."
|
||||
},
|
||||
"unsavedChanges": "You have unsaved changes",
|
||||
"confirmReset": "Confirm Reset",
|
||||
"resetToDefaultDescription": "This will reset all settings in this section to their default values. This action cannot be undone.",
|
||||
"resetToGlobalDescription": "This will reset the settings in this section to the global defaults. This action cannot be undone."
|
||||
}
|
||||
|
||||
56
web/src/components/card/SettingsGroupCard.tsx
Normal file
56
web/src/components/card/SettingsGroupCard.tsx
Normal file
@ -0,0 +1,56 @@
|
||||
import { ReactNode } from "react";
|
||||
import { Label } from "../ui/label";
|
||||
|
||||
export const SPLIT_ROW_CLASS_NAME =
|
||||
"space-y-2 md:grid md:grid-cols-[minmax(14rem,24rem)_minmax(0,1fr)] md:items-start md:gap-x-6 md:space-y-0";
|
||||
export const DESCRIPTION_CLASS_NAME = "text-sm text-muted-foreground";
|
||||
export const CONTROL_COLUMN_CLASS_NAME = "w-full md:max-w-2xl";
|
||||
|
||||
type SettingsGroupCardProps = {
|
||||
title: string | ReactNode;
|
||||
children: ReactNode;
|
||||
};
|
||||
|
||||
export function SettingsGroupCard({ title, children }: SettingsGroupCardProps) {
|
||||
return (
|
||||
<div className="space-y-4 rounded-lg border border-border/70 bg-card/30 p-4">
|
||||
<div className="text-md border-b border-border/60 pb-4 font-semibold text-primary-variant">
|
||||
{title}
|
||||
</div>
|
||||
{children}
|
||||
</div>
|
||||
);
|
||||
}
|
||||
|
||||
type SplitCardRowProps = {
|
||||
label: ReactNode;
|
||||
description?: ReactNode;
|
||||
content: ReactNode;
|
||||
};
|
||||
|
||||
export function SplitCardRow({
|
||||
label,
|
||||
description,
|
||||
content,
|
||||
}: SplitCardRowProps) {
|
||||
return (
|
||||
<div className={SPLIT_ROW_CLASS_NAME}>
|
||||
<div className="space-y-1.5">
|
||||
<Label className="text-sm font-medium">{label}</Label>
|
||||
{description && (
|
||||
<div className={`hidden md:block ${DESCRIPTION_CLASS_NAME}`}>
|
||||
{description}
|
||||
</div>
|
||||
)}
|
||||
</div>
|
||||
<div className={`${CONTROL_COLUMN_CLASS_NAME} space-y-1.5`}>
|
||||
{content}
|
||||
{description && (
|
||||
<div className={`md:hidden ${DESCRIPTION_CLASS_NAME}`}>
|
||||
{description}
|
||||
</div>
|
||||
)}
|
||||
</div>
|
||||
</div>
|
||||
);
|
||||
}
|
||||
370
web/src/components/config-form/ConfigForm.tsx
Normal file
370
web/src/components/config-form/ConfigForm.tsx
Normal file
@ -0,0 +1,370 @@
|
||||
// ConfigForm - Main RJSF form wrapper component
|
||||
import Form from "@rjsf/shadcn";
|
||||
import validator from "@rjsf/validator-ajv8";
|
||||
import type { FormValidation, RJSFSchema, UiSchema } from "@rjsf/utils";
|
||||
import type { IChangeEvent } from "@rjsf/core";
|
||||
import { frigateTheme } from "./theme";
|
||||
import { transformSchema } from "@/lib/config-schema";
|
||||
import { createErrorTransformer } from "@/lib/config-schema/errorMessages";
|
||||
import { useMemo, useCallback } from "react";
|
||||
import { useTranslation } from "react-i18next";
|
||||
import { cn, mergeUiSchema } from "@/lib/utils";
|
||||
import type { ConfigFormContext } from "@/types/configForm";
|
||||
|
||||
type SchemaWithProperties = RJSFSchema & {
|
||||
properties: Record<string, RJSFSchema>;
|
||||
};
|
||||
|
||||
type SchemaWithAdditionalProperties = RJSFSchema & {
|
||||
additionalProperties: RJSFSchema;
|
||||
};
|
||||
|
||||
// Runtime guards for schema fragments
|
||||
const hasSchemaProperties = (
|
||||
schema: RJSFSchema,
|
||||
): schema is SchemaWithProperties =>
|
||||
typeof schema === "object" &&
|
||||
schema !== null &&
|
||||
typeof schema.properties === "object" &&
|
||||
schema.properties !== null;
|
||||
|
||||
const hasSchemaAdditionalProperties = (
|
||||
schema: RJSFSchema,
|
||||
): schema is SchemaWithAdditionalProperties =>
|
||||
typeof schema === "object" &&
|
||||
schema !== null &&
|
||||
typeof schema.additionalProperties === "object" &&
|
||||
schema.additionalProperties !== null;
|
||||
|
||||
// Detects path-style uiSchema keys (e.g., "filters.*.mask")
|
||||
const isPathKey = (key: string) => key.includes(".") || key.includes("*");
|
||||
|
||||
type UiSchemaPathOverride = {
|
||||
path: string[];
|
||||
value: UiSchema;
|
||||
};
|
||||
|
||||
// Split uiSchema into normal keys vs path-based overrides
|
||||
const splitUiSchemaOverrides = (
|
||||
uiSchema?: UiSchema,
|
||||
): { baseUiSchema?: UiSchema; pathOverrides: UiSchemaPathOverride[] } => {
|
||||
if (!uiSchema) {
|
||||
return { baseUiSchema: undefined, pathOverrides: [] };
|
||||
}
|
||||
|
||||
const baseUiSchema: UiSchema = {};
|
||||
const pathOverrides: UiSchemaPathOverride[] = [];
|
||||
|
||||
Object.entries(uiSchema).forEach(([key, value]) => {
|
||||
if (isPathKey(key)) {
|
||||
pathOverrides.push({
|
||||
path: key.split("."),
|
||||
value: value as UiSchema,
|
||||
});
|
||||
} else {
|
||||
baseUiSchema[key] = value as UiSchema;
|
||||
}
|
||||
});
|
||||
|
||||
return { baseUiSchema, pathOverrides };
|
||||
};
|
||||
|
||||
// Apply wildcard path overrides to uiSchema using the schema structure
|
||||
const applyUiSchemaPathOverrides = (
|
||||
uiSchema: UiSchema,
|
||||
schema: RJSFSchema,
|
||||
overrides: UiSchemaPathOverride[],
|
||||
): UiSchema => {
|
||||
if (overrides.length === 0) {
|
||||
return uiSchema;
|
||||
}
|
||||
|
||||
// Recursively apply a path override; supports "*" to match any property.
|
||||
const applyOverride = (
|
||||
targetUi: UiSchema,
|
||||
targetSchema: RJSFSchema,
|
||||
path: string[],
|
||||
value: UiSchema,
|
||||
) => {
|
||||
if (path.length === 0) {
|
||||
Object.assign(targetUi, mergeUiSchema(targetUi, value));
|
||||
return;
|
||||
}
|
||||
|
||||
const [segment, ...rest] = path;
|
||||
const schemaObj = targetSchema;
|
||||
|
||||
if (segment === "*") {
|
||||
if (hasSchemaProperties(schemaObj)) {
|
||||
Object.entries(schemaObj.properties).forEach(
|
||||
([propertyName, propertySchema]) => {
|
||||
const existing =
|
||||
(targetUi[propertyName] as UiSchema | undefined) || {};
|
||||
targetUi[propertyName] = { ...existing };
|
||||
applyOverride(
|
||||
targetUi[propertyName] as UiSchema,
|
||||
propertySchema,
|
||||
rest,
|
||||
value,
|
||||
);
|
||||
},
|
||||
);
|
||||
} else if (hasSchemaAdditionalProperties(schemaObj)) {
|
||||
// For dict schemas, apply override to additionalProperties
|
||||
const existing =
|
||||
(targetUi.additionalProperties as UiSchema | undefined) || {};
|
||||
targetUi.additionalProperties = { ...existing };
|
||||
applyOverride(
|
||||
targetUi.additionalProperties as UiSchema,
|
||||
schemaObj.additionalProperties,
|
||||
rest,
|
||||
value,
|
||||
);
|
||||
}
|
||||
return;
|
||||
}
|
||||
|
||||
if (hasSchemaProperties(schemaObj)) {
|
||||
const propertySchema = schemaObj.properties[segment];
|
||||
if (propertySchema) {
|
||||
const existing = (targetUi[segment] as UiSchema | undefined) || {};
|
||||
targetUi[segment] = { ...existing };
|
||||
applyOverride(
|
||||
targetUi[segment] as UiSchema,
|
||||
propertySchema,
|
||||
rest,
|
||||
value,
|
||||
);
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const updated = { ...uiSchema };
|
||||
overrides.forEach(({ path, value }) => {
|
||||
applyOverride(updated, schema, path, value);
|
||||
});
|
||||
|
||||
return updated;
|
||||
};
|
||||
|
||||
const applyLayoutGridFieldDefaults = (uiSchema: UiSchema): UiSchema => {
|
||||
const applyDefaults = (node: unknown): unknown => {
|
||||
if (Array.isArray(node)) {
|
||||
return node.map((item) => applyDefaults(item));
|
||||
}
|
||||
|
||||
if (typeof node !== "object" || node === null) {
|
||||
return node;
|
||||
}
|
||||
|
||||
const nextNode: Record<string, unknown> = {};
|
||||
|
||||
Object.entries(node).forEach(([key, value]) => {
|
||||
nextNode[key] = applyDefaults(value);
|
||||
});
|
||||
|
||||
if (
|
||||
Array.isArray(nextNode["ui:layoutGrid"]) &&
|
||||
nextNode["ui:field"] === undefined
|
||||
) {
|
||||
nextNode["ui:field"] = "LayoutGridField";
|
||||
}
|
||||
|
||||
return nextNode;
|
||||
};
|
||||
|
||||
return applyDefaults(uiSchema) as UiSchema;
|
||||
};
|
||||
|
||||
export interface ConfigFormProps {
|
||||
/** JSON Schema for the form */
|
||||
schema: RJSFSchema;
|
||||
/** Current form data */
|
||||
formData?: unknown;
|
||||
/** Called when form data changes */
|
||||
onChange?: (data: unknown) => void;
|
||||
/** Called when form is submitted */
|
||||
onSubmit?: (data: unknown) => void;
|
||||
/** Called when form has errors on submit */
|
||||
onError?: (errors: unknown[]) => void;
|
||||
/** Additional uiSchema overrides */
|
||||
uiSchema?: UiSchema;
|
||||
/** Field ordering */
|
||||
fieldOrder?: string[];
|
||||
/** Field groups for layout */
|
||||
fieldGroups?: Record<string, string[]>;
|
||||
/** Fields to hide */
|
||||
hiddenFields?: string[];
|
||||
/** Fields marked as advanced (collapsed by default) */
|
||||
advancedFields?: string[];
|
||||
/** Whether form is disabled */
|
||||
disabled?: boolean;
|
||||
/** Whether form is read-only */
|
||||
readonly?: boolean;
|
||||
/** Whether to show submit button */
|
||||
showSubmit?: boolean;
|
||||
/** Custom class name */
|
||||
className?: string;
|
||||
/** Live validation mode */
|
||||
liveValidate?: boolean;
|
||||
/** Form context passed to all widgets */
|
||||
formContext?: ConfigFormContext;
|
||||
/** i18n namespace for field labels */
|
||||
i18nNamespace?: string;
|
||||
/** Optional custom validation */
|
||||
customValidate?: (
|
||||
formData: unknown,
|
||||
errors: FormValidation,
|
||||
) => FormValidation;
|
||||
/** Called whenever form validation state changes */
|
||||
onValidationChange?: (hasErrors: boolean) => void;
|
||||
}
|
||||
|
||||
export function ConfigForm({
|
||||
schema,
|
||||
formData,
|
||||
onChange,
|
||||
onSubmit,
|
||||
onError,
|
||||
uiSchema: customUiSchema,
|
||||
fieldOrder,
|
||||
fieldGroups,
|
||||
hiddenFields,
|
||||
advancedFields,
|
||||
disabled = false,
|
||||
readonly = false,
|
||||
showSubmit = false,
|
||||
className,
|
||||
liveValidate = true,
|
||||
formContext,
|
||||
i18nNamespace,
|
||||
customValidate,
|
||||
onValidationChange,
|
||||
}: ConfigFormProps) {
|
||||
const { t, i18n } = useTranslation([
|
||||
i18nNamespace || "common",
|
||||
"views/settings",
|
||||
"config/validation",
|
||||
]);
|
||||
|
||||
// Determine which fields to hide based on advanced toggle
|
||||
const effectiveHiddenFields = useMemo(() => {
|
||||
return hiddenFields;
|
||||
}, [hiddenFields]);
|
||||
|
||||
// Transform schema and generate uiSchema
|
||||
const { schema: transformedSchema, uiSchema: generatedUiSchema } = useMemo(
|
||||
() =>
|
||||
transformSchema(schema, {
|
||||
fieldOrder,
|
||||
hiddenFields: effectiveHiddenFields,
|
||||
advancedFields: advancedFields,
|
||||
i18nNamespace,
|
||||
}),
|
||||
[schema, fieldOrder, effectiveHiddenFields, advancedFields, i18nNamespace],
|
||||
);
|
||||
|
||||
const { baseUiSchema, pathOverrides } = useMemo(
|
||||
() => splitUiSchemaOverrides(customUiSchema),
|
||||
[customUiSchema],
|
||||
);
|
||||
|
||||
// Merge generated uiSchema with custom overrides
|
||||
const finalUiSchema = useMemo(() => {
|
||||
// Start with generated schema
|
||||
const expandedUiSchema = applyUiSchemaPathOverrides(
|
||||
generatedUiSchema,
|
||||
transformedSchema,
|
||||
pathOverrides,
|
||||
);
|
||||
const merged = applyLayoutGridFieldDefaults(
|
||||
mergeUiSchema(expandedUiSchema, baseUiSchema),
|
||||
);
|
||||
|
||||
// Add field groups
|
||||
if (fieldGroups) {
|
||||
merged["ui:groups"] = fieldGroups;
|
||||
}
|
||||
|
||||
// Set submit button options
|
||||
merged["ui:submitButtonOptions"] = showSubmit
|
||||
? { norender: false }
|
||||
: { norender: true };
|
||||
|
||||
// Ensure hiddenFields take precedence over any custom uiSchema overrides
|
||||
// Build path-based overrides for hidden fields and apply them after merging
|
||||
if (hiddenFields && hiddenFields.length > 0) {
|
||||
const hiddenOverrides = hiddenFields.map((field) => ({
|
||||
path: field.split("."),
|
||||
value: { "ui:widget": "hidden" } as UiSchema,
|
||||
}));
|
||||
|
||||
return applyUiSchemaPathOverrides(
|
||||
merged,
|
||||
transformedSchema,
|
||||
hiddenOverrides,
|
||||
);
|
||||
}
|
||||
|
||||
return merged;
|
||||
}, [
|
||||
generatedUiSchema,
|
||||
transformedSchema,
|
||||
pathOverrides,
|
||||
baseUiSchema,
|
||||
showSubmit,
|
||||
fieldGroups,
|
||||
hiddenFields,
|
||||
]);
|
||||
|
||||
// Create error transformer for user-friendly error messages
|
||||
const errorTransformer = useMemo(() => createErrorTransformer(i18n), [i18n]);
|
||||
|
||||
const handleChange = useCallback(
|
||||
(e: IChangeEvent) => {
|
||||
onValidationChange?.(Array.isArray(e.errors) && e.errors.length > 0);
|
||||
onChange?.(e.formData);
|
||||
},
|
||||
[onChange, onValidationChange],
|
||||
);
|
||||
|
||||
const handleSubmit = useCallback(
|
||||
(e: IChangeEvent) => {
|
||||
onSubmit?.(e.formData);
|
||||
},
|
||||
[onSubmit],
|
||||
);
|
||||
|
||||
// Extended form context with i18n info
|
||||
const extendedFormContext = useMemo(
|
||||
() => ({
|
||||
...formContext,
|
||||
i18nNamespace,
|
||||
t,
|
||||
}),
|
||||
[formContext, i18nNamespace, t],
|
||||
);
|
||||
|
||||
return (
|
||||
<div className={cn("config-form w-full max-w-5xl", className)}>
|
||||
<Form
|
||||
schema={transformedSchema}
|
||||
uiSchema={finalUiSchema}
|
||||
formData={formData}
|
||||
validator={validator}
|
||||
onChange={handleChange}
|
||||
onSubmit={handleSubmit}
|
||||
onError={onError}
|
||||
disabled={disabled}
|
||||
readonly={readonly}
|
||||
liveValidate={liveValidate}
|
||||
formContext={extendedFormContext}
|
||||
transformErrors={errorTransformer}
|
||||
customValidate={customValidate}
|
||||
{...frigateTheme}
|
||||
/>
|
||||
</div>
|
||||
);
|
||||
}
|
||||
|
||||
export default ConfigForm;
|
||||
42
web/src/components/config-form/section-configs/audio.ts
Normal file
42
web/src/components/config-form/section-configs/audio.ts
Normal file
@ -0,0 +1,42 @@
|
||||
import type { SectionConfigOverrides } from "./types";
|
||||
|
||||
const audio: SectionConfigOverrides = {
|
||||
base: {
|
||||
sectionDocs: "/configuration/audio_detectors",
|
||||
restartRequired: [],
|
||||
fieldOrder: [
|
||||
"enabled",
|
||||
"listen",
|
||||
"filters",
|
||||
"min_volume",
|
||||
"max_not_heard",
|
||||
"num_threads",
|
||||
],
|
||||
fieldGroups: {
|
||||
detection: ["enabled", "listen", "filters"],
|
||||
sensitivity: ["min_volume", "max_not_heard"],
|
||||
},
|
||||
hiddenFields: ["enabled_in_config"],
|
||||
advancedFields: ["min_volume", "max_not_heard", "num_threads"],
|
||||
uiSchema: {
|
||||
listen: {
|
||||
"ui:widget": "audioLabels",
|
||||
},
|
||||
},
|
||||
},
|
||||
global: {
|
||||
restartRequired: [
|
||||
"enabled",
|
||||
"listen",
|
||||
"filters",
|
||||
"min_volume",
|
||||
"max_not_heard",
|
||||
"num_threads",
|
||||
],
|
||||
},
|
||||
camera: {
|
||||
restartRequired: ["num_threads"],
|
||||
},
|
||||
};
|
||||
|
||||
export default audio;
|
||||
@ -0,0 +1,19 @@
|
||||
import type { SectionConfigOverrides } from "./types";
|
||||
|
||||
const audioTranscription: SectionConfigOverrides = {
|
||||
base: {
|
||||
sectionDocs: "/configuration/audio_detectors#audio-transcription",
|
||||
restartRequired: [],
|
||||
fieldOrder: ["enabled", "language", "device", "model_size"],
|
||||
hiddenFields: ["enabled_in_config", "live_enabled"],
|
||||
advancedFields: ["language", "device", "model_size"],
|
||||
overrideFields: ["enabled", "live_enabled"],
|
||||
},
|
||||
global: {
|
||||
fieldOrder: ["enabled", "language", "device", "model_size"],
|
||||
advancedFields: ["language", "device", "model_size"],
|
||||
restartRequired: ["enabled", "language", "device", "model_size"],
|
||||
},
|
||||
};
|
||||
|
||||
export default audioTranscription;
|
||||
49
web/src/components/config-form/section-configs/auth.ts
Normal file
49
web/src/components/config-form/section-configs/auth.ts
Normal file
@ -0,0 +1,49 @@
|
||||
import type { SectionConfigOverrides } from "./types";
|
||||
|
||||
const auth: SectionConfigOverrides = {
|
||||
base: {
|
||||
sectionDocs: "/configuration/authentication",
|
||||
restartRequired: [
|
||||
"enabled",
|
||||
"reset_admin_password",
|
||||
"failed_login_rate_limit",
|
||||
],
|
||||
fieldOrder: [
|
||||
"enabled",
|
||||
"reset_admin_password",
|
||||
"cookie_name",
|
||||
"cookie_secure",
|
||||
"session_length",
|
||||
"refresh_time",
|
||||
"native_oauth_url",
|
||||
"failed_login_rate_limit",
|
||||
"trusted_proxies",
|
||||
"hash_iterations",
|
||||
"roles",
|
||||
],
|
||||
hiddenFields: ["admin_first_time_login", "roles"],
|
||||
advancedFields: [
|
||||
"cookie_name",
|
||||
"cookie_secure",
|
||||
"session_length",
|
||||
"refresh_time",
|
||||
"failed_login_rate_limit",
|
||||
"trusted_proxies",
|
||||
"hash_iterations",
|
||||
"roles",
|
||||
],
|
||||
uiSchema: {
|
||||
reset_admin_password: {
|
||||
"ui:widget": "switch",
|
||||
},
|
||||
native_oauth_url: {
|
||||
"ui:options": { size: "lg" },
|
||||
},
|
||||
failed_login_rate_limit: {
|
||||
"ui:options": { size: "md" },
|
||||
},
|
||||
},
|
||||
},
|
||||
};
|
||||
|
||||
export default auth;
|
||||
45
web/src/components/config-form/section-configs/birdseye.ts
Normal file
45
web/src/components/config-form/section-configs/birdseye.ts
Normal file
@ -0,0 +1,45 @@
|
||||
import type { SectionConfigOverrides } from "./types";
|
||||
|
||||
const birdseye: SectionConfigOverrides = {
|
||||
base: {
|
||||
sectionDocs: "/configuration/birdseye",
|
||||
restartRequired: [],
|
||||
fieldOrder: ["enabled", "mode", "order"],
|
||||
hiddenFields: [],
|
||||
advancedFields: [],
|
||||
overrideFields: ["enabled", "mode"],
|
||||
},
|
||||
global: {
|
||||
fieldOrder: [
|
||||
"enabled",
|
||||
"restream",
|
||||
"width",
|
||||
"height",
|
||||
"quality",
|
||||
"mode",
|
||||
"layout",
|
||||
"inactivity_threshold",
|
||||
"idle_heartbeat_fps",
|
||||
],
|
||||
advancedFields: ["width", "height", "quality", "inactivity_threshold"],
|
||||
restartRequired: [
|
||||
"enabled",
|
||||
"restream",
|
||||
"width",
|
||||
"height",
|
||||
"quality",
|
||||
"mode",
|
||||
"layout.scaling_factor",
|
||||
"inactivity_threshold",
|
||||
"layout.max_cameras",
|
||||
"idle_heartbeat_fps",
|
||||
],
|
||||
uiSchema: {
|
||||
mode: {
|
||||
"ui:size": "xs",
|
||||
},
|
||||
},
|
||||
},
|
||||
};
|
||||
|
||||
export default birdseye;
|
||||
@ -0,0 +1,12 @@
|
||||
import type { SectionConfigOverrides } from "./types";
|
||||
|
||||
const classification: SectionConfigOverrides = {
|
||||
base: {
|
||||
sectionDocs: "/configuration/custom_classification/object_classification",
|
||||
restartRequired: ["bird.enabled", "bird.threshold"],
|
||||
hiddenFields: ["custom"],
|
||||
advancedFields: [],
|
||||
},
|
||||
};
|
||||
|
||||
export default classification;
|
||||
17
web/src/components/config-form/section-configs/database.ts
Normal file
17
web/src/components/config-form/section-configs/database.ts
Normal file
@ -0,0 +1,17 @@
|
||||
import type { SectionConfigOverrides } from "./types";
|
||||
|
||||
const database: SectionConfigOverrides = {
|
||||
base: {
|
||||
sectionDocs: "/configuration/advanced#database",
|
||||
restartRequired: ["path"],
|
||||
fieldOrder: ["path"],
|
||||
advancedFields: [],
|
||||
uiSchema: {
|
||||
path: {
|
||||
"ui:options": { size: "md" },
|
||||
},
|
||||
},
|
||||
},
|
||||
};
|
||||
|
||||
export default database;
|
||||
49
web/src/components/config-form/section-configs/detect.ts
Normal file
49
web/src/components/config-form/section-configs/detect.ts
Normal file
@ -0,0 +1,49 @@
|
||||
import type { SectionConfigOverrides } from "./types";
|
||||
|
||||
const detect: SectionConfigOverrides = {
|
||||
base: {
|
||||
sectionDocs: "/configuration/camera_specific",
|
||||
fieldOrder: [
|
||||
"enabled",
|
||||
"width",
|
||||
"height",
|
||||
"fps",
|
||||
"min_initialized",
|
||||
"max_disappeared",
|
||||
"annotation_offset",
|
||||
"stationary",
|
||||
"interval",
|
||||
"threshold",
|
||||
"max_frames",
|
||||
],
|
||||
restartRequired: [],
|
||||
fieldGroups: {
|
||||
resolution: ["enabled", "width", "height", "fps"],
|
||||
tracking: ["min_initialized", "max_disappeared"],
|
||||
},
|
||||
hiddenFields: ["enabled_in_config"],
|
||||
advancedFields: [
|
||||
"min_initialized",
|
||||
"max_disappeared",
|
||||
"annotation_offset",
|
||||
"stationary",
|
||||
],
|
||||
},
|
||||
global: {
|
||||
restartRequired: [
|
||||
"enabled",
|
||||
"width",
|
||||
"height",
|
||||
"fps",
|
||||
"min_initialized",
|
||||
"max_disappeared",
|
||||
"annotation_offset",
|
||||
"stationary",
|
||||
],
|
||||
},
|
||||
camera: {
|
||||
restartRequired: ["width", "height", "min_initialized", "max_disappeared"],
|
||||
},
|
||||
};
|
||||
|
||||
export default detect;
|
||||
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue
Block a user