mirror of
https://github.com/blakeblackshear/frigate.git
synced 2026-05-09 15:05:26 +03:00
Compare commits
6 Commits
d6d636ea27
...
b218221a60
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
b218221a60 | ||
|
|
1a6d04fde7 | ||
|
|
4a1b7a1629 | ||
|
|
8eace9c3e7 | ||
|
|
b5a360be39 | ||
|
|
54a7c5015e |
@ -590,112 +590,92 @@ class BirdsEyeFrameManager:
|
|||||||
) -> Optional[list[list[Any]]]:
|
) -> Optional[list[list[Any]]]:
|
||||||
"""Calculate the optimal layout for 2+ cameras."""
|
"""Calculate the optimal layout for 2+ cameras."""
|
||||||
|
|
||||||
def map_layout(
|
def find_available_x(
|
||||||
camera_layout: list[list[Any]], row_height: int
|
current_x: int,
|
||||||
) -> tuple[int, int, Optional[list[list[Any]]]]:
|
width: int,
|
||||||
"""Map the calculated layout."""
|
reserved_ranges: list[tuple[int, int]],
|
||||||
candidate_layout = []
|
max_width: int,
|
||||||
starting_x = 0
|
) -> Optional[int]:
|
||||||
x = 0
|
"""Find the first horizontal slot that does not collide with reservations."""
|
||||||
max_width = 0
|
x = current_x
|
||||||
y = 0
|
|
||||||
|
|
||||||
for row in camera_layout:
|
for reserved_start, reserved_end in sorted(reserved_ranges):
|
||||||
final_row = []
|
if x >= reserved_end:
|
||||||
max_width = max(max_width, x)
|
continue
|
||||||
x = starting_x
|
|
||||||
for cameras in row:
|
|
||||||
camera_dims = self.cameras[cameras[0]]["dimensions"].copy()
|
|
||||||
camera_aspect = cameras[1]
|
|
||||||
|
|
||||||
if camera_dims[1] > camera_dims[0]:
|
if x + width <= reserved_start:
|
||||||
scaled_height = int(row_height * 2)
|
return x
|
||||||
scaled_width = int(scaled_height * camera_aspect)
|
|
||||||
starting_x = scaled_width
|
|
||||||
else:
|
|
||||||
scaled_height = row_height
|
|
||||||
scaled_width = int(scaled_height * camera_aspect)
|
|
||||||
|
|
||||||
# layout is too large
|
x = max(x, reserved_end)
|
||||||
if (
|
|
||||||
x + scaled_width > self.canvas.width
|
|
||||||
or y + scaled_height > self.canvas.height
|
|
||||||
):
|
|
||||||
return x + scaled_width, y + scaled_height, None
|
|
||||||
|
|
||||||
final_row.append((cameras[0], (x, y, scaled_width, scaled_height)))
|
if x + width <= max_width:
|
||||||
x += scaled_width
|
return x
|
||||||
|
|
||||||
y += row_height
|
|
||||||
candidate_layout.append(final_row)
|
|
||||||
|
|
||||||
if max_width == 0:
|
|
||||||
max_width = x
|
|
||||||
|
|
||||||
return max_width, y, candidate_layout
|
|
||||||
|
|
||||||
canvas_aspect_x, canvas_aspect_y = self.canvas.get_aspect(coefficient)
|
|
||||||
camera_layout: list[list[Any]] = []
|
|
||||||
camera_layout.append([])
|
|
||||||
starting_x = 0
|
|
||||||
x = starting_x
|
|
||||||
y = 0
|
|
||||||
y_i = 0
|
|
||||||
max_y = 0
|
|
||||||
for camera in cameras_to_add:
|
|
||||||
camera_dims = self.cameras[camera]["dimensions"].copy()
|
|
||||||
camera_aspect_x, camera_aspect_y = self.canvas.get_camera_aspect(
|
|
||||||
camera, camera_dims[0], camera_dims[1]
|
|
||||||
)
|
|
||||||
|
|
||||||
if camera_dims[1] > camera_dims[0]:
|
|
||||||
portrait = True
|
|
||||||
else:
|
|
||||||
portrait = False
|
|
||||||
|
|
||||||
if (x + camera_aspect_x) <= canvas_aspect_x:
|
|
||||||
# insert if camera can fit on current row
|
|
||||||
camera_layout[y_i].append(
|
|
||||||
(
|
|
||||||
camera,
|
|
||||||
camera_aspect_x / camera_aspect_y,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
if portrait:
|
|
||||||
starting_x = camera_aspect_x
|
|
||||||
else:
|
|
||||||
max_y = max(
|
|
||||||
max_y,
|
|
||||||
camera_aspect_y,
|
|
||||||
)
|
|
||||||
|
|
||||||
x += camera_aspect_x
|
|
||||||
else:
|
|
||||||
# move on to the next row and insert
|
|
||||||
y += max_y
|
|
||||||
y_i += 1
|
|
||||||
camera_layout.append([])
|
|
||||||
x = starting_x
|
|
||||||
|
|
||||||
if x + camera_aspect_x > canvas_aspect_x:
|
|
||||||
return None
|
|
||||||
|
|
||||||
camera_layout[y_i].append(
|
|
||||||
(
|
|
||||||
camera,
|
|
||||||
camera_aspect_x / camera_aspect_y,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
x += camera_aspect_x
|
|
||||||
|
|
||||||
if y + max_y > canvas_aspect_y:
|
|
||||||
return None
|
return None
|
||||||
|
|
||||||
row_height = int(self.canvas.height / coefficient)
|
def map_layout(row_height: int) -> tuple[int, int, Optional[list[list[Any]]]]:
|
||||||
total_width, total_height, standard_candidate_layout = map_layout(
|
"""Lay out cameras row by row while reserving portrait spans for the next row."""
|
||||||
camera_layout, row_height
|
candidate_layout: list[list[Any]] = []
|
||||||
)
|
reserved_ranges: dict[int, list[tuple[int, int]]] = {}
|
||||||
|
current_row: list[Any] = []
|
||||||
|
row_index = 0
|
||||||
|
row_y = 0
|
||||||
|
row_x = 0
|
||||||
|
max_width = 0
|
||||||
|
max_height = 0
|
||||||
|
|
||||||
|
for camera in cameras_to_add:
|
||||||
|
camera_dims = self.cameras[camera]["dimensions"].copy()
|
||||||
|
camera_aspect_x, camera_aspect_y = self.canvas.get_camera_aspect(
|
||||||
|
camera, camera_dims[0], camera_dims[1]
|
||||||
|
)
|
||||||
|
portrait = camera_dims[1] > camera_dims[0]
|
||||||
|
scaled_height = row_height * 2 if portrait else row_height
|
||||||
|
scaled_width = int(scaled_height * (camera_aspect_x / camera_aspect_y))
|
||||||
|
|
||||||
|
while True:
|
||||||
|
x = find_available_x(
|
||||||
|
row_x,
|
||||||
|
scaled_width,
|
||||||
|
reserved_ranges.get(row_index, []),
|
||||||
|
self.canvas.width,
|
||||||
|
)
|
||||||
|
|
||||||
|
if x is not None and row_y + scaled_height <= self.canvas.height:
|
||||||
|
current_row.append(
|
||||||
|
(camera, (x, row_y, scaled_width, scaled_height))
|
||||||
|
)
|
||||||
|
row_x = x + scaled_width
|
||||||
|
max_width = max(max_width, row_x)
|
||||||
|
max_height = max(max_height, row_y + scaled_height)
|
||||||
|
|
||||||
|
if portrait:
|
||||||
|
reserved_ranges.setdefault(row_index + 1, []).append(
|
||||||
|
(x, row_x)
|
||||||
|
)
|
||||||
|
|
||||||
|
break
|
||||||
|
|
||||||
|
if current_row:
|
||||||
|
candidate_layout.append(current_row)
|
||||||
|
current_row = []
|
||||||
|
|
||||||
|
row_index += 1
|
||||||
|
row_y = row_index * row_height
|
||||||
|
row_x = 0
|
||||||
|
|
||||||
|
if row_y + scaled_height > self.canvas.height:
|
||||||
|
overflow_width = max(max_width, scaled_width)
|
||||||
|
overflow_height = row_y + scaled_height
|
||||||
|
return overflow_width, overflow_height, None
|
||||||
|
|
||||||
|
if current_row:
|
||||||
|
candidate_layout.append(current_row)
|
||||||
|
|
||||||
|
return max_width, max_height, candidate_layout
|
||||||
|
|
||||||
|
row_height = max(1, int(self.canvas.height / coefficient))
|
||||||
|
total_width, total_height, standard_candidate_layout = map_layout(row_height)
|
||||||
|
|
||||||
if not standard_candidate_layout:
|
if not standard_candidate_layout:
|
||||||
# if standard layout didn't work
|
# if standard layout didn't work
|
||||||
@ -704,9 +684,9 @@ class BirdsEyeFrameManager:
|
|||||||
total_width / self.canvas.width,
|
total_width / self.canvas.width,
|
||||||
total_height / self.canvas.height,
|
total_height / self.canvas.height,
|
||||||
)
|
)
|
||||||
row_height = int(row_height / scale_down_percent)
|
row_height = max(1, int(row_height / scale_down_percent))
|
||||||
total_width, total_height, standard_candidate_layout = map_layout(
|
total_width, total_height, standard_candidate_layout = map_layout(
|
||||||
camera_layout, row_height
|
row_height
|
||||||
)
|
)
|
||||||
|
|
||||||
if not standard_candidate_layout:
|
if not standard_candidate_layout:
|
||||||
@ -720,8 +700,8 @@ class BirdsEyeFrameManager:
|
|||||||
1 / (total_width / self.canvas.width),
|
1 / (total_width / self.canvas.width),
|
||||||
1 / (total_height / self.canvas.height),
|
1 / (total_height / self.canvas.height),
|
||||||
)
|
)
|
||||||
row_height = int(row_height * scale_up_percent)
|
row_height = max(1, int(row_height * scale_up_percent))
|
||||||
_, _, scaled_layout = map_layout(camera_layout, row_height)
|
_, _, scaled_layout = map_layout(row_height)
|
||||||
|
|
||||||
if scaled_layout:
|
if scaled_layout:
|
||||||
return scaled_layout
|
return scaled_layout
|
||||||
|
|||||||
@ -1,11 +1,64 @@
|
|||||||
"""Test camera user and password cleanup."""
|
"""Tests for Birdseye canvas sizing and layout behavior."""
|
||||||
|
|
||||||
import unittest
|
import unittest
|
||||||
|
from multiprocessing import Event
|
||||||
|
|
||||||
from frigate.output.birdseye import get_canvas_shape
|
from frigate.config import FrigateConfig
|
||||||
|
from frigate.output.birdseye import BirdsEyeFrameManager, get_canvas_shape
|
||||||
|
|
||||||
|
|
||||||
class TestBirdseye(unittest.TestCase):
|
class TestBirdseye(unittest.TestCase):
|
||||||
|
def _build_manager(
|
||||||
|
self, camera_dimensions: dict[str, tuple[int, int]]
|
||||||
|
) -> BirdsEyeFrameManager:
|
||||||
|
config = {
|
||||||
|
"mqtt": {"host": "mqtt"},
|
||||||
|
"birdseye": {"width": 1280, "height": 720},
|
||||||
|
"cameras": {},
|
||||||
|
}
|
||||||
|
|
||||||
|
for order, (camera, dimensions) in enumerate(
|
||||||
|
camera_dimensions.items(), start=1
|
||||||
|
):
|
||||||
|
config["cameras"][camera] = {
|
||||||
|
"ffmpeg": {
|
||||||
|
"inputs": [
|
||||||
|
{
|
||||||
|
"path": f"rtsp://10.0.0.1:554/{camera}",
|
||||||
|
"roles": ["detect"],
|
||||||
|
}
|
||||||
|
]
|
||||||
|
},
|
||||||
|
"detect": {
|
||||||
|
"width": dimensions[0],
|
||||||
|
"height": dimensions[1],
|
||||||
|
"fps": 5,
|
||||||
|
},
|
||||||
|
"birdseye": {"order": order},
|
||||||
|
}
|
||||||
|
|
||||||
|
return BirdsEyeFrameManager(FrigateConfig(**config), Event())
|
||||||
|
|
||||||
|
def _assert_no_overlaps(
|
||||||
|
self, layout: list[list[tuple[str, tuple[int, int, int, int]]]]
|
||||||
|
):
|
||||||
|
rectangles = [position for row in layout for _, position in row]
|
||||||
|
|
||||||
|
for index, rect in enumerate(rectangles):
|
||||||
|
x1, y1, width1, height1 = rect
|
||||||
|
for other in rectangles[index + 1 :]:
|
||||||
|
x2, y2, width2, height2 = other
|
||||||
|
overlap = (
|
||||||
|
x1 < x2 + width2
|
||||||
|
and x2 < x1 + width1
|
||||||
|
and y1 < y2 + height2
|
||||||
|
and y2 < y1 + height1
|
||||||
|
)
|
||||||
|
self.assertFalse(
|
||||||
|
overlap,
|
||||||
|
msg=f"Overlapping rectangles found: {rect} and {other}",
|
||||||
|
)
|
||||||
|
|
||||||
def test_16x9(self):
|
def test_16x9(self):
|
||||||
"""Test 16x9 aspect ratio works as expected for birdseye."""
|
"""Test 16x9 aspect ratio works as expected for birdseye."""
|
||||||
width = 1280
|
width = 1280
|
||||||
@ -45,3 +98,104 @@ class TestBirdseye(unittest.TestCase):
|
|||||||
canvas_width, canvas_height = get_canvas_shape(width, height)
|
canvas_width, canvas_height = get_canvas_shape(width, height)
|
||||||
assert canvas_width == width # width will be the same
|
assert canvas_width == width # width will be the same
|
||||||
assert canvas_height != height
|
assert canvas_height != height
|
||||||
|
|
||||||
|
def test_portrait_camera_does_not_overlap_next_row(self):
|
||||||
|
"""Portrait cameras should reserve their real horizontal position on the next row."""
|
||||||
|
manager = self._build_manager(
|
||||||
|
{
|
||||||
|
"cam_a": (1280, 720),
|
||||||
|
"cam_p": (360, 640),
|
||||||
|
"cam_b": (1280, 720),
|
||||||
|
"cam_c": (640, 480),
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
layout = manager.calculate_layout(["cam_a", "cam_p", "cam_b", "cam_c"], 3)
|
||||||
|
|
||||||
|
self.assertIsNotNone(layout)
|
||||||
|
assert layout is not None
|
||||||
|
self._assert_no_overlaps(layout)
|
||||||
|
|
||||||
|
cam_c = [
|
||||||
|
position for row in layout for camera, position in row if camera == "cam_c"
|
||||||
|
][0]
|
||||||
|
self.assertEqual(cam_c[0], 0)
|
||||||
|
|
||||||
|
def test_portrait_reservation_only_applies_to_next_row(self):
|
||||||
|
"""Portrait reservations should not push later rows after the span ends."""
|
||||||
|
manager = self._build_manager(
|
||||||
|
{
|
||||||
|
"cam_a": (1280, 720),
|
||||||
|
"cam_p": (360, 640),
|
||||||
|
"cam_b": (1280, 720),
|
||||||
|
"cam_c": (1280, 720),
|
||||||
|
"cam_d": (1280, 720),
|
||||||
|
"cam_e": (1280, 720),
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
layout = manager.calculate_layout(
|
||||||
|
["cam_a", "cam_p", "cam_b", "cam_c", "cam_d", "cam_e"],
|
||||||
|
3,
|
||||||
|
)
|
||||||
|
|
||||||
|
self.assertIsNotNone(layout)
|
||||||
|
assert layout is not None
|
||||||
|
self._assert_no_overlaps(layout)
|
||||||
|
|
||||||
|
cam_e = [
|
||||||
|
position for row in layout for camera, position in row if camera == "cam_e"
|
||||||
|
][0]
|
||||||
|
self.assertEqual(cam_e[0], 0)
|
||||||
|
|
||||||
|
def test_multiple_portraits_reserve_distinct_ranges(self):
|
||||||
|
"""Multiple portrait cameras in one row should reserve separate spans below them."""
|
||||||
|
manager = self._build_manager(
|
||||||
|
{
|
||||||
|
"cam_a": (640, 480),
|
||||||
|
"cam_p1": (360, 640),
|
||||||
|
"cam_p2": (360, 640),
|
||||||
|
"cam_b": (640, 480),
|
||||||
|
"cam_c": (1280, 720),
|
||||||
|
"cam_d": (640, 480),
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
layout = manager.calculate_layout(
|
||||||
|
["cam_a", "cam_p1", "cam_p2", "cam_b", "cam_c", "cam_d"],
|
||||||
|
4,
|
||||||
|
)
|
||||||
|
|
||||||
|
self.assertIsNotNone(layout)
|
||||||
|
assert layout is not None
|
||||||
|
self._assert_no_overlaps(layout)
|
||||||
|
|
||||||
|
def test_two_landscapes_then_portrait_then_two_landscapes(self):
|
||||||
|
"""A portrait after two landscapes should reserve only its own tail span."""
|
||||||
|
manager = self._build_manager(
|
||||||
|
{
|
||||||
|
"cam_a": (1280, 720),
|
||||||
|
"cam_b": (1280, 720),
|
||||||
|
"cam_p": (360, 640),
|
||||||
|
"cam_c": (1280, 720),
|
||||||
|
"cam_d": (1280, 720),
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
layout = manager.calculate_layout(
|
||||||
|
["cam_a", "cam_b", "cam_p", "cam_c", "cam_d"],
|
||||||
|
3,
|
||||||
|
)
|
||||||
|
|
||||||
|
self.assertIsNotNone(layout)
|
||||||
|
assert layout is not None
|
||||||
|
self._assert_no_overlaps(layout)
|
||||||
|
|
||||||
|
cam_c = [
|
||||||
|
position for row in layout for camera, position in row if camera == "cam_c"
|
||||||
|
][0]
|
||||||
|
cam_d = [
|
||||||
|
position for row in layout for camera, position in row if camera == "cam_d"
|
||||||
|
][0]
|
||||||
|
self.assertEqual(cam_c[0], 0)
|
||||||
|
self.assertEqual(cam_d[0], cam_c[0] + cam_c[2])
|
||||||
|
|||||||
@ -24,8 +24,12 @@ from frigate.log import redirect_output_to_logger, suppress_stderr_during
|
|||||||
from frigate.models import Event, Recordings, ReviewSegment
|
from frigate.models import Event, Recordings, ReviewSegment
|
||||||
from frigate.types import ModelStatusTypesEnum
|
from frigate.types import ModelStatusTypesEnum
|
||||||
from frigate.util.downloader import ModelDownloader
|
from frigate.util.downloader import ModelDownloader
|
||||||
from frigate.util.file import get_event_thumbnail_bytes
|
from frigate.util.file import get_event_thumbnail_bytes, load_event_snapshot_image
|
||||||
from frigate.util.image import get_image_from_recording
|
from frigate.util.image import (
|
||||||
|
calculate_region,
|
||||||
|
get_image_from_recording,
|
||||||
|
relative_box_to_absolute,
|
||||||
|
)
|
||||||
from frigate.util.process import FrigateProcess
|
from frigate.util.process import FrigateProcess
|
||||||
|
|
||||||
BATCH_SIZE = 16
|
BATCH_SIZE = 16
|
||||||
@ -713,7 +717,7 @@ def collect_object_classification_examples(
|
|||||||
This function:
|
This function:
|
||||||
1. Queries events for the specified label
|
1. Queries events for the specified label
|
||||||
2. Selects 100 balanced events across different cameras and times
|
2. Selects 100 balanced events across different cameras and times
|
||||||
3. Retrieves thumbnails for selected events (with 33% center crop applied)
|
3. Crops each event's clean snapshot around the object bounding box
|
||||||
4. Selects 24 most visually distinct thumbnails
|
4. Selects 24 most visually distinct thumbnails
|
||||||
5. Saves to dataset directory
|
5. Saves to dataset directory
|
||||||
|
|
||||||
@ -832,66 +836,106 @@ def _select_balanced_events(
|
|||||||
|
|
||||||
def _extract_event_thumbnails(events: list[Event], output_dir: str) -> list[str]:
|
def _extract_event_thumbnails(events: list[Event], output_dir: str) -> list[str]:
|
||||||
"""
|
"""
|
||||||
Extract thumbnails from events and save to disk.
|
Extract a training image for each event.
|
||||||
|
|
||||||
|
Preferred path: load the full-frame clean snapshot and crop around the
|
||||||
|
stored bounding box with the same calculate_region(..., max(w, h), 1.0)
|
||||||
|
call the live ObjectClassificationProcessor uses, so wizard examples
|
||||||
|
are framed like inference-time inputs.
|
||||||
|
|
||||||
|
Fallback: if no clean snapshot exists (snapshots disabled, or only a
|
||||||
|
legacy annotated JPG is on disk), center-crop the stored thumbnail
|
||||||
|
using a step ladder sized from the box/region area ratio.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
events: List of Event objects
|
events: List of Event objects
|
||||||
output_dir: Directory to save thumbnails
|
output_dir: Directory to save crops
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
List of paths to successfully extracted thumbnail images
|
List of paths to successfully extracted images
|
||||||
"""
|
"""
|
||||||
thumbnail_paths = []
|
image_paths = []
|
||||||
|
|
||||||
for idx, event in enumerate(events):
|
for idx, event in enumerate(events):
|
||||||
try:
|
try:
|
||||||
thumbnail_bytes = get_event_thumbnail_bytes(event)
|
img = _load_event_classification_crop(event)
|
||||||
|
if img is None:
|
||||||
|
continue
|
||||||
|
|
||||||
if thumbnail_bytes:
|
resized = cv2.resize(img, (224, 224))
|
||||||
nparr = np.frombuffer(thumbnail_bytes, np.uint8)
|
output_path = os.path.join(output_dir, f"thumbnail_{idx:04d}.jpg")
|
||||||
img = cv2.imdecode(nparr, cv2.IMREAD_COLOR)
|
cv2.imwrite(output_path, resized)
|
||||||
|
image_paths.append(output_path)
|
||||||
if img is not None:
|
|
||||||
height, width = img.shape[:2]
|
|
||||||
|
|
||||||
crop_size = 1.0
|
|
||||||
if event.data and "box" in event.data and "region" in event.data:
|
|
||||||
box = event.data["box"]
|
|
||||||
region = event.data["region"]
|
|
||||||
|
|
||||||
if len(box) == 4 and len(region) == 4:
|
|
||||||
box_w, box_h = box[2], box[3]
|
|
||||||
region_w, region_h = region[2], region[3]
|
|
||||||
|
|
||||||
box_area = (box_w * box_h) / (region_w * region_h)
|
|
||||||
|
|
||||||
if box_area < 0.05:
|
|
||||||
crop_size = 0.4
|
|
||||||
elif box_area < 0.10:
|
|
||||||
crop_size = 0.5
|
|
||||||
elif box_area < 0.20:
|
|
||||||
crop_size = 0.65
|
|
||||||
elif box_area < 0.35:
|
|
||||||
crop_size = 0.80
|
|
||||||
else:
|
|
||||||
crop_size = 0.95
|
|
||||||
|
|
||||||
crop_width = int(width * crop_size)
|
|
||||||
crop_height = int(height * crop_size)
|
|
||||||
|
|
||||||
x1 = (width - crop_width) // 2
|
|
||||||
y1 = (height - crop_height) // 2
|
|
||||||
x2 = x1 + crop_width
|
|
||||||
y2 = y1 + crop_height
|
|
||||||
|
|
||||||
cropped = img[y1:y2, x1:x2]
|
|
||||||
resized = cv2.resize(cropped, (224, 224))
|
|
||||||
output_path = os.path.join(output_dir, f"thumbnail_{idx:04d}.jpg")
|
|
||||||
cv2.imwrite(output_path, resized)
|
|
||||||
thumbnail_paths.append(output_path)
|
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.debug(f"Failed to extract thumbnail for event {event.id}: {e}")
|
logger.debug(f"Failed to extract image for event {event.id}: {e}")
|
||||||
continue
|
continue
|
||||||
|
|
||||||
return thumbnail_paths
|
return image_paths
|
||||||
|
|
||||||
|
|
||||||
|
def _load_event_classification_crop(event: Event) -> np.ndarray | None:
|
||||||
|
"""Prefer a snapshot-based object crop; fall back to a center-cropped thumbnail."""
|
||||||
|
if event.data and "box" in event.data:
|
||||||
|
snapshot, _ = load_event_snapshot_image(event, clean_only=True)
|
||||||
|
if snapshot is not None:
|
||||||
|
abs_box = relative_box_to_absolute(snapshot.shape, event.data["box"])
|
||||||
|
if abs_box is not None:
|
||||||
|
xmin, ymin, xmax, ymax = abs_box
|
||||||
|
box_w = xmax - xmin
|
||||||
|
box_h = ymax - ymin
|
||||||
|
if box_w > 0 and box_h > 0:
|
||||||
|
x1, y1, x2, y2 = calculate_region(
|
||||||
|
snapshot.shape,
|
||||||
|
xmin,
|
||||||
|
ymin,
|
||||||
|
xmax,
|
||||||
|
ymax,
|
||||||
|
max(box_w, box_h),
|
||||||
|
1.0,
|
||||||
|
)
|
||||||
|
cropped = snapshot[y1:y2, x1:x2]
|
||||||
|
if cropped.size > 0:
|
||||||
|
return cropped
|
||||||
|
|
||||||
|
thumbnail_bytes = get_event_thumbnail_bytes(event)
|
||||||
|
if not thumbnail_bytes:
|
||||||
|
return None
|
||||||
|
|
||||||
|
nparr = np.frombuffer(thumbnail_bytes, np.uint8)
|
||||||
|
img = cv2.imdecode(nparr, cv2.IMREAD_COLOR)
|
||||||
|
if img is None or img.size == 0:
|
||||||
|
return None
|
||||||
|
|
||||||
|
height, width = img.shape[:2]
|
||||||
|
crop_size = 1.0
|
||||||
|
|
||||||
|
if event.data and "box" in event.data and "region" in event.data:
|
||||||
|
box = event.data["box"]
|
||||||
|
region = event.data["region"]
|
||||||
|
|
||||||
|
if len(box) == 4 and len(region) == 4:
|
||||||
|
box_w, box_h = box[2], box[3]
|
||||||
|
region_w, region_h = region[2], region[3]
|
||||||
|
box_area = (box_w * box_h) / (region_w * region_h)
|
||||||
|
|
||||||
|
if box_area < 0.05:
|
||||||
|
crop_size = 0.4
|
||||||
|
elif box_area < 0.10:
|
||||||
|
crop_size = 0.5
|
||||||
|
elif box_area < 0.20:
|
||||||
|
crop_size = 0.65
|
||||||
|
elif box_area < 0.35:
|
||||||
|
crop_size = 0.80
|
||||||
|
else:
|
||||||
|
crop_size = 0.95
|
||||||
|
|
||||||
|
crop_width = int(width * crop_size)
|
||||||
|
crop_height = int(height * crop_size)
|
||||||
|
x1 = (width - crop_width) // 2
|
||||||
|
y1 = (height - crop_height) // 2
|
||||||
|
cropped = img[y1 : y1 + crop_height, x1 : x1 + crop_width]
|
||||||
|
if cropped.size == 0:
|
||||||
|
return None
|
||||||
|
|
||||||
|
return cropped
|
||||||
|
|||||||
@ -726,7 +726,20 @@ def ffprobe_stream(ffmpeg, path: str, detailed: bool = False) -> sp.CompletedPro
|
|||||||
if detailed and format_entries:
|
if detailed and format_entries:
|
||||||
cmd.extend(["-show_entries", f"format={format_entries}"])
|
cmd.extend(["-show_entries", f"format={format_entries}"])
|
||||||
cmd.extend(["-loglevel", "error", clean_path])
|
cmd.extend(["-loglevel", "error", clean_path])
|
||||||
return sp.run(cmd, capture_output=True)
|
try:
|
||||||
|
return sp.run(cmd, capture_output=True, timeout=6)
|
||||||
|
except sp.TimeoutExpired as e:
|
||||||
|
logger.info(
|
||||||
|
"ffprobe timed out while probing %s (transport=%s)",
|
||||||
|
clean_camera_user_pass(path),
|
||||||
|
rtsp_transport or "default",
|
||||||
|
)
|
||||||
|
return sp.CompletedProcess(
|
||||||
|
args=cmd,
|
||||||
|
returncode=1,
|
||||||
|
stdout=e.stdout or b"",
|
||||||
|
stderr=(e.stderr or b"") + b"\nffprobe timed out",
|
||||||
|
)
|
||||||
|
|
||||||
result = run()
|
result = run()
|
||||||
|
|
||||||
@ -832,11 +845,23 @@ async def get_video_properties(
|
|||||||
"-show_streams",
|
"-show_streams",
|
||||||
url,
|
url,
|
||||||
]
|
]
|
||||||
|
proc = None
|
||||||
try:
|
try:
|
||||||
proc = await asyncio.create_subprocess_exec(
|
proc = await asyncio.create_subprocess_exec(
|
||||||
*cmd, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE
|
*cmd, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE
|
||||||
)
|
)
|
||||||
stdout, _ = await proc.communicate()
|
try:
|
||||||
|
stdout, _ = await asyncio.wait_for(proc.communicate(), timeout=6)
|
||||||
|
except asyncio.TimeoutError:
|
||||||
|
logger.info(
|
||||||
|
"ffprobe timed out while probing %s (transport=%s)",
|
||||||
|
clean_camera_user_pass(url),
|
||||||
|
rtsp_transport or "default",
|
||||||
|
)
|
||||||
|
proc.kill()
|
||||||
|
await proc.wait()
|
||||||
|
return False, 0, 0, None, -1
|
||||||
|
|
||||||
if proc.returncode != 0:
|
if proc.returncode != 0:
|
||||||
return False, 0, 0, None, -1
|
return False, 0, 0, None, -1
|
||||||
|
|
||||||
|
|||||||
@ -17,6 +17,9 @@ import { useUserPersistence } from "@/hooks/use-user-persistence";
|
|||||||
import { Skeleton } from "../ui/skeleton";
|
import { Skeleton } from "../ui/skeleton";
|
||||||
import { Button } from "../ui/button";
|
import { Button } from "../ui/button";
|
||||||
import { FaCircleCheck } from "react-icons/fa6";
|
import { FaCircleCheck } from "react-icons/fa6";
|
||||||
|
import { FaExclamationTriangle } from "react-icons/fa";
|
||||||
|
import { MdOutlinePersonSearch } from "react-icons/md";
|
||||||
|
import { ThreatLevel } from "@/types/review";
|
||||||
import { cn } from "@/lib/utils";
|
import { cn } from "@/lib/utils";
|
||||||
import { useTranslation } from "react-i18next";
|
import { useTranslation } from "react-i18next";
|
||||||
import { getTranslatedLabel } from "@/utils/i18n";
|
import { getTranslatedLabel } from "@/utils/i18n";
|
||||||
@ -127,6 +130,11 @@ export function AnimatedEventCard({
|
|||||||
true,
|
true,
|
||||||
);
|
);
|
||||||
|
|
||||||
|
const threatLevel = useMemo<ThreatLevel>(
|
||||||
|
() => (event.data.metadata?.potential_threat_level ?? 0) as ThreatLevel,
|
||||||
|
[event],
|
||||||
|
);
|
||||||
|
|
||||||
const aspectRatio = useMemo(() => {
|
const aspectRatio = useMemo(() => {
|
||||||
if (
|
if (
|
||||||
!config ||
|
!config ||
|
||||||
@ -152,7 +160,15 @@ export function AnimatedEventCard({
|
|||||||
<Tooltip>
|
<Tooltip>
|
||||||
<TooltipTrigger asChild>
|
<TooltipTrigger asChild>
|
||||||
<Button
|
<Button
|
||||||
className="pointer-events-none absolute left-2 top-1 z-40 bg-gray-500 bg-gradient-to-br from-gray-400 to-gray-500 opacity-0 transition-opacity group-hover:pointer-events-auto group-hover:opacity-100"
|
className={cn(
|
||||||
|
"absolute left-2 top-1 z-40 transition-opacity",
|
||||||
|
threatLevel === ThreatLevel.SECURITY_CONCERN &&
|
||||||
|
"pointer-events-auto bg-severity_alert opacity-100 hover:bg-severity_alert",
|
||||||
|
threatLevel === ThreatLevel.NEEDS_REVIEW &&
|
||||||
|
"pointer-events-auto bg-severity_detection opacity-100 hover:bg-severity_detection",
|
||||||
|
threatLevel === ThreatLevel.NORMAL &&
|
||||||
|
"pointer-events-none bg-gray-500 bg-gradient-to-br from-gray-400 to-gray-500 opacity-0 group-hover:pointer-events-auto group-hover:opacity-100",
|
||||||
|
)}
|
||||||
size="xs"
|
size="xs"
|
||||||
aria-label={t("markAsReviewed")}
|
aria-label={t("markAsReviewed")}
|
||||||
onClick={async () => {
|
onClick={async () => {
|
||||||
@ -160,7 +176,13 @@ export function AnimatedEventCard({
|
|||||||
updateEvents();
|
updateEvents();
|
||||||
}}
|
}}
|
||||||
>
|
>
|
||||||
<FaCircleCheck className="size-3 text-white" />
|
{threatLevel === ThreatLevel.SECURITY_CONCERN ? (
|
||||||
|
<FaExclamationTriangle className="size-3 text-white" />
|
||||||
|
) : threatLevel === ThreatLevel.NEEDS_REVIEW ? (
|
||||||
|
<MdOutlinePersonSearch className="size-3 text-white" />
|
||||||
|
) : (
|
||||||
|
<FaCircleCheck className="size-3 text-white" />
|
||||||
|
)}
|
||||||
</Button>
|
</Button>
|
||||||
</TooltipTrigger>
|
</TooltipTrigger>
|
||||||
<TooltipContent>{t("markAsReviewed")}</TooltipContent>
|
<TooltipContent>{t("markAsReviewed")}</TooltipContent>
|
||||||
|
|||||||
@ -389,7 +389,7 @@ export default function LiveCameraView({
|
|||||||
return "mse";
|
return "mse";
|
||||||
}, [lowBandwidth, mic, webRTC, isRestreamed]);
|
}, [lowBandwidth, mic, webRTC, isRestreamed]);
|
||||||
|
|
||||||
useKeyboardListener(["m"], (key, modifiers) => {
|
useKeyboardListener(["m", "Escape"], (key, modifiers) => {
|
||||||
if (!modifiers.down) {
|
if (!modifiers.down) {
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
@ -407,6 +407,12 @@ export default function LiveCameraView({
|
|||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
|
case "Escape":
|
||||||
|
if (!fullscreen) {
|
||||||
|
navigate(-1);
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
return false;
|
return false;
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user