mirror of
https://github.com/blakeblackshear/frigate.git
synced 2025-12-06 21:44:13 +03:00
improve seeking logic in tracking details
mimic the logic of DynamicVideoController
This commit is contained in:
parent
0dbf648b3d
commit
fd06c272d2
@ -12,7 +12,11 @@ import { cn } from "@/lib/utils";
|
|||||||
import HlsVideoPlayer from "@/components/player/HlsVideoPlayer";
|
import HlsVideoPlayer from "@/components/player/HlsVideoPlayer";
|
||||||
import { baseUrl } from "@/api/baseUrl";
|
import { baseUrl } from "@/api/baseUrl";
|
||||||
import { REVIEW_PADDING } from "@/types/review";
|
import { REVIEW_PADDING } from "@/types/review";
|
||||||
import { ASPECT_VERTICAL_LAYOUT, ASPECT_WIDE_LAYOUT } from "@/types/record";
|
import {
|
||||||
|
ASPECT_VERTICAL_LAYOUT,
|
||||||
|
ASPECT_WIDE_LAYOUT,
|
||||||
|
Recording,
|
||||||
|
} from "@/types/record";
|
||||||
import {
|
import {
|
||||||
DropdownMenu,
|
DropdownMenu,
|
||||||
DropdownMenuTrigger,
|
DropdownMenuTrigger,
|
||||||
@ -75,6 +79,139 @@ export function TrackingDetails({
|
|||||||
|
|
||||||
const { data: config } = useSWR<FrigateConfig>("config");
|
const { data: config } = useSWR<FrigateConfig>("config");
|
||||||
|
|
||||||
|
// Fetch recording segments for the event's time range to handle motion-only gaps
|
||||||
|
const eventStartRecord = useMemo(
|
||||||
|
() => (event.start_time ?? 0) + annotationOffset / 1000,
|
||||||
|
[event.start_time, annotationOffset],
|
||||||
|
);
|
||||||
|
const eventEndRecord = useMemo(
|
||||||
|
() => (event.end_time ?? Date.now() / 1000) + annotationOffset / 1000,
|
||||||
|
[event.end_time, annotationOffset],
|
||||||
|
);
|
||||||
|
|
||||||
|
const { data: recordings } = useSWR<Recording[]>(
|
||||||
|
event.camera
|
||||||
|
? [
|
||||||
|
`${event.camera}/recordings`,
|
||||||
|
{
|
||||||
|
after: eventStartRecord - REVIEW_PADDING,
|
||||||
|
before: eventEndRecord + REVIEW_PADDING,
|
||||||
|
},
|
||||||
|
]
|
||||||
|
: null,
|
||||||
|
);
|
||||||
|
|
||||||
|
// Convert a timeline timestamp to actual video player time, accounting for
|
||||||
|
// motion-only recording gaps. Uses the same algorithm as DynamicVideoController.
|
||||||
|
const timestampToVideoTime = useCallback(
|
||||||
|
(timestamp: number): number => {
|
||||||
|
if (!recordings || recordings.length === 0) {
|
||||||
|
// Fallback to simple calculation if no recordings data
|
||||||
|
return timestamp - (eventStartRecord - REVIEW_PADDING);
|
||||||
|
}
|
||||||
|
|
||||||
|
const videoStartTime = eventStartRecord - REVIEW_PADDING;
|
||||||
|
|
||||||
|
// If timestamp is before video start, return 0
|
||||||
|
if (timestamp < videoStartTime) return 0;
|
||||||
|
|
||||||
|
// Check if timestamp is before the first recording or after the last
|
||||||
|
if (
|
||||||
|
timestamp < recordings[0].start_time ||
|
||||||
|
timestamp > recordings[recordings.length - 1].end_time
|
||||||
|
) {
|
||||||
|
// No recording available at this timestamp
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Calculate the inpoint offset - the HLS video may start partway through the first segment
|
||||||
|
let inpointOffset = 0;
|
||||||
|
if (
|
||||||
|
videoStartTime > recordings[0].start_time &&
|
||||||
|
videoStartTime < recordings[0].end_time
|
||||||
|
) {
|
||||||
|
inpointOffset = videoStartTime - recordings[0].start_time;
|
||||||
|
}
|
||||||
|
|
||||||
|
let seekSeconds = 0;
|
||||||
|
for (const segment of recordings) {
|
||||||
|
// Skip segments that end before our timestamp
|
||||||
|
if (segment.end_time <= timestamp) {
|
||||||
|
// Add this segment's duration, but subtract inpoint offset from first segment
|
||||||
|
if (segment === recordings[0]) {
|
||||||
|
seekSeconds += segment.duration - inpointOffset;
|
||||||
|
} else {
|
||||||
|
seekSeconds += segment.duration;
|
||||||
|
}
|
||||||
|
} else if (segment.start_time <= timestamp) {
|
||||||
|
// The timestamp is within this segment
|
||||||
|
if (segment === recordings[0]) {
|
||||||
|
// For the first segment, account for the inpoint offset
|
||||||
|
seekSeconds +=
|
||||||
|
timestamp - Math.max(segment.start_time, videoStartTime);
|
||||||
|
} else {
|
||||||
|
seekSeconds += timestamp - segment.start_time;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return seekSeconds;
|
||||||
|
},
|
||||||
|
[recordings, eventStartRecord],
|
||||||
|
);
|
||||||
|
|
||||||
|
// Convert video player time back to timeline timestamp, accounting for
|
||||||
|
// motion-only recording gaps. Reverse of timestampToVideoTime.
|
||||||
|
const videoTimeToTimestamp = useCallback(
|
||||||
|
(playerTime: number): number => {
|
||||||
|
if (!recordings || recordings.length === 0) {
|
||||||
|
// Fallback to simple calculation if no recordings data
|
||||||
|
const videoStartTime = eventStartRecord - REVIEW_PADDING;
|
||||||
|
return playerTime + videoStartTime;
|
||||||
|
}
|
||||||
|
|
||||||
|
const videoStartTime = eventStartRecord - REVIEW_PADDING;
|
||||||
|
|
||||||
|
// Calculate the inpoint offset - the video may start partway through the first segment
|
||||||
|
let inpointOffset = 0;
|
||||||
|
if (
|
||||||
|
videoStartTime > recordings[0].start_time &&
|
||||||
|
videoStartTime < recordings[0].end_time
|
||||||
|
) {
|
||||||
|
inpointOffset = videoStartTime - recordings[0].start_time;
|
||||||
|
}
|
||||||
|
|
||||||
|
let timestamp = 0;
|
||||||
|
let totalTime = 0;
|
||||||
|
|
||||||
|
for (const segment of recordings) {
|
||||||
|
const segmentDuration =
|
||||||
|
segment === recordings[0]
|
||||||
|
? segment.duration - inpointOffset
|
||||||
|
: segment.duration;
|
||||||
|
|
||||||
|
if (totalTime + segmentDuration > playerTime) {
|
||||||
|
// The player time is within this segment
|
||||||
|
if (segment === recordings[0]) {
|
||||||
|
// For the first segment, add the inpoint offset
|
||||||
|
timestamp =
|
||||||
|
Math.max(segment.start_time, videoStartTime) +
|
||||||
|
(playerTime - totalTime);
|
||||||
|
} else {
|
||||||
|
timestamp = segment.start_time + (playerTime - totalTime);
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
} else {
|
||||||
|
totalTime += segmentDuration;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return timestamp;
|
||||||
|
},
|
||||||
|
[recordings, eventStartRecord],
|
||||||
|
);
|
||||||
|
|
||||||
eventSequence?.map((event) => {
|
eventSequence?.map((event) => {
|
||||||
event.data.zones_friendly_names = event.data?.zones?.map((zone) => {
|
event.data.zones_friendly_names = event.data?.zones?.map((zone) => {
|
||||||
return resolveZoneName(config, zone);
|
return resolveZoneName(config, zone);
|
||||||
@ -148,17 +285,14 @@ export function TrackingDetails({
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// For video mode: convert to video-relative time and seek player
|
// For video mode: convert to video-relative time (accounting for motion-only gaps)
|
||||||
const eventStartRecord =
|
const relativeTime = timestampToVideoTime(targetTimeRecord);
|
||||||
(event.start_time ?? 0) + annotationOffset / 1000;
|
|
||||||
const videoStartTime = eventStartRecord - REVIEW_PADDING;
|
|
||||||
const relativeTime = targetTimeRecord - videoStartTime;
|
|
||||||
|
|
||||||
if (videoRef.current) {
|
if (videoRef.current) {
|
||||||
videoRef.current.currentTime = relativeTime;
|
videoRef.current.currentTime = relativeTime;
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
[event.start_time, annotationOffset, displaySource],
|
[annotationOffset, displaySource, timestampToVideoTime],
|
||||||
);
|
);
|
||||||
|
|
||||||
const formattedStart = config
|
const formattedStart = config
|
||||||
@ -211,24 +345,14 @@ export function TrackingDetails({
|
|||||||
}
|
}
|
||||||
|
|
||||||
// seekToTimestamp is a record stream timestamp
|
// seekToTimestamp is a record stream timestamp
|
||||||
// event.start_time is detect stream time, convert to record
|
// Convert to video position (accounting for motion-only recording gaps)
|
||||||
// The video clip starts at (eventStartRecord - REVIEW_PADDING)
|
|
||||||
if (!videoRef.current) return;
|
if (!videoRef.current) return;
|
||||||
const eventStartRecord = event.start_time + annotationOffset / 1000;
|
const relativeTime = timestampToVideoTime(seekToTimestamp);
|
||||||
const videoStartTime = eventStartRecord - REVIEW_PADDING;
|
|
||||||
const relativeTime = seekToTimestamp - videoStartTime;
|
|
||||||
if (relativeTime >= 0) {
|
if (relativeTime >= 0) {
|
||||||
videoRef.current.currentTime = relativeTime;
|
videoRef.current.currentTime = relativeTime;
|
||||||
}
|
}
|
||||||
setSeekToTimestamp(null);
|
setSeekToTimestamp(null);
|
||||||
}, [
|
}, [seekToTimestamp, displaySource, timestampToVideoTime]);
|
||||||
seekToTimestamp,
|
|
||||||
event.start_time,
|
|
||||||
annotationOffset,
|
|
||||||
apiHost,
|
|
||||||
event.camera,
|
|
||||||
displaySource,
|
|
||||||
]);
|
|
||||||
|
|
||||||
const isWithinEventRange = useMemo(() => {
|
const isWithinEventRange = useMemo(() => {
|
||||||
if (effectiveTime === undefined || event.start_time === undefined) {
|
if (effectiveTime === undefined || event.start_time === undefined) {
|
||||||
@ -335,14 +459,13 @@ export function TrackingDetails({
|
|||||||
|
|
||||||
const handleTimeUpdate = useCallback(
|
const handleTimeUpdate = useCallback(
|
||||||
(time: number) => {
|
(time: number) => {
|
||||||
// event.start_time is detect stream time, convert to record
|
// Convert video player time back to timeline timestamp
|
||||||
const eventStartRecord = event.start_time + annotationOffset / 1000;
|
// accounting for motion-only recording gaps
|
||||||
const videoStartTime = eventStartRecord - REVIEW_PADDING;
|
const absoluteTime = videoTimeToTimestamp(time);
|
||||||
const absoluteTime = time + videoStartTime;
|
|
||||||
|
|
||||||
setCurrentTime(absoluteTime);
|
setCurrentTime(absoluteTime);
|
||||||
},
|
},
|
||||||
[event.start_time, annotationOffset],
|
[videoTimeToTimestamp],
|
||||||
);
|
);
|
||||||
|
|
||||||
const [src, setSrc] = useState(
|
const [src, setSrc] = useState(
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user