diff --git a/frigate/detectors/detection_runners.py b/frigate/detectors/detection_runners.py index 8911c26ed..a034d2379 100644 --- a/frigate/detectors/detection_runners.py +++ b/frigate/detectors/detection_runners.py @@ -391,7 +391,9 @@ def get_optimized_runner( return RKNNModelRunner(rknn_path) if device != "CPU" and is_openvino_gpu_npu_available(): - return OpenVINOModelRunner(model_path, device or "AUTO", complex_model, **kwargs) + return OpenVINOModelRunner( + model_path, device or "AUTO", complex_model, **kwargs + ) providers, options = get_ort_providers(device == "CPU", device, **kwargs) ortSession = ort.InferenceSession( diff --git a/frigate/detectors/plugins/openvino.py b/frigate/detectors/plugins/openvino.py index 9bb5f9ee3..789912b4a 100644 --- a/frigate/detectors/plugins/openvino.py +++ b/frigate/detectors/plugins/openvino.py @@ -43,7 +43,9 @@ class OvDetector(DetectionApi): self.w = detector_config.model.width self.runner = OpenVINOModelRunner( - model_path=detector_config.model.path, device=detector_config.device, complex_model=False + model_path=detector_config.model.path, + device=detector_config.device, + complex_model=False, ) # For dfine models, also pre-allocate target sizes tensor