From e60b51d036555efcdb08ef355a1f5c508341e61b Mon Sep 17 00:00:00 2001 From: Josh Hawkins <32435876+hawkeye217@users.noreply.github.com> Date: Mon, 26 May 2025 07:49:27 -0500 Subject: [PATCH] fix linter --- .../data_processing/real_time/whisper_online.py | 15 +++------------ 1 file changed, 3 insertions(+), 12 deletions(-) diff --git a/frigate/data_processing/real_time/whisper_online.py b/frigate/data_processing/real_time/whisper_online.py index d9bfb0348..6122dbdd9 100644 --- a/frigate/data_processing/real_time/whisper_online.py +++ b/frigate/data_processing/real_time/whisper_online.py @@ -119,15 +119,6 @@ class FasterWhisperASR(ASRBase): from faster_whisper import WhisperModel logging.getLogger("faster_whisper").setLevel(logging.WARNING) - if model_dir is not None: - logger.debug( - f"Loading whisper model from model_dir {model_dir}. modelsize and cache_dir parameters are not used." - ) - model_size_or_path = model_dir - elif modelsize is not None: - model_size_or_path = modelsize - else: - raise ValueError("modelsize or model_dir parameter must be set") # this worked fast and reliably on NVIDIA L40 model = WhisperModel( @@ -514,10 +505,10 @@ class OnlineASRProcessor: p = self.commited[:k] p = [t for _, _, t in p] prompt = [] - l = 0 - while p and l < 200: # 200 characters prompt size + y = 0 + while p and y < 200: # 200 characters prompt size x = p.pop(-1) - l += len(x) + 1 + y += len(x) + 1 prompt.append(x) non_prompt = self.commited[k:] return self.asr.sep.join(prompt[::-1]), self.asr.sep.join(