diff --git a/frigate/data_processing/real_time/whisper_online.py b/frigate/data_processing/real_time/whisper_online.py index 9b81d7fbe..024b19fba 100644 --- a/frigate/data_processing/real_time/whisper_online.py +++ b/frigate/data_processing/real_time/whisper_online.py @@ -141,6 +141,8 @@ class FasterWhisperASR(ASRBase): def transcribe(self, audio, init_prompt=""): from faster_whisper import BatchedInferencePipeline + logging.getLogger("faster_whisper").setLevel(logging.WARNING) + # tested: beam_size=5 is faster and better than 1 (on one 200 second document from En ESIC, min chunk 0.01) batched_model = BatchedInferencePipeline(model=self.model) segments, info = batched_model.transcribe(