From b4ba2ce7c453adc0fd780b2ddbb252b005b9382c Mon Sep 17 00:00:00 2001 From: Nicolas Mowen Date: Fri, 31 Oct 2025 12:26:18 -0600 Subject: [PATCH] Don't use GPU for training --- docker/tensorrt/requirements-amd64.txt | 1 - .../configuration/custom_classification/object_classification.md | 1 - .../configuration/custom_classification/state_classification.md | 1 - 3 files changed, 3 deletions(-) diff --git a/docker/tensorrt/requirements-amd64.txt b/docker/tensorrt/requirements-amd64.txt index a7853aeec..63c68b583 100644 --- a/docker/tensorrt/requirements-amd64.txt +++ b/docker/tensorrt/requirements-amd64.txt @@ -13,7 +13,6 @@ nvidia_cusolver_cu12==11.6.3.*; platform_machine == 'x86_64' nvidia_cusparse_cu12==12.5.1.*; platform_machine == 'x86_64' nvidia_nccl_cu12==2.23.4; platform_machine == 'x86_64' nvidia_nvjitlink_cu12==12.5.82; platform_machine == 'x86_64' -tensorflow==2.19.*; platform_machine == 'x86_64' onnx==1.16.*; platform_machine == 'x86_64' onnxruntime-gpu==1.22.*; platform_machine == 'x86_64' protobuf==3.20.3; platform_machine == 'x86_64' diff --git a/docs/docs/configuration/custom_classification/object_classification.md b/docs/docs/configuration/custom_classification/object_classification.md index 983fce852..cff8a6cad 100644 --- a/docs/docs/configuration/custom_classification/object_classification.md +++ b/docs/docs/configuration/custom_classification/object_classification.md @@ -10,7 +10,6 @@ Object classification allows you to train a custom MobileNetV2 classification mo Object classification models are lightweight and run very fast on CPU. Inference should be usable on virtually any machine that can run Frigate. Training the model does briefly use a high amount of system resources for about 1–3 minutes per training run. On lower-power devices, training may take longer. -When running the `-tensorrt` image, Nvidia GPUs will automatically be used to accelerate training. ## Classes diff --git a/docs/docs/configuration/custom_classification/state_classification.md b/docs/docs/configuration/custom_classification/state_classification.md index c22661f26..caaeaed5a 100644 --- a/docs/docs/configuration/custom_classification/state_classification.md +++ b/docs/docs/configuration/custom_classification/state_classification.md @@ -10,7 +10,6 @@ State classification allows you to train a custom MobileNetV2 classification mod State classification models are lightweight and run very fast on CPU. Inference should be usable on virtually any machine that can run Frigate. Training the model does briefly use a high amount of system resources for about 1–3 minutes per training run. On lower-power devices, training may take longer. -When running the `-tensorrt` image, Nvidia GPUs will automatically be used to accelerate training. ## Classes