Update benchmark

This commit is contained in:
Nicolas Mowen 2025-09-23 13:05:25 -06:00
parent b9001ae2f0
commit ec4ca54b85

View File

@ -180,7 +180,7 @@ Inference speeds will vary greatly depending on the GPU and the model used.
✅ - Accelerated with CUDA Graphs ✅ - Accelerated with CUDA Graphs
❌ - Not accelerated with CUDA Graphs ❌ - Not accelerated with CUDA Graphs
| Name | ✅ YOLOv9 Inference Time | ✅ RF-DETR Inference Time | ❌ YOLO-NAS Inference Time | Name | ✅ YOLOv9 Inference Time | ✅ RF-DETR Inference Time | ❌ YOLO-NAS Inference Time
| --------------- | ------------------------ | ------------------------- | -------------------------- | | --------------- | ------------------------ | ------------------------- | -------------------------- |
| RTX 3050 | t-320: 8 ms s-320: 10 ms | Nano-320: ~ 12 ms | 320: ~ 10 ms 640: ~ 16 ms | | RTX 3050 | t-320: 8 ms s-320: 10 ms | Nano-320: ~ 12 ms | 320: ~ 10 ms 640: ~ 16 ms |
| RTX 3070 | t-320: 6 ms s-320: 8 ms | Nano-320: ~ 9 ms | 320: ~ 8 ms 640: ~ 14 ms | | RTX 3070 | t-320: 6 ms s-320: 8 ms | Nano-320: ~ 9 ms | 320: ~ 8 ms 640: ~ 14 ms |
@ -197,10 +197,11 @@ Apple Silicon can not run within a container, so a ZMQ proxy is utilized to comm
::: :::
| Name | YOLOv9 Inference Time | | Name | YOLOv9 Inference Time |
| --------- | ---------------------- | | --------- | ------------------------------------ |
| M3 Pro | t-320: 6 ms s-320: 8ms | | M4 | s-20: 10 ms |
| M1 | s-320: 9ms | | M3 Pro | t-320: 6 ms s-320: 8 ms s-640: 20 ms |
| M1 | s-320: 9ms |
### ROCm - AMD GPU ### ROCm - AMD GPU
@ -234,7 +235,7 @@ The MX3 is a pipelined architecture, where the maximum frames per second support
| YOLOv9s | 640 | ~ 41 ms | ~ 110 | | YOLOv9s | 640 | ~ 41 ms | ~ 110 |
| YOLOX-Small | 640 | ~ 16 ms | ~ 263 | | YOLOX-Small | 640 | ~ 16 ms | ~ 263 |
| SSDlite MobileNet v2 | 320 | ~ 5 ms | ~ 1056 | | SSDlite MobileNet v2 | 320 | ~ 5 ms | ~ 1056 |
Inference speeds may vary depending on the host platform. The above data was measured on an **Intel 13700 CPU**. Platforms like Raspberry Pi, Orange Pi, and other ARM-based SBCs have different levels of processing capability, which may limit total FPS. Inference speeds may vary depending on the host platform. The above data was measured on an **Intel 13700 CPU**. Platforms like Raspberry Pi, Orange Pi, and other ARM-based SBCs have different levels of processing capability, which may limit total FPS.
### Nvidia Jetson ### Nvidia Jetson