You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
on RVC will result in lower latency, but also in lower FPS.
128
+
129
+
By default, NN nodes are running 2 threads, 1 NCE/thread, and we suggest compiling the model for half of the
130
+
available SHAVE cores of the pipeline. This configuration will provide best throughput, as all threads can run freely.
131
+
Compiling the model for more SHAVE cores will only provide marginal improvement, due to:
132
+
133
+
1. `Model optimizer <https://docs.luxonis.com/en/latest/pages/model_conversion/#model-optimizer>`__ doing a great work at optimizing the model
134
+
2. On-device parallelization of NN operations (splitting the operation task between multiple SHAVEs) doesn't scale linearly due to " `memory wall <https://en.wikipedia.org/wiki/Random-access_memory#Memory_wall>`__ "
135
+
136
+
To minimize the latency, though, we should allocate all resources to the single inference. To get lowest latency (which will result in much lower FPS),
137
+
we suggest the following:
138
+
139
+
- Setting the number of threads to 1
140
+
- Setting the number of NCE per thread to 2
141
+
- Compiling the model for all available SHAVE cores - `documentation here <https://docs.luxonis.com/en/latest/pages/model_conversion/#compile-tool>`__)
142
+
143
+
.. code-block:: python
144
+
145
+
nn = pipeline.create(dai.node.NeuralNetwork)
146
+
# Same for Yolo/MobileNet (Spatial) Detection node
0 commit comments