The custom NN information you provided will help as it will allow most of computation to occur on the device side (Intel Myriad) and to pipe the custom NN output to the video encoder . This leads to another question, does depthAI’s API support linking custom NN with two inputs? For example, in “rgb_mobilenet_4k.py” line 52 ( camRgb.preview.link(nn.input)), shows piping the RGB data into NN node with one input. However, PyTorch supports the ability to create models (custom NN) with two inputs ( https://discuss.pytorch.org/t/nn-module-with-multiple-inputs/237 ). The inputs of the custom NN would be the detection results of a object detection model and the camera's RGB data.
By the way, I just want to say the team, product, documentation, and examples are amazing. I am an embedded SW engineer with zero computer vision experience and less than a week I was able to get a custom object detection model working that allowed me to annotate a video stream in real time.
Thanks!