I think I'm finally with the OAK-D and DepthAI in terms of ability to determine depth. I'd like now to move on to object detection. I've run many of the examples that combine object detection and depth, but what I really want to do is define a custom set of objects in my environment; initially they will be used to assist navigation of that environment.
I don't really understand enough about NNs, AI or ML to even get started, at least not without some extra help. Maybe pointers to some good references would be very helpful, but I've also got some specific questions below.
I've found the "Custom Training" page under Tutorials, and have opened and browsed through the "Easy Object Detector Training" Colab. I have not yet had the guts to run it. That Colab suggests you can train using your own images. But, they have to be annotated. I did a bit of research and it seems that there may be different forms of annotation. And it is quite clear there are many different annotation tools. So, the I have to ask
- What is the form of annotation required to run the Colab successfully?
- Are there annotation tools that produce this format? Are there free tools that do so?
I'd also appreciate some guidelines on the training, validation, and final test images. I've read that for any object there should be images from different views, different lighting conditions, probably other things.
- What factors really matter?
- Also, how many images of a single object are sufficient. Generally there will be significant differences between most of the the objects I want to detect; I'm not sure if that matters.
Thanks very much for any assistance.