ThasnimolVSam22d007 jakaskerl i saw this sir, but this is by saving video or directly?? bcausae when i pass arguments it is showing the frame, but how come i know which kind of results it is saving instaed of showing video to me?
jakaskerl Hi ThasnimolVSam22d007 This example works by running gaze estimation on either live camera feed or pre-saved video on your machine. Either way, nothing gets saved back to your machine. It is only used to display the results. -cam will use your camera and run neural network models on the live video; -vid will load a saved video file and use it instead of live video from camera Hope this helps, Jaka
ThasnimolVSam22d007 jakaskerl Hi, i have one more query, this camera is actually for edge computing knw (means need to do all processing in camera itself), i am using gaze estimation code and face detection codes, but how can i verify whether that code processing is happening in camera ? not consuming cpu??
erik Hi ThasnimolVSam22d007 , You can just check in the Task Manager how much the python script consumes CPU/GPU. All the AI happens on the camera itself for this demo. Note that it sends the results back to the host computer for decoding, and for that I already started working on PR here that also does decoding+cropping on the device (via script node). Thoughts? Thanks, Erik
ThasnimolVSam22d007 erik the pr link you given know? i didnt get any data from that page with that code i can only preview , i tried to save video but not saving bbox in it, so if i want to save result only to the file, how to do it?? is it possible by using openvino mobile ssd models only detecting person, and some other objects, is it possible to detect face,shoes, goggles and all using it? if you have any ideas regarding this just let me know
ThasnimolVSam22d007 erik i tried this today, but that code with preview video is fully consuming cpu, after running the code consumption is incresed to 30,35 and all, so i am trying to save the results, do you have any ideas for it?? how to take output
erik Hi ThasnimolVSam22d007 , It's just a work in progress PR. Yes, you can save only the video. See DepthAI SDK's Recording module: https://docs.luxonis.com/projects/sdk/en/latest/features/recording/ You'd need to check the objects the pretrained mobilenet-ssd was trained on - I don't think it detects shoes and goggles. You could find a different pretrained model, or train your own neural network.
ThasnimolVSam22d007 erik no i was asked, what is the purpose that link i couldnt find anything there okay images am saving actually, in that color camera code but size is set to 300, when am trying to change it it is showing error only not changing the size? is possible to change the size of image or fixed??i mean is it possible to increase the field of view of image from where i should get the objects pretrained in that network obkjects from that model, if you have any links please share here, am checking it also
ThasnimolVSam22d007 jakaskerl means in that face detection code in github , using one model for object detection, so i want to know what all models are already defined in it, if it is not i want to include those things in the model, whether it is possible to add it in the same model,is it possible do you know whether anyone tried this, or any pretrained models with shoes, goggles, person, face and all already available?? also erik i tried that recording code it is saving some .dat file what is the advantage of those files?/ where i can use that??
jakaskerl Hi ThasnimolVSam22d007 Its a link to a pull request which has the code that aims to do decoding/cropping via script node. Your preview img size is (usually) set to whatever your NN expects. You can instead save video/still/isp to get a larger frame size. Info here. Basically have one stream (preview) go through NN, and the second (video) go directly through xlink/encoder to a file. i'm sorry, what object are you referring to? Thanks, Jaka
jakaskerl Hi ThasnimolVSam22d007 You can add your own classes to an already existing/trained model, by retraining it from a checkpoint along with your new training data. I suggest you look up a few tutorials. I've seen shoes, person, face (haven't yet seen one for goggles), but all as separate models. Your best bet is looking for something like yolo, which usually already has ~ 80 classes. Hope this helps, Jaka
ThasnimolVSam22d007 jakaskerl means, currently am using the code which is already given in the example, so it is using mobilenet ssd, so your saying i need train object and add to the same model or like mobilenet i saw yolo, we can use it same way as yolo in code??
jakaskerl Hi ThasnimolVSam22d007 If your model (be it SSD or yolo) supports retraining with new classes, then yes. Otherwise you will have to freshly train those models to detect your objects. Thanks, Jaka
ThasnimolVSam22d007 @jakaskerl , @erik i have one doubt, any custom trained model will work with oak d camera, now am trying hand gesture detection with yolo8, do you have any idea on the process, actually with this oak d any random model will work or any restriction is there?
erik ThasnimolVSam22d007 yes, as long as operations are supported then you can run the model on OAK. For hand gestures, I would strongly recommend building on top of this project: https://github.com/geaxgx/depthai_hand_tracker
ThasnimolVSam22d007 @erik here in this code which model is using?? do you have any custom model example with yolo - then it will be helpful for me , then i can add more things on it??
erik ThasnimolVSam22d007 it's mediapipe, as written in readme (readme is there for a reason 🙂). Yes, we have custom yolo training, conversion and deployment notebooks here: https://github.com/luxonis/depthai-ml-training/tree/master/colab-notebooks