Hey guys! Is it possible to trigger any OAK camera with HW signal? To make synchronization with external devices? Any idea to expand HW to add it? And any info about capturing delay in each cycle?
HW Trigger on OAK hw
- Edited
Yes technically there is an FSIN input that allows hardware-triggering as many stereo cameras as you'd like together.
That said it is significantly easier to use our software sync:
https://github.com/luxonis/depthai-experiments/tree/master/gen2-syncing
If your are building custom hardware you can design you could tie all the hardware camera sync pins together.
Also with our coming 4-port FFC, there is a connector on the board that has the sync pin exposed for this purpose:
https://github.com/luxonis/depthai-hardware/issues/85#issuecomment-909450066
Thanks,
Brandon
Hello, any tuto on where is this FSIN input and how it synchronises with the software ?
OK, I am not used to Discord .
I did send an explaination on DIscord (Harware / Hardware triggering / FranckMpa
Brandon the link above is broken. Here is the updated link: https://github.com/luxonis/depthai-experiments/tree/master/gen2-syncing
Also, a couple questions:
1) is there anything in that code example that is causing software sync? Does it happen automatically?
2) How accurate can I expect software sync to be? The pictures in that demo show time resolution of milliseconds, but I assume your monitor isn't refreshing at 1000 FPS so accuracy of this measurement isn't actually to the millisecond. It seems the monitor is likely running at 60fps or 120fps which means that we can trust the devices are in sync to within 16 ms or 8 ms. Do you have any more precise data on just how close the multi-device synchronization is?
Thanks
Thanks. Fixed it above. And for the questions, tagging @Luxonis-Alex .
@zdwiel
1) The software sync for the camera modules within one device (rgb, left, right) happens automatically. The mechanism is explained here:
https://docs.luxonis.com/projects/hardware/en/latest/pages/guides/sync_frames.html?#software-soft-sync
Software “soft” sync
Through firmware sync, we’re monitoring for drift and aligning the capture timestamps of the cameras, which are taken at the MIPI Start-of-Frame event. The Left/Right global shutter cameras are driven by the same clock, started by broadcast write on I2C, so no drift will happen over time, even when running freely without a sync. With the above functionality it would be also possible to configure FSIN as an output on one sensor, and an input to the other sensor. The RGB rolling shutter has a slight difference in clocking/frame-time, so when we detect a small drift, we’re modifying the frame time (number of lines) for the next frame by a small amount to compensate.
All 3 cameras are soft-synced by default using the above method, as long as they are configured with the same FPS (default is 30).
For multiple devices, currently we don't have a way to do soft sync. host-multiple-OAK-sync.py
is actually just matching frames on host as close as possible based on their timestamps.
But we were thinking to implement a soft sync mechanism for multiple devices connected to the same host as: a time-sync mechanism already runs in the background, and all devices are aware of the host time. Two options from here:
- each device will try to align the capture time for its RGB/L/R cameras at host time intervals multiple of a period, based on the configured FPS. For example for an FPS of 20, the interval would be 50ms, and the devices will try to align the capture at host timestamps like: 1000.2s, 1000.25s, 1000.3s, ...
- the host could monitor the frame timestamps from each device, and decide to issue frame-time adjustment commands (+/- a small time interval) for cameras that are drifting, that would be applied once (one frame-time is corrected, then the camera goes back to its configured frame-time -- based on FPS).
2) As you said, capturing the stopwatch shown on the monitor isn't a good way to measure how good the sync is, it was added just to show that the frames captured are actually in sync, and not far apart by a large number. For each camera, we do capture a timestamp on device and attach it as metadata to the ImgFrame
object. You can check the timestamp diff between cameras as here:
https://github.com/luxonis/depthai-experiments/blob/master/gen2-syncing/README.md#output-with-logging-enabled
Seq Left_tstamp RGB-Left Right-Left Dropped
num [seconds] diff[ms] diff[ms] delta
0 0.055592 -0.785 0.017
1 0.088907 -0.771 0.011
2 0.122222 -0.758 0.009
3 0.155537 -0.745 0.010
4 0.188852 -0.340 0.011
5 0.222167 -0.326 0.011
6 0.255482 0.063 0.010
7 0.288796 0.078 0.010
8 0.322111 0.257 0.010
9 0.355426 0.270 0.009
10 0.388741 0.246 0.010
11 0.422056 0.260 0.011
12 0.455371 0.146 0.009
13 0.488686 0.160 0.009
14 0.522000 0.046 0.010
15 0.555315 0.061 0.010
16 0.588644 -0.015 0.010
17 0.621945 0.020 0.011
18 0.655267 -0.011 0.008
19 0.688575 0.019 0.010
20 0.721890 0.018 0.009
...
Luxonis-Alex Has this soft sync mechanism for multiple devices connected to the same host been implemented? Both of your ideas sound great. Or can we implement this ourselves, using script nodes or some other lower level access?
Note that I have tried using FSYNC for this purpose, but failed to get the FSYNC example to work: I adapted Luxonis OAK camera trigger FSYNC pin via Script node for DM9098 board (OAK-D S2/Pro/W/Pro W) (github.com) to OAK-1 (disabling left/right) and get the message [184430100113480E00] [1.1.4] [2.306] [Script(7)] [warning] Trigger successful
- but then no frame is received. This could be an error caused by my adaptation attempt, … or is the FSYNC pin even functional on OAK-1?
Any help would be greatly appreciated because I otherwise have to give up my expensive experiment and resort to PTP-synced industrial vision cameras.