Stereo Neural Inference

In this mode, the neural inference (object detection, landmark detection, etc.) is run on the left and right cameras to produce stereo inference results. Unlike monocular neural inference fused with stereo depth - there is no max disparity search limit - so the minimum distance is purely limited by the greater of (a) horizontal field of view (HFOV) of the stereo cameras themselves and (b) the hyperfocal distance of the cameras.

The hyperfocal distance of the global shutter synchronized stereo pair is 19.6cm. So objects closer than 19.6cm will appear out of focus. This is effectively the minimum distance for this mode of operation, as in most cases (except for very wide stereo baselines with the OAK-FFC-3P-OG), this effective minimum distance is higher than the actual minimum distance as a result of the stereo camera field of views. For example, the objects will be fully out of the field of view of both grayscale cameras when less than 5.25cm (marked M on the picture below) from the OAK-D, but that is closer than the hyperfocal distance of the grayscale cameras (which is 19.6cm, marked as Y), so the actual minimum depth is this hyperfocal distance.

Minimum perceiving distance

Accordingly, to calculate the minimum distance for this mode of operation, use the following formula:

min_distance = max(tan((90 - HFOV/2) * pi/2) * base_line_dist/2, 19.6)

This formula implements the maximum of the HFOV-imposed minimum distance, and 19.6cm, which is the hyperfocal-distance-imposed minimum distance.


Triangulation Demo

For more infromation check out the gen2-triangulation demo that performs the stereo neural interface.

Got questions?

We’re always happy to help with code or other questions you might have.