Depth Perception
Depth perception is a key feature of the DepthAI platform, which supports different methods of perceiving depth:- Passive Stereo Depth - Used by non-Pro version of OAK-D cameras.
- Active Stereo Depth - Used by Pro version of OAK-D cameras.
- Time-of-Flight Depth - Used by OAK-ToF.
Passive Stereo Depth
Passive stereo works similarly to human vision. Our brains estimate the depth of objects based on the difference in perception between our left and right eyes. In OAK-D cameras, a stereo camera pair (left and right monocular cameras) performs a similar function. The VPU (Visual Processing Unit) of the OAK cameras does disparity matching to estimate the depth.Disparity is the pixel distance between the same point in the left and right images of the stereo pair camera. The OAK-D camera calculates disparity for every pixel in the mono frame, assigning a disparity value with some confidence level. This process occurs inside the StereoDepth node. The depth map is calculated from the disparity map using this formula.Passive stereo depth perception may not work well with featureless surfaces like walls or ceilings, as disparity matching relies on distinctive points or features.
- Lighting/Texture: Stereo depth depends on feature matching, which can be challenging in low light or on featureless surfaces. Active stereo can resolve both texture and lighting issues.
- Calibration: Factory calibration is typically optimal.
- Postprocessing filters: More information can be found here (under Depth Filters). Additional filtering can be performed on the host side as well, e.g., WLS filter.