Set up your development environment by following the instructions in getting started.
(Optional) Install dependencies for any sensors you want to use by following the sensor-specific guides.
Note
We strongly recommend installing all sensor dependencies before starting any quickstarts.
Some sensor dependencies require restarting the development environment during installation, which will interrupt the quickstart process.
FoundationStereo supports two fixed resolution configurations:
high_res: 576x960 resolution (default)
low_res: 320x736 resolution (original training resolution)
You can set the default model resolution using the FOUNDATIONSTEREO_MODEL_RES
environment variable. Use the --model_res argument to override the default or
to explicitly select between these options.
The high_res (576x960) model requires more than 16GB of VRAM, so it only works on GPUs with more than 16GB of VRAM.
FoundationStereo supports two fixed resolution configurations:
high_res: 576x960 resolution (default)
low_res: 320x736 resolution (original training resolution)
You can set the default model resolution using the FOUNDATIONSTEREO_MODEL_RES
environment variable. Use the --model_res argument to override the default or
to explicitly select between these options.
The high_res (576x960) model requires more than 16GB of VRAM, so it only works on GPUs with more than 16GB of VRAM.
This tutorial requires a compatible RealSense camera from
the list of available
cameras.
Ensure that you have already set up your RealSense camera using the RealSense setup tutorial. If you have not, set up the sensor and then restart this quickstart from the beginning.
Continuing inside the Isaac ROS environment, install the following dependencies:
Complete steps to set up the FoundationStereo model as described in the
quickstart.
Continuing inside the Isaac ROS environment, run the following launch file to spin up a demo using a RealSense stereo camera:
Note
The RealSense camera may be unstable and the streams may drop with the high resolution model within a minute after starting. If you encounter issues, try the low resolution model or running the camera at a lower frame rate (for example, 15 FPS).
The isaac_ros_foundationstereo package offers functionality to generate a stereo
disparity map from stereo images using a trained FoundationStereo model. Given a pair
of stereo input images, the package generates a continuous disparity
image for the left input image. The package consists of the following node:
FoundationStereoDecoderNode: Processes the model output and generates the disparity map
The isaac_ros_foundationstereo package outputs a disparity image with dimension same as the FoundationStereo model output dimension.
The input images are rescaled and padded to the FoundationStereo model input dimension
before inferencing. The disparity output is published as a continuous disparity map.
The disparity output is filtered to remove invalid values:
- Values below min_disparity (default: 0.0)
- Values above max_disparity (default: 10000.0)
- Invalid regions (inf, Nan) are set to 0.0
The right CameraInfo is used to composite a
NitrosDisparityImage. If you only care about the disparity
image, and don’t need the baseline and focal length information, you
can pass dummy camera messages.