zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

zivid_camera repository

Repository Summary

Checkout URI https://github.com/zivid/zivid-ros.git
VCS Type git
VCS Version master
Last Updated 2025-04-03
Dev Status MAINTAINED
CI status No Continuous Integration
Released UNRELEASED
Tags No category tags.
Contributing Help Wanted (0)
Good First Issues (0)
Pull Requests to Review (0)

Packages

Name Version
zivid_camera 3.1.0
zivid_interfaces 3.1.0
zivid_rviz_plugin 3.0.0
zivid_samples 3.1.0

README

Zivid ROS driver

This is the official ROS driver for Zivid 3D cameras.

This driver provides support for ROS2.

If you are looking for the Zivid ROS1 driver, please use the ros1-master branch if you are using Zivid SDK 2.13 or older, or use the ros1-sdk-2.14.0 branch if you are using Zivid SDK 2.14 or newer.

Build Status Zivid Image


Contents: Installation | Getting Started | Launching | Configuration | Services | Topics | Samples | RViz Plugin | FAQ


Installation

Support

This driver supports Ubuntu 20.04 / 22.04 / 24.04 with ROS2. Follow the official ROS installation instructions for your OS.

If you are looking for the Zivid ROS1 driver, please use the ros1-master branch for Zivid SDK 2.13 or older, or use the ros1-sdk-2.14.0 branch for Zivid SDK 2.14 or newer.

Zivid SDK

To use the ROS driver you need to download and install the “Zivid Core” package. Zivid SDK version 2.15.0 is supported. See releases for older ROS driver releases that supports older SDK versions.

Follow this guide to install “Zivid Core” for your version of Ubuntu. The “Zivid Studio” and “Zivid Tools” packages can be useful to test your system setup and camera, but are not required by the driver.

An OpenCL 1.2 compatible GPU with driver installed is required by the SDK. Follow this guide to install OpenCL drivers for your system.

C++ compiler

A C++17 compiler is required.

sudo apt-get install -y g++

Downloading and building Zivid ROS driver

First, source the setup.bash script for your ROS distribution in your terminal:

source /opt/ros/$ROS_DISTRO/setup.bash

Then create the workspace and src directory:

mkdir -p ~/ros2_ws/src

Clone the Zivid ROS project into the src directory:

cd ~/ros2_ws/src
git clone https://github.com/zivid/zivid-ros.git

Initialize rosdep:

cd ~/ros2_ws/src
sudo rosdep init
rosdep update

Install dependencies:

cd ~/ros2_ws/src
rosdep install -i --from-path ./ -y

Finally, build the driver:

cd ~/ros2_ws
colcon build --symlink-install

Getting started

Connect the Zivid camera to your PC. You can use the ZividListCameras command-line tool available in the “Zivid Tools” package to confirm that your system has been configured correctly, and that the camera is discovered by your PC. You can also open Zivid Studio and connect to the camera. Close Zivid Studio before continuing with the rest of this guide.

Run the sample_capture_cpp via the launch script to check that everything is working.

cd ~/ros2_ws && source install/setup.bash
ros2 launch zivid_samples sample_with_rviz.launch sample:=sample_capture_cpp

This will start the zivid_camera driver node, the sample_capture_cpp node, and rviz. The zivid_camera driver will connect to the first available Zivid camera, and then sample_capture_cpp will trigger captures repeatedly. The resulting point cloud and color image should be visible in rviz.

A more detailed description of the zivid_camera driver follows below.

For sample code, see the Samples section.

Launching the driver

To launch the driver, use ros2 run:

cd ~/ros2_ws && source install/setup.bash
ros2 run zivid_camera zivid_camera

The driver will by default connect to the first available Zivid camera. This behavior can be overridden by setting the serial_number launch parameter, see below.

Launch Parameters (advanced)

The following parameters can be specified when starting the driver. Note that all the parameters are optional.

For example, to run the zivid_camera driver with a specific serial_number specified:

ros2 run zivid_camera zivid_camera --ros-args -p serial_number:=ABCD1234

Or you can use a launch file and invoke it as:

ros2 launch zivid_samples zivid_camera_with_serial_number.launch serial_number:=ABCD1234

file_camera_path (string, default: “”)

Specify the path to a file camera to use instead of a real Zivid camera. This can be used to develop without access to hardware. The file camera returns the same point cloud for every capture. Visit our knowledgebase to download file camera.

frame_id (string, default: “zivid_optical_frame”)

Specify the frame_id used for all published images and point clouds.

serial_number (string, default: “”)

Specify the serial number of the Zivid camera to use. This parameter is optional. By default, the driver will connect to the first available camera.

update_firmware_automatically (bool, default: true)

Specify if the firmware of the connected camera should be automatically updated to the correct version when the Zivid ROS driver starts. If set to false, if the firmware version is out of date then camera must be manually updated, for example using Zivid Studio or ZividFirmwareUpdater. Read more about firmware update in our knowledgebase. This parameter is optional, and by default it is true.

Configuration

The capture settings used by the zivid_camera ROS driver must be configured using YAML, which can be exported from Zivid Studio or the API, or downloaded as .yml files from our knowledge base.

For convenience, the Zivid ROS driver supports configuring capture settings in two ways: Using file path to a .yml file, or as a YAML string.

The following ROS parameters control which settings are used when capturing with the driver. Note that you must set either the _file_path or the _yaml parameter. If both _file_path and _yaml parameters are set to a non-empty string at the same time, then the driver will return an error when capturing. By default, all settings parameters are empty.

3D capture

settings_file_path (string, default: “”)

Specify the path to a .yml file that contains the settings you want to use.

settings_yaml (string, default: “”)

Specify a YAML string that contains the settings you want to use. For example, you can copy the contents of a .yml file saved from Zivid Studio.

The service capture_assistant/suggest_settings will modify the settings parameters automatically.

2D capture

settings_2d_file_path (string, default: “”)

Specify the path to a .yml file that contains the settings you want to use.

settings_2d_yaml (string, default: “”)

Specify a YAML string that contains the 2D settings you want to use. For example, you can copy the contents of a .yml file saved from Zivid Studio.

Services

capture

std_srvs/srv/Trigger

Invoke this service to trigger a 3D capture. See section Configuration for how to configure the 3D capture settings. The resulting point cloud is published on topics points/xyz and points/xyzrgba, color image is published on topic color/image_color, and depth image is published on topic depth/image. Camera calibration is published on topics color/camera_info and depth/camera_info.

See Sample Capture for code example.

capture_2d

std_srvs/srv/Trigger

Invoke this service to trigger a 2D capture. See section Configuration for how to configure the 2D capture settings. The resulting 2D image is published to topic color/image_color. Note: 2D RGB image is also published as a part of 3D capture, see capture.

See Sample Capture 2D for code example.

capture_and_save

zivid_interfaces/srv/CaptureAndSave.srv

It does exactly the same as the capture service, in addition it will save the frame to a file. This service takes a path as an argument. The chosen format is detected via the file extension. See knowledge base for a list of available output formats.

See Sample Capture and Save for code example.

capture_assistant/suggest_settings

zivid_interfaces/srv/CaptureAssistantSuggestSettings.srv

Invoke this service to analyze your scene and find suggested settings for your particular scene, camera distance, ambient lighting conditions, etc. This service will automatically update the node parameter settings_yaml with the suggested settings, see section Configuration. When this service has returned, you can invoke the capture service to trigger a 3D capture using these suggested settings.

This service has two parameters:

max_capture_time (duration):

Specify the maximum capture time for the settings suggested by the Capture Assistant. A longer capture time may be required to get good data for more challenging scenes. Minimum value is 0.2 sec and maximum value is 10.0 sec.

ambient_light_frequency (uint8):

Possible values are: AMBIENT_LIGHT_FREQUENCY_NONE, AMBIENT_LIGHT_FREQUENCY_50HZ, AMBIENT_LIGHT_FREQUENCY_60HZ. Can be used to ensure that the suggested settings are compatible with the frequency of the ambient light in the scene. If ambient light is unproblematic, use AMBIENT_LIGHT_FREQUENCY_NONE for optimal performance. Default is AMBIENT_LIGHT_FREQUENCY_NONE.

See Sample Capture Assistant for code example.

camera_info/model_name

zivid_interfaces/srv/CameraInfoModelName.srv

Returns the camera’s model name.

camera_info/serial_number

zivid_interfaces/srv/CameraInfoSerialNumber.srv

Returns the camera’s serial number.

is_connected

zivid_interfaces/srv/IsConnected.srv

Returns if the camera is currently in Connected state (from the perspective of the ROS driver). The connection status is updated by the driver every 10 seconds and before each capture service call. If the camera is not in Connected state the driver will attempt to re-connect to the camera when it detects that the camera is available. This can happen if the camera is power-cycled or the USB/Ethernet cable is unplugged and then replugged.

hand_eye_calibration/start

zivid_interfaces/srv/HandEyeCalibrationStart.srv

Prepares the node for hand-eye calibration and clears any previously collected hand-eye captures. The type of the calibration object to be used during capture must be provided. A working directory can optionally be provided so that all captures are saved to this directory. If provided, the directory must be given as an absolute path and the directory must be empty.

This service must be called first before capturing data for hand-eye calibration. It can also be used to restart an active hand-eye calibration session. After calling this service, proceed with captures by calling the hand_eye_calibration/capture service.

Please refer to the Zivid knowledge base for more information on hand-eye calibration.

hand_eye_calibration/capture

zivid_interfaces/srv/HandEyeCalibrationCapture.srv

Performs a hand-eye calibration capture. This service takes the robot pose as an input. Then it performs a capture, and detects any calibration objects. The resulting point cloud and color image will be published just like during a normal call to the capture service.

If the detections is successful, the result is stored locally in the driver. Additionally, if a working directory was specified during start, the captured frame and robot pose is saved to that directory.

This service uses the currently stored 3D capture settings to perform the capture. Ensure that the camera is properly configured first. Please see the knowledge base for how to get good quality data.

The camera and robot should be appropriately positioned so that the calibration object is visible in the frame. Multiple captures in different poses are necessary. After performing several captures, one can proceed by calling the hand_eye_calibration/calibrate service.

hand_eye_calibration/calibrate

zivid_interfaces/srv/HandEyeCalibrationCalibrate.srv

Computes the hand-eye calibration transform based on the captures gathered during the current hand-eye calibration session. Both eye-to-hand and eye-in-hand configurations are supported. If successful, the computed hand-eye transform is returned, see the knowledge base for more information on the hand-eye calibration solution.

The calibration procedure requires all robot poses to be different. At least 2 poses are required when using a calibration board, or 6 poses when using fiducial markers. For fiducial markers, each marker must be detected across 2 poses at minimum.

Low degrees-of-freedom (DOF) calibration is also supported (experimental) by supplying the fixed placement of calibration objects. This is not needed for regular (6-DOF) calibration.

hand_eye_calibration/load

zivid_interfaces/srv/HandEyeCalibrationLoad.srv

Loads a working directory from a previous hand-eye calibration session. See the hand_eye_calibration/start service for how to start a session with a working directory. The captures and robot poses are loaded from the provided directory. The directory is opened as read-only, and no new captures can be made during this session. However, calls to the hand_eye_calibration/calibrate service can be made to compute the hand-eye transform from the loaded data.

infield_correction/read

zivid_interfaces/srv/InfieldCorrectionRead.srv

Returns the state of the infield correction of the camera.

infield_correction/reset

std_srvs/srv/Trigger

Resets the infield correction on the camera to factory settings.

infield_correction/start

std_srvs/srv/Trigger

Prepares for infield correction, and clears any infield correction captures gathered so far in the zivid_camera node.

This service must be called before gathering captures using the infield_correction/capture service. However, other infield correction services can be used without calling the start service first. This service can also be used to restart an active infield correction session.

Please refer to the Zivid knowledge base for further information and guidelines for infield correction.

infield_correction/capture

zivid_interfaces/srv/InfieldCorrectionCapture.srv

Takes a capture to be used for infield correction. Please point the camera at a Zivid infield calibration object. It is recommended to cover several distances, with one or more captures at each distance. Successful captures are stored in the zivid_camera node.

Before calling this service, a call to infield_correction/start must have been made first.

This service call will perform a relatively slow but high-quality point cloud capture with the connected camera. Any settings applied to the camera in the ROS driver will be ignored for calls to this service, appropriate settings are automatically used. The resulting point cloud and color image will be published just like during a normal call to the capture service.

After sufficient number of captures, proceed by calling the infield_correction/compute service to compute the verification metrics and the correction based on the captures gathered so far. To additionally write the correction to camera, proceed by calling the infield_correction/compute_and_write service.

The captured data is cleared if the node is stopped, or after a successful call to either of the services infield_correction/start or infield_correction/compute_and_write.

infield_correction/compute

zivid_interfaces/srv/InfieldCorrectionCompute.srv

Calculates the new infield correction based the captured data gathered so far through the service infield_correction/capture.

The quantity and range of data is up to the user, but generally a larger dataset will yield a more accurate and reliable correction. If all measurements are taken at approximately the same distance, the resulting correction will mainly be valid at those distances. If several measurements are taken at significantly different distances, the resulting correction will likely be more suitable for extrapolation to distances beyond where the dataset is collected.

This service also acts as verification of the quality of the infield correction on a camera, or the need for one if none exists already. It returns an indication of the dimension trueness at the location where the input data was captured.

If the returned assessment indicates a trueness error that is above the threshold for your application, consider using infield_correction/compute_and_write. The service also returns information regarding the proposed working range and the accuracy that can be expected within the working range, if the correction is later written to the camera.

infield_correction/compute_and_write

zivid_interfaces/srv/InfieldCorrectionCompute.srv

Calculates the new infield correction based the captured data gathered so far through the service infield_correction/capture, and writes the result to the camera.

If the write operation is successful, the infield correction capture data is cleared. To perform infield correction again, a new session must be started with a call to the infield_correction/start service.

Please see the infield_correction/compute service for more information on the computed correction.

infield_correction/remove_last_capture

std_srvs/srv/Trigger

Removes the last infield correction capture gathered in the zivid_camera node.

Topics

The Zivid ROS driver provides several topics providing 3D, color, SNR and camera calibration data as a result of calling capture/capture_2d services. The different output topics provides flexibility for different use cases. Note that for performance reasons no messages are generated or sent on topics with zero active subscribers.

color/camera_info

sensor_msgs/msg/CameraInfo

Camera calibration and metadata.

color/image_color

sensor_msgs/msg/Image

Color/RGBA image. RGBA image is published as a result of invoking the capture or capture_2d service. Images are encoded as “rgba8”, where the alpha channel is always 255.

depth/camera_info

sensor_msgs/msg/CameraInfo

Camera calibration and metadata.

depth/image

sensor_msgs/msg/Image

Depth image. Each pixel contains the z-value (along the camera Z axis) in meters. The image is encoded as 32-bit float. Pixels where z-value is missing are NaN.

points/xyzrgba

sensor_msgs/msg/PointCloud2

Point cloud data. Sent as a result of the capture service. The output is in the camera’s optical frame, where x is right, y is down and z is forward. The included point fields are x, y, z (in meters) and rgba (color).

points/xyz

sensor_msgs/msg/PointCloud2

Point cloud data. This topic is similar to topic points/xyzrgba, except that only the XYZ 3D coordinates are included. This topic is recommended if you don’t need the RGBA values.

snr/camera_info

sensor_msgs/msg/CameraInfo

Camera calibration and metadata.

snr/image

sensor_msgs/msg/Image

Each pixel contains the SNR (signal-to-noise ratio) value. The image is encoded as 32-bit float. Published as a part of the capture service.

normals/xyz

sensor_msgs/msg/PointCloud2

Normals for the point cloud. The included fields are normal x, y and z coordinates. Each coordinate is a float value. There are no additional padding floats, so point-step is 12 bytes (3*4 bytes). The normals are unit vectors. Note that subscribing to this topic will cause some additional processing time for calculating the normals.

Samples

In the zivid_samples package we have added samples that demonstrate how to use the Zivid ROS driver. These samples can be used as a starting point for your project.

To launch the Python samples using ros2 launch, you need python to be available as a command. For example, the python-is-python3 package can be installed to achieve this, by running the following command:

sudo apt install python-is-python3

On Windows, the Python samples cannot be launched using ros2 launch. Instead, please launch the samples using ros2 run zivid_samples <sample_name>.py together with ros2 run zivid_camera zivid_camera in another terminal window.

Sample Capture

This sample performs single-acquisition 3D captures in a loop. This sample shows how to configure the capture settings, how to subscribe to the points/xyzrgba topic, and how to invoke the capture service.

Source code: C++ Python

ros2 launch zivid_samples sample.launch sample:=sample_capture_cpp
ros2 launch zivid_samples sample.launch sample:=sample_capture.py

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_capture_cpp
ros2 run zivid_samples sample_capture.py

Sample Capture 2D

This sample performs single-acquisition 2D captures in a loop. This sample shows how to configure the capture 2d settings with a YAML string, how to subscribe to the color/image_color topic, and how to invoke the capture_2d service.

Source code: C++ Python

ros2 launch zivid_samples sample.launch sample:=sample_capture_2d_cpp
ros2 launch zivid_samples sample.launch sample:=sample_capture_2d.py

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_capture_2d_cpp
ros2 run zivid_samples sample_capture_2d.py

Sample Capture and Save

This sample performs a capture, and stores the resulting frame to file. This sample shows how to configure the capture settings with a YAML string, how to invoke the capture_and_save service, and how to read the response from the service call.

Source code: C++ Python

ros2 launch zivid_samples sample.launch sample:=sample_capture_and_save_cpp
ros2 launch zivid_samples sample.launch sample:=sample_capture_and_save.py

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_capture_and_save_cpp
ros2 run zivid_samples sample_capture_and_save.py

Sample Capture Assistant

This sample shows how to invoke the capture_assistant/suggest_settings service to suggest and set capture settings. Then, it shows how to subscribe to the points/xyzrgba and color/image_color topics, and finally invoke the capture service.

Source code: C++ Python

ros2 launch zivid_samples sample.launch sample:=sample_capture_assistant_cpp
ros2 launch zivid_samples sample.launch sample:=sample_capture_assistant.py

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_capture_assistant_cpp
ros2 run zivid_samples sample_capture_assistant.py

Sample Capture with Settings from File

This sample performs single-acquisition 3D captures in a loop. This sample shows how to configure the capture settings from a yaml file, how to subscribe to the points/xyzrgba topic, and how to invoke the capture service.

Source code: C++ Python

ros2 launch zivid_samples sample.launch sample:=sample_capture_with_settings_from_file_cpp
ros2 launch zivid_samples sample.launch sample:=sample_capture_with_settings_from_file.py

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_capture_with_settings_from_file_cpp
ros2 run zivid_samples sample_capture_with_settings_from_file.py

Sample Infield Correction

This sample shows how to invoke the various infield_correction/[…] services to perform infield correction on Zivid cameras.

Source code: C++

ros2 launch zivid_samples sample.launch sample:=sample_infield_correction_cpp operation:=<operation>

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_infield_correction_cpp --ros-args -p operation:=<operation>

With the following argument:

operation (string, required)

Specify the infield correction operation to perform, one of:

Please see the infield correction documentation for prerequisites and guidelines on how to perform the correction.

For a more interactive experience, we recommend using the infield correction panel from the Zivid RViz plugin.

The typical procedure for performing a new infield correction is:

  1. start: Prepare for infield correction, clears any existing infield correction captures.
  2. capture: Take multiple captures from different angles and distances in accordance with the typical operating conditions of the camera.
  3. compute: Check the computed correction results and the estimated errors, verify that they give satisfactory results.
  4. compute_and_write: Compute the correction and write the results to camera.

The zivid_camera node persists the infield correction dataset as long as it is running. To start the infield correction captures over again, use the start operation which clears all infield correction captures previously gathered. The remove_last_capture can be used to remove just the last capture.

Sample Hand-Eye Calibration

This sample shows how to invoke various hand_eye_calibration/[…] services to perform hand-eye calibration on Zivid cameras. The sample is for exposition only, to demonstrate how the services can be called.

The sample begins by preparing the camera node for hand-eye calibration. Then a fixed number of captures is gathered at a fixed duration between captures, using a simulated robot pose. Finally, hand-eye calibration is run using the gathered data.

Source code: C++

ros2 launch zivid_samples sample.launch sample:=sample_hand_eye_calibration_cpp configuration:=<configuration> marker_ids:=<marker_ids> working_directory:=<working_directory>

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_hand_eye_calibration_cpp --ros-args -p configuration:=<configuration> -p marker_ids:=<marker_ids> -p working_directory:=<working_directory>

With the following arguments:

configuration (string, required)

Specify the configuration for the hand-eye calibration, one of:

marker_ids (dynamic array of integers, default: empty):

Specifies a list of fiducial marker IDs used for detection (e.g. [1,2,3]), or empty if a Zivid calibration board is used instead.

working_directory (string, default: empty)

argument specifies the working directory used to store the gathered data, or empty to indicate that the data should not be stored on disk. A non-empty value must specify an absolute path to an empty directory.

For more information on performing the calibration, please see the Zivid hand-eye calibration documentation.

RViz Plugin

Infield Correction Panel

The Zivid RViz plugin provides a panel to interactively perform infield correction with a Zivid camera.

To see the panel in RViz, go to Panels -> Add New Panel, then select Zivid Infield Correction and click OK. Infield correction can now be performed by interacting with the newly added panel.

The panel is also visible when launching a Zivid sample with RViz, e.g.:

ros2 launch zivid_samples sample_with_rviz.launch sample:=sample_capture_with_settings_from_file_cpp

When the zivid_camera node is already running, RViz can also be started directly with the panel visible by using the Zivid RViz display configuration:

rviz2 -d $(ros2 pkg prefix zivid_camera)/share/zivid_camera/config/rviz/zivid.rviz

Please see the infield correction documentation for prerequisites and guidelines on how to perform the correction.

Frequently Asked Questions

How to visualize the output from the camera in rviz

ros2 launch zivid_camera visualize.launch

How to connect to specific Zivid serial number

ros2 run zivid_camera zivid_camera --ros-args -p serial_number:=ABCD1234

How to start the driver using settings.yml files

See section Configuration for more details.

ros2 run zivid_camera zivid_camera --ros-args -p settings_file_path:=/path/to/settings.yml -p settings_2d_file_path:=/path/to/settings2D.yml

How to trigger 3D/2D capture via terminal

ros2 service call /capture std_srvs/srv/Trigger
ros2 service call /capture_2d std_srvs/srv/Trigger

How to use a file camera

ros2 run zivid_camera zivid_camera --ros-args -p file_camera_path:=/usr/share/Zivid/data/FileCameraZivid2L100.zfc

Visit our knowledgebase to download file camera.

How to use multiple Zivid cameras

You can use multiple Zivid cameras simultaneously by starting one node per camera and specifying unique namespaces per node.

ros2 run zivid_camera zivid_camera --ros-args --remap __ns:=/zivid_camera1

ros2 run zivid_camera zivid_camera --ros-args --remap __ns:=/zivid_camera2

You can combine this with a serial_number parameter (see above) to control which node uses which camera. By default, the zivid_camera node will connect to the first available (unused) camera. We recommend that you first start the first node, then wait for it to be ready (for example, by waiting for the capture service to be available), and then start the second node. This avoids any race conditions where both nodes may try to connect to the same camera at the same time.

How to run the unit and module tests

This project comes with a set of unit and module tests to verify the provided functionality. To run the tests locally, first download and install the required data used for testing:

for sample in "FileCameraZivid2M70.zip" "BinWithCalibrationBoard.zip"; do
    echo "Downloading ${sample}"
    wget -q "https://www.zivid.com/software/${sample}" || exit $?
    mkdir -p /usr/share/Zivid/data/ || exit $?
    unzip "./${sample}" -d /usr/share/Zivid/data/ || exit $?
    rm "./${sample}" || exit $?
done

Then run the tests:

cd ~/ros2_ws/ && source install/setup.bash
colcon test --event-handlers console_direct+ && colcon test-result --all

The tests can also be run via docker. See the GitHub Actions configuration file for details.

How to enable debug logging

The node logs extra information at log level debug, including the settings used when capturing. Enable debug logging to troubleshoot issues:

ros2 run zivid_camera zivid_camera  --ros-args --log-level debug

Above will enable debug logging for all components, you can also specify just the zivid_camera logger like so:

ros2 run zivid_camera zivid_camera  --ros-args --log-level zivid_camera:=debug

How to compile the project with warnings enabled

colcon build --cmake-args -DCOMPILER_WARNINGS=ON

How to format the source code

The CI test for this package enforces the linting defined by clang-format. From the code analysis docker image, run:

find /host -name '*.cpp' -or -name '*.hpp' | xargs clang-format -i

The style follows the one from ament_clang_format.

License

This project is licensed under BSD 3-clause license, see the LICENSE file for details.

Support

Please report any issues or feature requests related to the ROS driver in the issue tracker. Visit Zivid Knowledge Base for general help on using Zivid 3D cameras. If you cannot find a solution to your issue, please contact customersuccess@zivid.com.

Acknowledgements

This FTP (Focused Technical Project) has received funding from the European Union’s Horizon 2020 research and innovation programme under the project ROSIN with the grant agreement No 732287. For more information, visit rosin-project.eu.

CONTRIBUTING

Any contribution that you make to this repository will be under the 3-Clause BSD License, as dictated by that license.


Repository Summary

Checkout URI https://github.com/zivid/zivid-ros.git
VCS Type git
VCS Version master
Last Updated 2025-04-03
Dev Status MAINTAINED
CI status No Continuous Integration
Released UNRELEASED
Tags No category tags.
Contributing Help Wanted (0)
Good First Issues (0)
Pull Requests to Review (0)

Packages

Name Version
zivid_camera 3.1.0
zivid_interfaces 3.1.0
zivid_rviz_plugin 3.0.0
zivid_samples 3.1.0

README

Zivid ROS driver

This is the official ROS driver for Zivid 3D cameras.

This driver provides support for ROS2.

If you are looking for the Zivid ROS1 driver, please use the ros1-master branch if you are using Zivid SDK 2.13 or older, or use the ros1-sdk-2.14.0 branch if you are using Zivid SDK 2.14 or newer.

Build Status Zivid Image


Contents: Installation | Getting Started | Launching | Configuration | Services | Topics | Samples | RViz Plugin | FAQ


Installation

Support

This driver supports Ubuntu 20.04 / 22.04 / 24.04 with ROS2. Follow the official ROS installation instructions for your OS.

If you are looking for the Zivid ROS1 driver, please use the ros1-master branch for Zivid SDK 2.13 or older, or use the ros1-sdk-2.14.0 branch for Zivid SDK 2.14 or newer.

Zivid SDK

To use the ROS driver you need to download and install the “Zivid Core” package. Zivid SDK version 2.15.0 is supported. See releases for older ROS driver releases that supports older SDK versions.

Follow this guide to install “Zivid Core” for your version of Ubuntu. The “Zivid Studio” and “Zivid Tools” packages can be useful to test your system setup and camera, but are not required by the driver.

An OpenCL 1.2 compatible GPU with driver installed is required by the SDK. Follow this guide to install OpenCL drivers for your system.

C++ compiler

A C++17 compiler is required.

sudo apt-get install -y g++

Downloading and building Zivid ROS driver

First, source the setup.bash script for your ROS distribution in your terminal:

source /opt/ros/$ROS_DISTRO/setup.bash

Then create the workspace and src directory:

mkdir -p ~/ros2_ws/src

Clone the Zivid ROS project into the src directory:

cd ~/ros2_ws/src
git clone https://github.com/zivid/zivid-ros.git

Initialize rosdep:

cd ~/ros2_ws/src
sudo rosdep init
rosdep update

Install dependencies:

cd ~/ros2_ws/src
rosdep install -i --from-path ./ -y

Finally, build the driver:

cd ~/ros2_ws
colcon build --symlink-install

Getting started

Connect the Zivid camera to your PC. You can use the ZividListCameras command-line tool available in the “Zivid Tools” package to confirm that your system has been configured correctly, and that the camera is discovered by your PC. You can also open Zivid Studio and connect to the camera. Close Zivid Studio before continuing with the rest of this guide.

Run the sample_capture_cpp via the launch script to check that everything is working.

cd ~/ros2_ws && source install/setup.bash
ros2 launch zivid_samples sample_with_rviz.launch sample:=sample_capture_cpp

This will start the zivid_camera driver node, the sample_capture_cpp node, and rviz. The zivid_camera driver will connect to the first available Zivid camera, and then sample_capture_cpp will trigger captures repeatedly. The resulting point cloud and color image should be visible in rviz.

A more detailed description of the zivid_camera driver follows below.

For sample code, see the Samples section.

Launching the driver

To launch the driver, use ros2 run:

cd ~/ros2_ws && source install/setup.bash
ros2 run zivid_camera zivid_camera

The driver will by default connect to the first available Zivid camera. This behavior can be overridden by setting the serial_number launch parameter, see below.

Launch Parameters (advanced)

The following parameters can be specified when starting the driver. Note that all the parameters are optional.

For example, to run the zivid_camera driver with a specific serial_number specified:

ros2 run zivid_camera zivid_camera --ros-args -p serial_number:=ABCD1234

Or you can use a launch file and invoke it as:

ros2 launch zivid_samples zivid_camera_with_serial_number.launch serial_number:=ABCD1234

file_camera_path (string, default: “”)

Specify the path to a file camera to use instead of a real Zivid camera. This can be used to develop without access to hardware. The file camera returns the same point cloud for every capture. Visit our knowledgebase to download file camera.

frame_id (string, default: “zivid_optical_frame”)

Specify the frame_id used for all published images and point clouds.

serial_number (string, default: “”)

Specify the serial number of the Zivid camera to use. This parameter is optional. By default, the driver will connect to the first available camera.

update_firmware_automatically (bool, default: true)

Specify if the firmware of the connected camera should be automatically updated to the correct version when the Zivid ROS driver starts. If set to false, if the firmware version is out of date then camera must be manually updated, for example using Zivid Studio or ZividFirmwareUpdater. Read more about firmware update in our knowledgebase. This parameter is optional, and by default it is true.

Configuration

The capture settings used by the zivid_camera ROS driver must be configured using YAML, which can be exported from Zivid Studio or the API, or downloaded as .yml files from our knowledge base.

For convenience, the Zivid ROS driver supports configuring capture settings in two ways: Using file path to a .yml file, or as a YAML string.

The following ROS parameters control which settings are used when capturing with the driver. Note that you must set either the _file_path or the _yaml parameter. If both _file_path and _yaml parameters are set to a non-empty string at the same time, then the driver will return an error when capturing. By default, all settings parameters are empty.

3D capture

settings_file_path (string, default: “”)

Specify the path to a .yml file that contains the settings you want to use.

settings_yaml (string, default: “”)

Specify a YAML string that contains the settings you want to use. For example, you can copy the contents of a .yml file saved from Zivid Studio.

The service capture_assistant/suggest_settings will modify the settings parameters automatically.

2D capture

settings_2d_file_path (string, default: “”)

Specify the path to a .yml file that contains the settings you want to use.

settings_2d_yaml (string, default: “”)

Specify a YAML string that contains the 2D settings you want to use. For example, you can copy the contents of a .yml file saved from Zivid Studio.

Services

capture

std_srvs/srv/Trigger

Invoke this service to trigger a 3D capture. See section Configuration for how to configure the 3D capture settings. The resulting point cloud is published on topics points/xyz and points/xyzrgba, color image is published on topic color/image_color, and depth image is published on topic depth/image. Camera calibration is published on topics color/camera_info and depth/camera_info.

See Sample Capture for code example.

capture_2d

std_srvs/srv/Trigger

Invoke this service to trigger a 2D capture. See section Configuration for how to configure the 2D capture settings. The resulting 2D image is published to topic color/image_color. Note: 2D RGB image is also published as a part of 3D capture, see capture.

See Sample Capture 2D for code example.

capture_and_save

zivid_interfaces/srv/CaptureAndSave.srv

It does exactly the same as the capture service, in addition it will save the frame to a file. This service takes a path as an argument. The chosen format is detected via the file extension. See knowledge base for a list of available output formats.

See Sample Capture and Save for code example.

capture_assistant/suggest_settings

zivid_interfaces/srv/CaptureAssistantSuggestSettings.srv

Invoke this service to analyze your scene and find suggested settings for your particular scene, camera distance, ambient lighting conditions, etc. This service will automatically update the node parameter settings_yaml with the suggested settings, see section Configuration. When this service has returned, you can invoke the capture service to trigger a 3D capture using these suggested settings.

This service has two parameters:

max_capture_time (duration):

Specify the maximum capture time for the settings suggested by the Capture Assistant. A longer capture time may be required to get good data for more challenging scenes. Minimum value is 0.2 sec and maximum value is 10.0 sec.

ambient_light_frequency (uint8):

Possible values are: AMBIENT_LIGHT_FREQUENCY_NONE, AMBIENT_LIGHT_FREQUENCY_50HZ, AMBIENT_LIGHT_FREQUENCY_60HZ. Can be used to ensure that the suggested settings are compatible with the frequency of the ambient light in the scene. If ambient light is unproblematic, use AMBIENT_LIGHT_FREQUENCY_NONE for optimal performance. Default is AMBIENT_LIGHT_FREQUENCY_NONE.

See Sample Capture Assistant for code example.

camera_info/model_name

zivid_interfaces/srv/CameraInfoModelName.srv

Returns the camera’s model name.

camera_info/serial_number

zivid_interfaces/srv/CameraInfoSerialNumber.srv

Returns the camera’s serial number.

is_connected

zivid_interfaces/srv/IsConnected.srv

Returns if the camera is currently in Connected state (from the perspective of the ROS driver). The connection status is updated by the driver every 10 seconds and before each capture service call. If the camera is not in Connected state the driver will attempt to re-connect to the camera when it detects that the camera is available. This can happen if the camera is power-cycled or the USB/Ethernet cable is unplugged and then replugged.

hand_eye_calibration/start

zivid_interfaces/srv/HandEyeCalibrationStart.srv

Prepares the node for hand-eye calibration and clears any previously collected hand-eye captures. The type of the calibration object to be used during capture must be provided. A working directory can optionally be provided so that all captures are saved to this directory. If provided, the directory must be given as an absolute path and the directory must be empty.

This service must be called first before capturing data for hand-eye calibration. It can also be used to restart an active hand-eye calibration session. After calling this service, proceed with captures by calling the hand_eye_calibration/capture service.

Please refer to the Zivid knowledge base for more information on hand-eye calibration.

hand_eye_calibration/capture

zivid_interfaces/srv/HandEyeCalibrationCapture.srv

Performs a hand-eye calibration capture. This service takes the robot pose as an input. Then it performs a capture, and detects any calibration objects. The resulting point cloud and color image will be published just like during a normal call to the capture service.

If the detections is successful, the result is stored locally in the driver. Additionally, if a working directory was specified during start, the captured frame and robot pose is saved to that directory.

This service uses the currently stored 3D capture settings to perform the capture. Ensure that the camera is properly configured first. Please see the knowledge base for how to get good quality data.

The camera and robot should be appropriately positioned so that the calibration object is visible in the frame. Multiple captures in different poses are necessary. After performing several captures, one can proceed by calling the hand_eye_calibration/calibrate service.

hand_eye_calibration/calibrate

zivid_interfaces/srv/HandEyeCalibrationCalibrate.srv

Computes the hand-eye calibration transform based on the captures gathered during the current hand-eye calibration session. Both eye-to-hand and eye-in-hand configurations are supported. If successful, the computed hand-eye transform is returned, see the knowledge base for more information on the hand-eye calibration solution.

The calibration procedure requires all robot poses to be different. At least 2 poses are required when using a calibration board, or 6 poses when using fiducial markers. For fiducial markers, each marker must be detected across 2 poses at minimum.

Low degrees-of-freedom (DOF) calibration is also supported (experimental) by supplying the fixed placement of calibration objects. This is not needed for regular (6-DOF) calibration.

hand_eye_calibration/load

zivid_interfaces/srv/HandEyeCalibrationLoad.srv

Loads a working directory from a previous hand-eye calibration session. See the hand_eye_calibration/start service for how to start a session with a working directory. The captures and robot poses are loaded from the provided directory. The directory is opened as read-only, and no new captures can be made during this session. However, calls to the hand_eye_calibration/calibrate service can be made to compute the hand-eye transform from the loaded data.

infield_correction/read

zivid_interfaces/srv/InfieldCorrectionRead.srv

Returns the state of the infield correction of the camera.

infield_correction/reset

std_srvs/srv/Trigger

Resets the infield correction on the camera to factory settings.

infield_correction/start

std_srvs/srv/Trigger

Prepares for infield correction, and clears any infield correction captures gathered so far in the zivid_camera node.

This service must be called before gathering captures using the infield_correction/capture service. However, other infield correction services can be used without calling the start service first. This service can also be used to restart an active infield correction session.

Please refer to the Zivid knowledge base for further information and guidelines for infield correction.

infield_correction/capture

zivid_interfaces/srv/InfieldCorrectionCapture.srv

Takes a capture to be used for infield correction. Please point the camera at a Zivid infield calibration object. It is recommended to cover several distances, with one or more captures at each distance. Successful captures are stored in the zivid_camera node.

Before calling this service, a call to infield_correction/start must have been made first.

This service call will perform a relatively slow but high-quality point cloud capture with the connected camera. Any settings applied to the camera in the ROS driver will be ignored for calls to this service, appropriate settings are automatically used. The resulting point cloud and color image will be published just like during a normal call to the capture service.

After sufficient number of captures, proceed by calling the infield_correction/compute service to compute the verification metrics and the correction based on the captures gathered so far. To additionally write the correction to camera, proceed by calling the infield_correction/compute_and_write service.

The captured data is cleared if the node is stopped, or after a successful call to either of the services infield_correction/start or infield_correction/compute_and_write.

infield_correction/compute

zivid_interfaces/srv/InfieldCorrectionCompute.srv

Calculates the new infield correction based the captured data gathered so far through the service infield_correction/capture.

The quantity and range of data is up to the user, but generally a larger dataset will yield a more accurate and reliable correction. If all measurements are taken at approximately the same distance, the resulting correction will mainly be valid at those distances. If several measurements are taken at significantly different distances, the resulting correction will likely be more suitable for extrapolation to distances beyond where the dataset is collected.

This service also acts as verification of the quality of the infield correction on a camera, or the need for one if none exists already. It returns an indication of the dimension trueness at the location where the input data was captured.

If the returned assessment indicates a trueness error that is above the threshold for your application, consider using infield_correction/compute_and_write. The service also returns information regarding the proposed working range and the accuracy that can be expected within the working range, if the correction is later written to the camera.

infield_correction/compute_and_write

zivid_interfaces/srv/InfieldCorrectionCompute.srv

Calculates the new infield correction based the captured data gathered so far through the service infield_correction/capture, and writes the result to the camera.

If the write operation is successful, the infield correction capture data is cleared. To perform infield correction again, a new session must be started with a call to the infield_correction/start service.

Please see the infield_correction/compute service for more information on the computed correction.

infield_correction/remove_last_capture

std_srvs/srv/Trigger

Removes the last infield correction capture gathered in the zivid_camera node.

Topics

The Zivid ROS driver provides several topics providing 3D, color, SNR and camera calibration data as a result of calling capture/capture_2d services. The different output topics provides flexibility for different use cases. Note that for performance reasons no messages are generated or sent on topics with zero active subscribers.

color/camera_info

sensor_msgs/msg/CameraInfo

Camera calibration and metadata.

color/image_color

sensor_msgs/msg/Image

Color/RGBA image. RGBA image is published as a result of invoking the capture or capture_2d service. Images are encoded as “rgba8”, where the alpha channel is always 255.

depth/camera_info

sensor_msgs/msg/CameraInfo

Camera calibration and metadata.

depth/image

sensor_msgs/msg/Image

Depth image. Each pixel contains the z-value (along the camera Z axis) in meters. The image is encoded as 32-bit float. Pixels where z-value is missing are NaN.

points/xyzrgba

sensor_msgs/msg/PointCloud2

Point cloud data. Sent as a result of the capture service. The output is in the camera’s optical frame, where x is right, y is down and z is forward. The included point fields are x, y, z (in meters) and rgba (color).

points/xyz

sensor_msgs/msg/PointCloud2

Point cloud data. This topic is similar to topic points/xyzrgba, except that only the XYZ 3D coordinates are included. This topic is recommended if you don’t need the RGBA values.

snr/camera_info

sensor_msgs/msg/CameraInfo

Camera calibration and metadata.

snr/image

sensor_msgs/msg/Image

Each pixel contains the SNR (signal-to-noise ratio) value. The image is encoded as 32-bit float. Published as a part of the capture service.

normals/xyz

sensor_msgs/msg/PointCloud2

Normals for the point cloud. The included fields are normal x, y and z coordinates. Each coordinate is a float value. There are no additional padding floats, so point-step is 12 bytes (3*4 bytes). The normals are unit vectors. Note that subscribing to this topic will cause some additional processing time for calculating the normals.

Samples

In the zivid_samples package we have added samples that demonstrate how to use the Zivid ROS driver. These samples can be used as a starting point for your project.

To launch the Python samples using ros2 launch, you need python to be available as a command. For example, the python-is-python3 package can be installed to achieve this, by running the following command:

sudo apt install python-is-python3

On Windows, the Python samples cannot be launched using ros2 launch. Instead, please launch the samples using ros2 run zivid_samples <sample_name>.py together with ros2 run zivid_camera zivid_camera in another terminal window.

Sample Capture

This sample performs single-acquisition 3D captures in a loop. This sample shows how to configure the capture settings, how to subscribe to the points/xyzrgba topic, and how to invoke the capture service.

Source code: C++ Python

ros2 launch zivid_samples sample.launch sample:=sample_capture_cpp
ros2 launch zivid_samples sample.launch sample:=sample_capture.py

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_capture_cpp
ros2 run zivid_samples sample_capture.py

Sample Capture 2D

This sample performs single-acquisition 2D captures in a loop. This sample shows how to configure the capture 2d settings with a YAML string, how to subscribe to the color/image_color topic, and how to invoke the capture_2d service.

Source code: C++ Python

ros2 launch zivid_samples sample.launch sample:=sample_capture_2d_cpp
ros2 launch zivid_samples sample.launch sample:=sample_capture_2d.py

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_capture_2d_cpp
ros2 run zivid_samples sample_capture_2d.py

Sample Capture and Save

This sample performs a capture, and stores the resulting frame to file. This sample shows how to configure the capture settings with a YAML string, how to invoke the capture_and_save service, and how to read the response from the service call.

Source code: C++ Python

ros2 launch zivid_samples sample.launch sample:=sample_capture_and_save_cpp
ros2 launch zivid_samples sample.launch sample:=sample_capture_and_save.py

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_capture_and_save_cpp
ros2 run zivid_samples sample_capture_and_save.py

Sample Capture Assistant

This sample shows how to invoke the capture_assistant/suggest_settings service to suggest and set capture settings. Then, it shows how to subscribe to the points/xyzrgba and color/image_color topics, and finally invoke the capture service.

Source code: C++ Python

ros2 launch zivid_samples sample.launch sample:=sample_capture_assistant_cpp
ros2 launch zivid_samples sample.launch sample:=sample_capture_assistant.py

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_capture_assistant_cpp
ros2 run zivid_samples sample_capture_assistant.py

Sample Capture with Settings from File

This sample performs single-acquisition 3D captures in a loop. This sample shows how to configure the capture settings from a yaml file, how to subscribe to the points/xyzrgba topic, and how to invoke the capture service.

Source code: C++ Python

ros2 launch zivid_samples sample.launch sample:=sample_capture_with_settings_from_file_cpp
ros2 launch zivid_samples sample.launch sample:=sample_capture_with_settings_from_file.py

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_capture_with_settings_from_file_cpp
ros2 run zivid_samples sample_capture_with_settings_from_file.py

Sample Infield Correction

This sample shows how to invoke the various infield_correction/[…] services to perform infield correction on Zivid cameras.

Source code: C++

ros2 launch zivid_samples sample.launch sample:=sample_infield_correction_cpp operation:=<operation>

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_infield_correction_cpp --ros-args -p operation:=<operation>

With the following argument:

operation (string, required)

Specify the infield correction operation to perform, one of:

Please see the infield correction documentation for prerequisites and guidelines on how to perform the correction.

For a more interactive experience, we recommend using the infield correction panel from the Zivid RViz plugin.

The typical procedure for performing a new infield correction is:

  1. start: Prepare for infield correction, clears any existing infield correction captures.
  2. capture: Take multiple captures from different angles and distances in accordance with the typical operating conditions of the camera.
  3. compute: Check the computed correction results and the estimated errors, verify that they give satisfactory results.
  4. compute_and_write: Compute the correction and write the results to camera.

The zivid_camera node persists the infield correction dataset as long as it is running. To start the infield correction captures over again, use the start operation which clears all infield correction captures previously gathered. The remove_last_capture can be used to remove just the last capture.

Sample Hand-Eye Calibration

This sample shows how to invoke various hand_eye_calibration/[…] services to perform hand-eye calibration on Zivid cameras. The sample is for exposition only, to demonstrate how the services can be called.

The sample begins by preparing the camera node for hand-eye calibration. Then a fixed number of captures is gathered at a fixed duration between captures, using a simulated robot pose. Finally, hand-eye calibration is run using the gathered data.

Source code: C++

ros2 launch zivid_samples sample.launch sample:=sample_hand_eye_calibration_cpp configuration:=<configuration> marker_ids:=<marker_ids> working_directory:=<working_directory>

Using ros2 run (when zivid_camera node is already running):

ros2 run zivid_samples sample_hand_eye_calibration_cpp --ros-args -p configuration:=<configuration> -p marker_ids:=<marker_ids> -p working_directory:=<working_directory>

With the following arguments:

configuration (string, required)

Specify the configuration for the hand-eye calibration, one of:

marker_ids (dynamic array of integers, default: empty):

Specifies a list of fiducial marker IDs used for detection (e.g. [1,2,3]), or empty if a Zivid calibration board is used instead.

working_directory (string, default: empty)

argument specifies the working directory used to store the gathered data, or empty to indicate that the data should not be stored on disk. A non-empty value must specify an absolute path to an empty directory.

For more information on performing the calibration, please see the Zivid hand-eye calibration documentation.

RViz Plugin

Infield Correction Panel

The Zivid RViz plugin provides a panel to interactively perform infield correction with a Zivid camera.

To see the panel in RViz, go to Panels -> Add New Panel, then select Zivid Infield Correction and click OK. Infield correction can now be performed by interacting with the newly added panel.

The panel is also visible when launching a Zivid sample with RViz, e.g.:

ros2 launch zivid_samples sample_with_rviz.launch sample:=sample_capture_with_settings_from_file_cpp

When the zivid_camera node is already running, RViz can also be started directly with the panel visible by using the Zivid RViz display configuration:

rviz2 -d $(ros2 pkg prefix zivid_camera)/share/zivid_camera/config/rviz/zivid.rviz

Please see the infield correction documentation for prerequisites and guidelines on how to perform the correction.

Frequently Asked Questions

How to visualize the output from the camera in rviz

ros2 launch zivid_camera visualize.launch

How to connect to specific Zivid serial number

ros2 run zivid_camera zivid_camera --ros-args -p serial_number:=ABCD1234

How to start the driver using settings.yml files

See section Configuration for more details.

ros2 run zivid_camera zivid_camera --ros-args -p settings_file_path:=/path/to/settings.yml -p settings_2d_file_path:=/path/to/settings2D.yml

How to trigger 3D/2D capture via terminal

ros2 service call /capture std_srvs/srv/Trigger
ros2 service call /capture_2d std_srvs/srv/Trigger

How to use a file camera

ros2 run zivid_camera zivid_camera --ros-args -p file_camera_path:=/usr/share/Zivid/data/FileCameraZivid2L100.zfc

Visit our knowledgebase to download file camera.

How to use multiple Zivid cameras

You can use multiple Zivid cameras simultaneously by starting one node per camera and specifying unique namespaces per node.

ros2 run zivid_camera zivid_camera --ros-args --remap __ns:=/zivid_camera1

ros2 run zivid_camera zivid_camera --ros-args --remap __ns:=/zivid_camera2

You can combine this with a serial_number parameter (see above) to control which node uses which camera. By default, the zivid_camera node will connect to the first available (unused) camera. We recommend that you first start the first node, then wait for it to be ready (for example, by waiting for the capture service to be available), and then start the second node. This avoids any race conditions where both nodes may try to connect to the same camera at the same time.

How to run the unit and module tests

This project comes with a set of unit and module tests to verify the provided functionality. To run the tests locally, first download and install the required data used for testing:

for sample in "FileCameraZivid2M70.zip" "BinWithCalibrationBoard.zip"; do
    echo "Downloading ${sample}"
    wget -q "https://www.zivid.com/software/${sample}" || exit $?
    mkdir -p /usr/share/Zivid/data/ || exit $?
    unzip "./${sample}" -d /usr/share/Zivid/data/ || exit $?
    rm "./${sample}" || exit $?
done

Then run the tests:

cd ~/ros2_ws/ && source install/setup.bash
colcon test --event-handlers console_direct+ && colcon test-result --all

The tests can also be run via docker. See the GitHub Actions configuration file for details.

How to enable debug logging

The node logs extra information at log level debug, including the settings used when capturing. Enable debug logging to troubleshoot issues:

ros2 run zivid_camera zivid_camera  --ros-args --log-level debug

Above will enable debug logging for all components, you can also specify just the zivid_camera logger like so:

ros2 run zivid_camera zivid_camera  --ros-args --log-level zivid_camera:=debug

How to compile the project with warnings enabled

colcon build --cmake-args -DCOMPILER_WARNINGS=ON

How to format the source code

The CI test for this package enforces the linting defined by clang-format. From the code analysis docker image, run:

find /host -name '*.cpp' -or -name '*.hpp' | xargs clang-format -i

The style follows the one from ament_clang_format.

License

This project is licensed under BSD 3-clause license, see the LICENSE file for details.

Support

Please report any issues or feature requests related to the ROS driver in the issue tracker. Visit Zivid Knowledge Base for general help on using Zivid 3D cameras. If you cannot find a solution to your issue, please contact customersuccess@zivid.com.

Acknowledgements

This FTP (Focused Technical Project) has received funding from the European Union’s Horizon 2020 research and innovation programme under the project ROSIN with the grant agreement No 732287. For more information, visit rosin-project.eu.

CONTRIBUTING

Any contribution that you make to this repository will be under the 3-Clause BSD License, as dictated by that license.