Repository Summary
Checkout URI | https://github.com/pal-robotics/aruco_ros.git |
VCS Type | git |
VCS Version | humble-devel |
Last Updated | 2024-05-09 |
Dev Status | DEVELOPED |
CI status | No Continuous Integration |
Released | RELEASED |
Tags | No category tags. |
Contributing |
Help Wanted (0)
Good First Issues (0) Pull Requests to Review (0) |
Packages
Name | Version |
---|---|
aruco | 5.0.5 |
aruco_msgs | 5.0.5 |
aruco_ros | 5.0.5 |
README
aruco_ros
Software package and ROS wrappers of the Aruco Augmented Reality marker detector library.
Features
-
High-framerate tracking of AR markers
-
Generate AR markers with given size and optimized for minimal perceptive ambiguity (when there are more markers to track)
-
Enhanced precision tracking by using boards of markers
-
ROS wrappers
Applications
- Object pose estimation
- Visual servoing: track object and hand at the same time
ROS API
Messages
-
aruco_ros/Marker.msg
Header header uint32 id geometry_msgs/PoseWithCovariance pose float64 confidence
-
aruco_ros/MarkerArray.msg
Header header aruco_ros/Marker[] markers
Kinetic changes
-
Updated the Aruco library to version 3.0.4
-
Changed the coordinate system to match the library’s, the convention is shown in the image below, following rviz conventions, X is red, Y is green and Z is blue.
Test it with REEM
- Open a REEM in simulation with a marker floating in front of the robot. This will start the stereo cameras of the robot too. Since this is only a vision test, there is nothing else in this world apart from the robot and a marker floating in front of it. An extra light source had to be added to compensate for the default darkness.
roslaunch reem_gazebo reem_gazebo.launch world:=floating_marker
- Launch the
image_proc
node to get undistorted images from the cameras of the robot.
ROS_NAMESPACE=/stereo/right rosrun image_proc image_proc image_raw:=image
- Start the
single
node which will start tracking the specified marker and will publish its pose in the camera frame
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right"
the frame in which the pose is refered to can be chosen with the 'ref_frame' argument. The next example forces the marker pose to
be published with respect to the robot base_link frame:
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right" ref_frame:=/base_link
- Visualize the result
rosrun image_view image_view image:=/aruco_single/result
CONTRIBUTING
Repository Summary
Checkout URI | https://github.com/pal-robotics/aruco_ros.git |
VCS Type | git |
VCS Version | humble-devel |
Last Updated | 2024-05-09 |
Dev Status | DEVELOPED |
CI status | No Continuous Integration |
Released | RELEASED |
Tags | No category tags. |
Contributing |
Help Wanted (0)
Good First Issues (0) Pull Requests to Review (0) |
Packages
Name | Version |
---|---|
aruco | 5.0.5 |
aruco_msgs | 5.0.5 |
aruco_ros | 5.0.5 |
README
aruco_ros
Software package and ROS wrappers of the Aruco Augmented Reality marker detector library.
Features
-
High-framerate tracking of AR markers
-
Generate AR markers with given size and optimized for minimal perceptive ambiguity (when there are more markers to track)
-
Enhanced precision tracking by using boards of markers
-
ROS wrappers
Applications
- Object pose estimation
- Visual servoing: track object and hand at the same time
ROS API
Messages
-
aruco_ros/Marker.msg
Header header uint32 id geometry_msgs/PoseWithCovariance pose float64 confidence
-
aruco_ros/MarkerArray.msg
Header header aruco_ros/Marker[] markers
Kinetic changes
-
Updated the Aruco library to version 3.0.4
-
Changed the coordinate system to match the library’s, the convention is shown in the image below, following rviz conventions, X is red, Y is green and Z is blue.
Test it with REEM
- Open a REEM in simulation with a marker floating in front of the robot. This will start the stereo cameras of the robot too. Since this is only a vision test, there is nothing else in this world apart from the robot and a marker floating in front of it. An extra light source had to be added to compensate for the default darkness.
roslaunch reem_gazebo reem_gazebo.launch world:=floating_marker
- Launch the
image_proc
node to get undistorted images from the cameras of the robot.
ROS_NAMESPACE=/stereo/right rosrun image_proc image_proc image_raw:=image
- Start the
single
node which will start tracking the specified marker and will publish its pose in the camera frame
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right"
the frame in which the pose is refered to can be chosen with the 'ref_frame' argument. The next example forces the marker pose to
be published with respect to the robot base_link frame:
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right" ref_frame:=/base_link
- Visualize the result
rosrun image_view image_view image:=/aruco_single/result
CONTRIBUTING
Repository Summary
Checkout URI | https://github.com/pal-robotics/aruco_ros.git |
VCS Type | git |
VCS Version | humble-devel |
Last Updated | 2024-05-09 |
Dev Status | DEVELOPED |
CI status | No Continuous Integration |
Released | RELEASED |
Tags | No category tags. |
Contributing |
Help Wanted (0)
Good First Issues (0) Pull Requests to Review (0) |
Packages
Name | Version |
---|---|
aruco | 5.0.5 |
aruco_msgs | 5.0.5 |
aruco_ros | 5.0.5 |
README
aruco_ros
Software package and ROS wrappers of the Aruco Augmented Reality marker detector library.
Features
-
High-framerate tracking of AR markers
-
Generate AR markers with given size and optimized for minimal perceptive ambiguity (when there are more markers to track)
-
Enhanced precision tracking by using boards of markers
-
ROS wrappers
Applications
- Object pose estimation
- Visual servoing: track object and hand at the same time
ROS API
Messages
-
aruco_ros/Marker.msg
Header header uint32 id geometry_msgs/PoseWithCovariance pose float64 confidence
-
aruco_ros/MarkerArray.msg
Header header aruco_ros/Marker[] markers
Kinetic changes
-
Updated the Aruco library to version 3.0.4
-
Changed the coordinate system to match the library’s, the convention is shown in the image below, following rviz conventions, X is red, Y is green and Z is blue.
Test it with REEM
- Open a REEM in simulation with a marker floating in front of the robot. This will start the stereo cameras of the robot too. Since this is only a vision test, there is nothing else in this world apart from the robot and a marker floating in front of it. An extra light source had to be added to compensate for the default darkness.
roslaunch reem_gazebo reem_gazebo.launch world:=floating_marker
- Launch the
image_proc
node to get undistorted images from the cameras of the robot.
ROS_NAMESPACE=/stereo/right rosrun image_proc image_proc image_raw:=image
- Start the
single
node which will start tracking the specified marker and will publish its pose in the camera frame
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right"
the frame in which the pose is refered to can be chosen with the 'ref_frame' argument. The next example forces the marker pose to
be published with respect to the robot base_link frame:
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right" ref_frame:=/base_link
- Visualize the result
rosrun image_view image_view image:=/aruco_single/result
CONTRIBUTING
Repository Summary
Checkout URI | https://github.com/pal-robotics/aruco_ros.git |
VCS Type | git |
VCS Version | humble-devel |
Last Updated | 2024-05-09 |
Dev Status | DEVELOPED |
CI status | No Continuous Integration |
Released | RELEASED |
Tags | No category tags. |
Contributing |
Help Wanted (0)
Good First Issues (0) Pull Requests to Review (0) |
Packages
Name | Version |
---|---|
aruco | 5.0.5 |
aruco_msgs | 5.0.5 |
aruco_ros | 5.0.5 |
README
aruco_ros
Software package and ROS wrappers of the Aruco Augmented Reality marker detector library.
Features
-
High-framerate tracking of AR markers
-
Generate AR markers with given size and optimized for minimal perceptive ambiguity (when there are more markers to track)
-
Enhanced precision tracking by using boards of markers
-
ROS wrappers
Applications
- Object pose estimation
- Visual servoing: track object and hand at the same time
ROS API
Messages
-
aruco_ros/Marker.msg
Header header uint32 id geometry_msgs/PoseWithCovariance pose float64 confidence
-
aruco_ros/MarkerArray.msg
Header header aruco_ros/Marker[] markers
Kinetic changes
-
Updated the Aruco library to version 3.0.4
-
Changed the coordinate system to match the library’s, the convention is shown in the image below, following rviz conventions, X is red, Y is green and Z is blue.
Test it with REEM
- Open a REEM in simulation with a marker floating in front of the robot. This will start the stereo cameras of the robot too. Since this is only a vision test, there is nothing else in this world apart from the robot and a marker floating in front of it. An extra light source had to be added to compensate for the default darkness.
roslaunch reem_gazebo reem_gazebo.launch world:=floating_marker
- Launch the
image_proc
node to get undistorted images from the cameras of the robot.
ROS_NAMESPACE=/stereo/right rosrun image_proc image_proc image_raw:=image
- Start the
single
node which will start tracking the specified marker and will publish its pose in the camera frame
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right"
the frame in which the pose is refered to can be chosen with the 'ref_frame' argument. The next example forces the marker pose to
be published with respect to the robot base_link frame:
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right" ref_frame:=/base_link
- Visualize the result
rosrun image_view image_view image:=/aruco_single/result
CONTRIBUTING
Repository Summary
Checkout URI | https://github.com/pal-robotics/aruco_ros.git |
VCS Type | git |
VCS Version | noetic-devel |
Last Updated | 2023-09-21 |
Dev Status | DEVELOPED |
CI status | No Continuous Integration |
Released | RELEASED |
Tags | No category tags. |
Contributing |
Help Wanted (0)
Good First Issues (0) Pull Requests to Review (0) |
Packages
Name | Version |
---|---|
aruco | 3.1.4 |
aruco_msgs | 3.1.4 |
aruco_ros | 3.1.4 |
README
aruco_ros
Software package and ROS wrappers of the Aruco Augmented Reality marker detector library.
Features
-
High-framerate tracking of AR markers
-
Generate AR markers with given size and optimized for minimal perceptive ambiguity (when there are more markers to track)
-
Enhanced precision tracking by using boards of markers
-
ROS wrappers
Applications
- Object pose estimation
- Visual servoing: track object and hand at the same time
ROS API
Messages
-
aruco_ros/Marker.msg
Header header uint32 id geometry_msgs/PoseWithCovariance pose float64 confidence
-
aruco_ros/MarkerArray.msg
Header header aruco_ros/Marker[] markers
Kinetic changes
-
Updated the Aruco library to version 3.0.4
-
Changed the coordinate system to match the library’s, the convention is shown in the image below, following rviz conventions, X is red, Y is green and Z is blue.
Test it with REEM
- Open a REEM in simulation with a marker floating in front of the robot. This will start the stereo cameras of the robot too. Since this is only a vision test, there is nothing else in this world apart from the robot and a marker floating in front of it. An extra light source had to be added to compensate for the default darkness.
roslaunch reem_gazebo reem_gazebo.launch world:=floating_marker
- Launch the
image_proc
node to get undistorted images from the cameras of the robot.
ROS_NAMESPACE=/stereo/right rosrun image_proc image_proc image_raw:=image
- Start the
single
node which will start tracking the specified marker and will publish its pose in the camera frame
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right"
the frame in which the pose is refered to can be chosen with the 'ref_frame' argument. The next example forces the marker pose to
be published with respect to the robot base_link frame:
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right" ref_frame:=/base_link
- Visualize the result
rosrun image_view image_view image:=/aruco_single/result
CONTRIBUTING
Repository Summary
Checkout URI | https://github.com/pal-robotics/aruco_ros.git |
VCS Type | git |
VCS Version | indigo-devel |
Last Updated | 2018-02-06 |
Dev Status | DEVELOPED |
CI status | No Continuous Integration |
Released | RELEASED |
Tags | No category tags. |
Contributing |
Help Wanted (0)
Good First Issues (0) Pull Requests to Review (0) |
Packages
Name | Version |
---|---|
aruco | 0.2.2 |
aruco_msgs | 0.2.2 |
aruco_ros | 0.2.2 |
README
aruco_ros
Software package and ROS wrappers of the Aruco Augmented Reality marker detector library.
Features
-
High-framerate tracking of AR markers
-
Generate AR markers with given size and optimized for minimal perceptive ambiguity (when there are more markers to track)
-
Enhanced precision tracking by using boards of markers
-
ROS wrappers
Applications
- Object pose estimation
- Visual servoing: track object and hand at the same time
ROS API
Messages
-
aruco_ros/Marker.msg
Header header uint32 id geometry_msgs/PoseWithCovariance pose float64 confidence
-
aruco_ros/MarkerArray.msg
Header header aruco_ros/Marker[] markers
Test it with REEM
- Open a REEM in simulation with a marker floating in front of the robot. This will start the stereo cameras of the robot too. Since this is only a vision test, there is nothing else in this world apart from the robot and a marker floating in front of it. An extra light source had to be added to compensate for the default darkness.
roslaunch reem_gazebo reem_gazebo.launch world:=floating_marker
- Launch the
image_proc
node to get undistorted images from the cameras of the robot.
ROS_NAMESPACE=/stereo/right rosrun image_proc image_proc image_raw:=image
- Start the
single
node which will start tracking the specified marker and will publish its pose in the camera frame
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right"
the frame in which the pose is refered to can be chosen with the 'ref_frame' argument. The next example forces the marker pose to
be published with respect to the robot base_link frame:
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right" ref_frame:=/base_link
- Visualize the result
rosrun image_view image_view image:=/aruco_single/result
CONTRIBUTING
Repository Summary
Checkout URI | https://github.com/pal-robotics/aruco_ros.git |
VCS Type | git |
VCS Version | indigo-devel |
Last Updated | 2018-02-06 |
Dev Status | DEVELOPED |
CI status | No Continuous Integration |
Released | RELEASED |
Tags | No category tags. |
Contributing |
Help Wanted (0)
Good First Issues (0) Pull Requests to Review (0) |
Packages
Name | Version |
---|---|
aruco | 0.2.2 |
aruco_msgs | 0.2.2 |
aruco_ros | 0.2.2 |
README
aruco_ros
Software package and ROS wrappers of the Aruco Augmented Reality marker detector library.
Features
-
High-framerate tracking of AR markers
-
Generate AR markers with given size and optimized for minimal perceptive ambiguity (when there are more markers to track)
-
Enhanced precision tracking by using boards of markers
-
ROS wrappers
Applications
- Object pose estimation
- Visual servoing: track object and hand at the same time
ROS API
Messages
-
aruco_ros/Marker.msg
Header header uint32 id geometry_msgs/PoseWithCovariance pose float64 confidence
-
aruco_ros/MarkerArray.msg
Header header aruco_ros/Marker[] markers
Test it with REEM
- Open a REEM in simulation with a marker floating in front of the robot. This will start the stereo cameras of the robot too. Since this is only a vision test, there is nothing else in this world apart from the robot and a marker floating in front of it. An extra light source had to be added to compensate for the default darkness.
roslaunch reem_gazebo reem_gazebo.launch world:=floating_marker
- Launch the
image_proc
node to get undistorted images from the cameras of the robot.
ROS_NAMESPACE=/stereo/right rosrun image_proc image_proc image_raw:=image
- Start the
single
node which will start tracking the specified marker and will publish its pose in the camera frame
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right"
the frame in which the pose is refered to can be chosen with the 'ref_frame' argument. The next example forces the marker pose to
be published with respect to the robot base_link frame:
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right" ref_frame:=/base_link
- Visualize the result
rosrun image_view image_view image:=/aruco_single/result
CONTRIBUTING
Repository Summary
Checkout URI | https://github.com/pal-robotics/aruco_ros.git |
VCS Type | git |
VCS Version | master |
Last Updated | 2016-10-18 |
Dev Status | DEVELOPED |
CI status | No Continuous Integration |
Released | UNRELEASED |
Tags | No category tags. |
Contributing |
Help Wanted (0)
Good First Issues (0) Pull Requests to Review (0) |
Packages
Name | Version |
---|---|
aruco | 0.0.1 |
aruco_msgs | 0.0.1 |
aruco_ros | 0.0.1 |
README
aruco_ros
Software package and ROS wrappers of the Aruco Augmented Reality marker detector library.
Features
- High-framerate tracking of AR markers
- Generate AR markers with given size and optimized for minimal perceptive ambiguity (when there are more markers to track)
- Enhanced precision tracking by using boards of markers
- ROS wrappers
Applications
- Object pose estimation
- Visual servoing: track object and hand at the same time
ROS API
Messages
-
aruco_ros/Marker.msg
Header header uint32 id geometry_msgs/PoseWithCovariance pose float64 confidence
-
aruco_ros/MarkerArray.msg
Header header aruco_ros/Marker[] markers
Test it with REEM
- Open a REEM in simulation with a marker floating in front of the robot. This will start the stereo cameras of the robot too. Since this is only a vision test, there is nothing else in this world apart from the robot and a marker floating in front of it. An extra light source had to be added to compensate for the default darkness.
roslaunch reem_gazebo reem_gazebo.launch world:=floating_marker
- Launch the
image_proc
node to get undistorted images from the cameras of the robot.
ROS_NAMESPACE=/stereo/right rosrun image_proc image_proc image_raw:=image
- Start the
single
node which will start tracking the specified marker and will publish its pose in the camera frame
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right"
the frame in which the pose is refered to can be chosen with the 'ref_frame' argument. The next example forces the marker pose to
be published with respect to the robot base_link frame:
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right" ref_frame:=/base_link
- Visualize the result
rosrun image_view image_view image:=/aruco_single/result
CONTRIBUTING
Repository Summary
Checkout URI | https://github.com/pal-robotics/aruco_ros.git |
VCS Type | git |
VCS Version | kinetic-devel |
Last Updated | 2020-08-28 |
Dev Status | DEVELOPED |
CI status | No Continuous Integration |
Released | RELEASED |
Tags | No category tags. |
Contributing |
Help Wanted (0)
Good First Issues (0) Pull Requests to Review (0) |
Packages
Name | Version |
---|---|
aruco | 0.2.4 |
aruco_msgs | 0.2.4 |
aruco_ros | 0.2.4 |
README
aruco_ros
Software package and ROS wrappers of the Aruco Augmented Reality marker detector library.
Features
-
High-framerate tracking of AR markers
-
Generate AR markers with given size and optimized for minimal perceptive ambiguity (when there are more markers to track)
-
Enhanced precision tracking by using boards of markers
-
ROS wrappers
Applications
- Object pose estimation
- Visual servoing: track object and hand at the same time
Generate markers
rosrun aruco optimalmarkers
ROS API
Messages
-
aruco_ros/Marker.msg
Header header uint32 id geometry_msgs/PoseWithCovariance pose float64 confidence
-
aruco_ros/MarkerArray.msg
Header header aruco_ros/Marker[] markers
Test it with REEM
- Open a REEM in simulation with a marker floating in front of the robot. This will start the stereo cameras of the robot too. Since this is only a vision test, there is nothing else in this world apart from the robot and a marker floating in front of it. An extra light source had to be added to compensate for the default darkness.
roslaunch reem_gazebo reem_gazebo.launch world:=floating_marker
- Launch the
image_proc
node to get undistorted images from the cameras of the robot.
ROS_NAMESPACE=/stereo/right rosrun image_proc image_proc image_raw:=image
- Start the
single
node which will start tracking the specified marker and will publish its pose in the camera frame
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right"
the frame in which the pose is refered to can be chosen with the 'ref_frame' argument. The next example forces the marker pose to
be published with respect to the robot base_link frame:
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right" ref_frame:=/base_link
- Visualize the result
rosrun image_view image_view image:=/aruco_single/result
CONTRIBUTING
Repository Summary
Checkout URI | https://github.com/pal-robotics/aruco_ros.git |
VCS Type | git |
VCS Version | melodic-devel |
Last Updated | 2023-09-21 |
Dev Status | DEVELOPED |
CI status | No Continuous Integration |
Released | RELEASED |
Tags | No category tags. |
Contributing |
Help Wanted (0)
Good First Issues (0) Pull Requests to Review (0) |
Packages
Name | Version |
---|---|
aruco | 2.2.3 |
aruco_msgs | 2.2.3 |
aruco_ros | 2.2.3 |
README
aruco_ros
Software package and ROS wrappers of the Aruco Augmented Reality marker detector library.
Features
-
High-framerate tracking of AR markers
-
Generate AR markers with given size and optimized for minimal perceptive ambiguity (when there are more markers to track)
-
Enhanced precision tracking by using boards of markers
-
ROS wrappers
Applications
- Object pose estimation
- Visual servoing: track object and hand at the same time
ROS API
Messages
-
aruco_ros/Marker.msg
Header header uint32 id geometry_msgs/PoseWithCovariance pose float64 confidence
-
aruco_ros/MarkerArray.msg
Header header aruco_ros/Marker[] markers
Kinetic changes
-
Updated the Aruco library to version 3.0.4
-
Changed the coordinate system to match the library’s, the convention is shown in the image below, following rviz conventions, X is red, Y is green and Z is blue.
Test it with REEM
- Open a REEM in simulation with a marker floating in front of the robot. This will start the stereo cameras of the robot too. Since this is only a vision test, there is nothing else in this world apart from the robot and a marker floating in front of it. An extra light source had to be added to compensate for the default darkness.
roslaunch reem_gazebo reem_gazebo.launch world:=floating_marker
- Launch the
image_proc
node to get undistorted images from the cameras of the robot.
ROS_NAMESPACE=/stereo/right rosrun image_proc image_proc image_raw:=image
- Start the
single
node which will start tracking the specified marker and will publish its pose in the camera frame
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right"
the frame in which the pose is refered to can be chosen with the 'ref_frame' argument. The next example forces the marker pose to
be published with respect to the robot base_link frame:
roslaunch aruco_ros single.launch markerId:=26 markerSize:=0.08 eye:="right" ref_frame:=/base_link
- Visualize the result
rosrun image_view image_view image:=/aruco_single/result