Version 1.1.0
new: Organize the code and documents; Use Patch-NetVLAD instead of NetVLAD for loop detect; Fix bugs and standardize algorithm input and output.
GAC-Mapping (GACM for short) is a multi-session SLAM to solve the mapping and localization of ground-aerial heterogeneous robot clusters in urban scenes. In GACM, a Visual-LiDAR ego-motion estimation module that considers point, line and planar constraints can provide robust odometry information. Thumbnail images representing obstacle outlines are generated and descriptors are extracted using a neural network to help perform data association between separate runs. Map segments and the robot poses are organized together and are updated during a pose graph optimization procedure. We provide guidance and examples how to run GACM on sample data and GrAco dataset (website or github).
Viedo (Replay: X6)
Contributors
Jinhao He 何晋豪,Yuming Zhou 周榆明,Lixiang Huang 黄立祥,Yang Kong 孔阳,Yilin Zhu 朱奕霖 (version 1.1.0 author) ,Cheng Hui 成慧 from SYSU RAPID Lab
Related Papers
- Jinhao He, Yuming Zhou, Lixiang Huang, Yang Kong, Hui Cheng*, “Ground and Aerial Collaborative Mapping in Urban Environments,”IEEE Robotics and Automation Letters, pp. 95-102, Oct, 2020
If our work has helped you, please cite:
@article{DBLP:journals/ral/HeZHKC21,
author = {Jinhao He and
Yuming Zhou and
Lixiang Huang and
Yang Kong and
Hui Cheng},
title = {Ground and Aerial Collaborative Mapping in Urban Environments},
journal = {{IEEE} Robotics Autom. Lett.},
volume = {6},
number = {1},
pages = {95--102},
year = {2021}
}
- Prerequisites
- Build GACM on ROS
- Run GACM on datasets
- Run GACM on your own device
- Acknowledgements
- License
1.1 Ubuntu and ROS
Ubuntu 18.04. ROS Melodic. ROS Installation additional ROS pacakge
sudo apt install ros-$ROS_DISTRO-cv-bridge ros-$ROS_DISTRO-tf ros-$ROS_DISTRO-message-filters ros-$ROS_DISTRO-image-transport
1.2 Ceres Solver
Test Version are 1.14.0 and 2.0.0. Follow Ceres Installation, remember to make install. (Our testing environment: Ubuntu 18.04, ROS Melodic, OpenCV 3.2.0, Eigen 3.3.4, PCL 1.8.1)
1.3 Patch-NetVLAD
Please refer to Patch-NetVLAD installation process. Our test is in anaconda.
If you need to replace the software source of anaconda, please follow the instructions in this blog.
# using gpu
conda create -n patchnetvlad python=3.8 numpy pytorch-gpu torchvision natsort tqdm opencv pillow scikit-learn faiss matplotlib-base -c conda-forge
# using cpu
conda create -n patchnetvlad python=3.8 numpy pytorch-cpu torchvision natsort tqdm opencv pillow scikit-learn faiss matplotlib-base -c conda-forge
Install rospy, rospkg and pyyaml by pip.
conda activate patchnetvlad
pip install pyyaml rospy rospkg
Clone the repository and catkin_make: (If camera_model exists in your workspace, please remove it first)
cd ~/catkin_ws/src
git clone https://github.com/SYSU-RoboticsLab/GAC-Mapping.git
cd ../
catkin_make
We provide some sample data (including two ground sequences and one aerial sequence). In addition, we also provide configuration files for the GrAco Dataset. You can also download the it for testing.
-
Open a terminal, run
roscore
. -
Open a new terminal, run Patch-NetVLAD to extract feature.
conda activate patchnetvlad source ~/catkin_ws/devel/setup.bash cd ~/catkin_ws/src/GAC-Mapping/Patch-NetVLAD python feature_extract_ros.py # use only cpu python feature_extract_ros.py --nocuda
GACM uses the pretrained model provided by Patch-NetVLAD. The script will automatically download the corresponding model file, or you can download it manually. Please refer to the README file of Patch-NetVLAD for the download address.
After testing, when using a 512-dimensional descriptor, the average thumbnail processing time is 0.007 seconds under GPU (NVIDIA TITAN X), and 0.085 seconds under CPU (i7-10710U).
-
Open a new terminal.
source ~/catkin_ws/devel/setup.bash roslaunch gacm gacm.launch
After the algorithm runs, you can choose whether to load the data file from
gacm_output\data\testSavemap
or not (see Output file for details). If you choose not to load, then you can play rosbag directly.If you choose to load, you need to wait for all data to be loaded. After loading, you can choose whether to perform pose graph optimization. The optimization looks for new inter-robot loops based on the global coordinates of each submap. If you think the mapping effect is good enough, or there are enough inter-robot loops, you don’t need to optimize it. Then you can play rosbag to run the algorithm.
After rosbag finishes, wait a moment (about 10 seconds), the algorithm will save data files and ask you whether to perform pose graph optimization. However, if it's the end of the first sequence (robot id is 0), the pose graph optimization will not be performed by default. After optimization, the algorithm will save the odometry file. When all files are saved, you can run the next rosbag or exit directly.
Finally, we recommend playing ground rosbags first, then air ones.
-
Run the
feature_extract_ros.py
script andgacm.launch
file according to the operation process section. -
Play car1.bag, car2.bag and uav.bag file in sequence and and follow the operation process section.
-
After all rosbags are played, you can see the map shown in the picture below on rviz. The trajectories and point cloud maps of different robots have different colors.
-
Also, you can see that there are yellow lines connecting some parts of the trajectory, which means that loop closures are not only detected but relative poses are obtained.
Dataset website: https://sites.google.com/view/graco-dataset
(or download data from github: https://github.com/SYSU-RoboticsLab/GrAco)
Notice: If you plan to run ground-05
(duration over 400 seconds), you need to set the submap_length
parameter in the configuration file to 70 or more.
-
play order: Aerial-07 -> Aerial-05
Change
graco.launch
configration file order<arg name="config_path" default = "$(find gacm)/config/graco_uav.yaml" /> <arg name="config_path1" default = "$(find gacm)/config/graco_uav.yaml" />
run
feature_extract_ros.py
andgraco.launch
. -
play order: Ground-06 -> Aerial-07
Change
graco.launch
configration file order<arg name="config_path" default = "$(find gacm)/config/graco_ugv.yaml" /> <arg name="config_path1" default = "$(find gacm)/config/graco_uav.yaml" />
run
feature_extract_ros.py
andgraco.launch
.
GACM uses Visual-LiDAR odometry as a front end, which means your device should contain LiDAR and camera. In addition, the front-end LiDAR part of GACM will generate a depth image based on the LiDAR point cloud to assist the visual part to restore the depth of the feature. So GACM also has certain requirements for the synchronization and placement of sensors.
- Do your best to keep in sync: Both the LiDAR and the camera in our device use hardware triggers to ensure time stamp synchronization. If your device synchronization is not good, you can reduce the weight of the visual part by reducing the value of
visual_weight_max
(in the configuration file). - Face the same direction: Because the point cloud is used to help image features recover depth, the overlapping of the field of view between the camera and LiDAR should be as much as possible. Our LiDAR uses Velodyne VLP16 (it has a vertical field of view of only 30 degrees). So if you have a mechanical LiDAR with a larger field of view, the effect of depth recovery is definitely better. However, the current algorithm does not support solid LiDAR (Livox).
- Rotation is important: For the extrinsic parameter of LiDAR and camera, we found that the error on the rotation has a greater impact on the accuracy of the odometry than the translation. For calibration, you can try adding an imu. Then complete the calibration of LiDAR-Imu and Camera-Imu respectively, get their respective extrinsic parameters. Finally, the extrinsic parameter of LiDAR-Camera are indirectly obtained by calculation LiDAR-Imu and Camera-Imu.
-
Create your own device configuration file
You can modify the configuration file we provide to adapt to your device. The configuration file is under the
GAC-Mapping/gacm/config/
folder.If your camera z-axis is above the horizontal plane, you can refer to the
car.yaml
orgraco_ugv.yaml
file.If your camera z-axis is below the horizontal plane (such as the way the camera is mounted on the drone in the GrAco dataset), you can refer to the
uav.yaml
orgraco_uav.yaml
. -
Create your own roslaunch file
GACM is multi-session SLAM, which supports running up to 5 different sequences. As shown below, you should modify the config file path for each sequence. When playing rosbag, be sure to play in the order of the configuration file.
<arg name="config_path" default = "$(find gacm)/config/car.yaml" /> <arg name="config_path1" default = "$(find gacm)/config/car.yaml" /> <arg name="config_path2" default = "$(find gacm)/config/uav.yaml" /> <arg name="config_path3" default = "$(find gacm)/config/uav.yaml" /> <arg name="config_path4" default = "$(find gacm)/config/uav.yaml" />
Also, you can change the pointcloud and image topic name by remap.
<node pkg="gacm" type="featureTrackingNode" name="featureTrackingNode" output="screen"> ... <!-- change your topic name --> <remap from="/image_raw" to="/camera_left/image_raw" /> <remap from="/velodyne_points" to="/velodyne/points" /> </node>
GACM will save all the data stored in the current algorithm at the end of each sequence of data (rosbag play ends). This allows GACM to start from the end of the previous sequence without re-running all sequences. These data files will be stored in the gacm_output/data/testSavemap
folder in the user's home directory (such as /home/sysu-rapid
).
The data files includes loop closure information for each sequence, pose graph node and edge information, pcd files of corner points, planar points and thumbnail point clouds of submap, and point cloud maps before and after fusion.
When a sequence ends, GACM automatically saves the odometry file for each sequence in tum format (timestamp[s] x y z q_x q_y q_z q_w). But you should modify the save path of the file in the configuration file! Otherwise the file will be saved in the gacm_output/data/testSavemap
folder.
You can view the thumbnails generated by the algorithm in real time under the gacm_ouput/pictures/submap_img
folder.
- In order to reduce the memory usage when the code is running, each sequence can only generate up to 100 submap. The number of lidar frames contained in each submap can be set (
submap_length
parameter). Assume that the period of each frame of lidar point cloud is 0.1 second, parametersubmap_length
is 50. Each sequence can only run for a maximum of 500 seconds (0.1 x 50 x 100 = 500). However, after testing, the running time of a sequence should not exceed 400 seconds, otherwise it may affect the real-time performance of the algorithm. - The inter-robot loop detection of the algorithm is unstable. After testing, for the same submap, the descriptors generated by Patch-NetVLAD may be different, which will affect loop closure detection. In addition, the point cloud registration results of NDT+GICP will also be greatly affected by the size of submap point cloud (
submap_length
,surf_threshold
), matching parameters (ndt_resolution_match
), and the initial values. Therefore, even if the correct loop closure is detected, the correct relative pose may not be obtained.
We use ceres solver for non-linear optimization and a generic camera model, Patch-NetVLAD for loop dectect and skipmap_ros to manage submap. In addition, this algorithm refers to the A-LOAM, LeGO-LOAM and VINS-mono. We really appreciate these open source projects!
The source code is released under GPLv3 license.