- follow original instruction to install all prerequisite
- follow below instruction to run hesai bag
# download and build
git clone https://github.com/ori-drs/ImMesh_hesai.git
mv ImMesh_hesai ImMesh
catkin build
# launch
roslaunch ImMesh mapping_new_college_hesai.launch
# play bag
rosbag play xxx.bag
The codes of ImMesh (pronounced as "I-am-Mesh") are now available! You can following our guidelines in this page to setup, run and evaluate our work.
The source code of this package is released under GPLv2 license. We only allow it free for personal and academic usage. For commercial use, please contact me <ziv.lin.ljrATgmail.com> and Dr. Fu Zhang <fuzhangAThku.hk> to negotiate a different license.
I hope you will love and enjoy our work, ImMesh. If you feel like ImMesh has indeed helped in your current research or work, I would greatly appreciate it if you could give a star to this repository or cite our paper in your academic research paper :)
ImMesh (pronounced as "I-am-Mesh") is a novel LiDAR(-inertial) odometry and meshing framework, which takes advantage of input of LiDAR data, achieving the goal of simultaneous localization and meshing in real-time. ImMesh comprises four tightly-coupled modules: receiver, localization, meshing, and broadcaster. The localization module utilizes the prepossessed sensor data from the receiver, estimates the sensor pose online by registering LiDAR scans to maps, and dynamically grows the map. Then, our meshing module takes the registered LiDAR scan for incrementally reconstructing the triangle mesh on the fly. Finally, the real-time odometry, map, and mesh are published via our broadcaster.
Our research has been accepted for publication in IEEE Transactions on Robotics (T-RO) (avaible here).
Our paper can be accessed from T-RO or be downloaded here.
Our accompanying videos are now available on YouTube (click below images to open) and Bilibili1, 2, 3.
Following this ROS Installation to install ROS and its additional pacakge:
sudo apt-get install ros-XXX-cv-bridge ros-XXX-tf ros-XXX-message-filters ros-XXX-image-transport ros-XXX-image-transport*
NOTICE: remember to replace "XXX" on above command as your ROS distributions, for example, if your use ROS-kinetic, the command should be:
sudo apt-get install ros-kinetic-cv-bridge ros-kinetic-tf ros-kinetic-message-filters ros-kinetic-image-transport*
Follow this livox_ros_driver Installation.
sudo apt-get install -y libcgal-dev pcl-tools
sudo apt-get install -y libgl-dev libglm-dev libglfw3-dev libglew-dev libglw1-mesa-dev
sudo apt-get install -y libcgal-dev libxkbcommon-x11-dev
Clone this repository and catkin_make:
cd ~/catkin_ws/src
git clone https://github.com/hku-mars/ImMesh.git
cd ../
catkin_make
source ~/catkin_ws/devel/setup.bash
ImMesh is capable run with solid-state LiDAR (e.g., Livox-avia form (r3live_dataset)).
Our example rosbag files for evaluation can be download from our Google drive or Baidu-NetDisk [百度网盘] (code提取码: wwxw).
After you have downloaded our bag files, you can now run our example ^_^
roslaunch ImMesh mapping_avia.launch
rosbag play YOUR_DOWNLOADED.bag
ImMesh is also capable to mechanical spinning LiDAR (e.g., Velodyne LiDAR form (KITTI-dataset, NCLT-dataset, and NTU-VIRAL ).
We also provide some example rosbag files for evaluation (thanks for KITTI-dataset and NCLT-dataset) , which can be download from our OneDrive.
After you have preparation of rosbag files, you can now run our example with:
For KITTI-dataset:
roslaunch ImMesh mapping_velody64.launch
rosbag play YOUR_DOWNLOADED.bag
For NCLT-dataset:
roslaunch ImMesh mapping_nclt.launch
rosbag play YOUR_DOWNLOADED.bag
For NTU-VIRAL:
roslaunch ImMesh mapping_ntu_viral.launch
rosbag play YOUR_DOWNLOADED.bag
ImMesh can also perform the fast offline mesh reconstruction with given point cloud file (*.pcd). Modified the pointcloud file name in mapping_pointcloud.launch:
# In mapping_pointcloud.launch file:
<param name="pc_name" type="string" value="YOUR_POINTCLOUD.pcd" />
Then, perform the mesh reconstruction with:
roslaunch ImMesh mapping_pointcloud.launch
If everything is correct, you will get the result same as the screenshot shown in below:
We provide some functions/features/demonstrations in our GUI (implemented based on ImGUI ), for example, as listed as below (press "F1" key for help):
Click "enable" checkbox (under the treenode of "LiDAR pointcloud reinforcement") to toggle realtime real-time depth rasterization and LiDAR pointcloud reinforcement.
Tips: You can click the "Dynamic configuration" checkbox to online dynamic reconfigure the rasterization parameter.
ImMesh allow you to save the online reconstructed mesh you build at anytime you wanted.
Click the "Save Mesh to PLY file" button (on the Main windows panel) to export the mesh to PLY file (default save in directory: ${HOME}/ImMesh_output).
Due to the space limitation, we are unable to list out all our features on this page. We recommend you to explore our work through GUI interaction by yourself.
I hope you will love and enjoy our work, ImMesh. If you feel like ImMesh has indeed helped in your current research or work, I would greatly appreciate it if you could give a star to this repository or cite our paper in your academic research paper.:)
ImMesh is building upon the foundations of our previous SLAM works including R3LIVE, VoxelMap, FAST-LIO, R2LIVE, and ikd-Tree. These works are all available on our GitHub, as listed below:
- R3LIVE: A Robust, Real-time, RGB-colored, LiDAR-Inertial-Visual tightly-coupled state Estimation and mapping package
- VoxelMap: An efficient and probabilistic adaptive(coarse-to-fine) voxel mapping method for 3D LiDAR
- FAST-LIO: A computationally efficient and robust LiDAR-inertial odometry (LIO) package
- R2LIVE: a robust, real-time tightly-coupled multi-sensor fusion framework
- ikd-Tree: an incremental k-d tree designed for robotic applications
We know our packages might not totally stable in this stage, and we are keep working on improving the performance and reliability of our codes. So, if you have met any bug or problem, please feel free to draw an issue or contact me via email <ziv.lin.ljrATgmail.com> and I will respond ASAP.
If you have any other question about this work, please feel free to contact me via www.jiaronglin.com and Dr. Fu Zhang <fuzhangAThku.hk> via email.