Skip to content

Latest commit

 

History

History
70 lines (54 loc) · 5.31 KB

data_srcipt.md

File metadata and controls

70 lines (54 loc) · 5.31 KB

Dataset of EVI-SAM

To benefit the community, we release the dataset that we utilized for real-time event-based dense mapping. This dataset also serves as a basis for comparing our event-based dense mapping approach with data captured from a depth camera.

Event-based Handheld Device

Our handheld device includes a power supply unit, an onboard computer NUC (equipped with Intel i7-1260P, 32GB RAM, and Ubuntu 20.04 operation system), a DAVIS346 event camera, and a Intel® RealSense™ D455 RGB-D camera.
We release the all of our schematic files with STL files format, which can be imported and printed directly. Moreover, we also release the CAD source files (with suffix “*.SLDPRT and *.SLDASM”), which can be opened and edited with Solidworks.

image

Data Sequence

Please note that the depth maps provided by the depth camera only serve as references for qualitative comparison.

Sequence Name Collection Date Total Size Rosbag (Baidu Disk)
HKU_LG_office_1 2023-11 8.91 GB Rosbag
HKU_LG_office_2 2023-11 7.71 GB Rosbag
HKU_LG_office_3 2023-11 9.49 GB Rosbag
HKU_LG_office_4 2023-11 8.55 GB Rosbag
HKU_LG_office_5 2023-11 8.01 GB Rosbag
HKU_LG_factory_1 2023-11 6.13 GB Rosbag
HKU_LG_factory_2 2023-11 5.98 GB Rosbag
HKU_LG_factory_3 2023-11 6.33 GB Rosbag
HKU_Logo_wall_1 2023-11 *** GB Rosbag
HKU_Logo_wall_2 2023-11 *** GB Rosbag
HKU_Logo_wall_3 2023-11 *** GB Rosbag

Evaluation using EVI-SAM

The video demos of evaluating EVI-SAM using this dataset are available on Bilibili 1 2 3 . The global mapping performance of our EVI-SAM is also illustrated in the following figur, showing the surface mesh generated through TSDF-based map fusion. Our global event-based dense mapping exhibits excellent global consistency. The supplemental video also illustrates the incremental reconstruction of the event-based surface mesh from updated TSDF voxels. This process enables on-demand meshing for visualization, allowing flexibility in generating the mesh at any time. Our TSDF-based map fusion for global mapping is designed to generate surface meshes that enable humans to assess the 3D reconstructed environment more effectively. This capability supports high-level mission goals, such as collision-free motion planning. During evaluation, an onboard computer NUC is utilized to support real-time pose estimation and local event-based dense mapping. However, the NUC lacks sufficient computational power to support a real-time meshing process. Therefore, we utilize a personal computer (Intel i7-11800H, 32GB RAM) without GPU to output the global mesh of EVI-SAM for the global mapping evaluation.

video

Visualization of the estimated camera trajectory and global 3D reconstruction (surface mesh) of our EVI-SAM. Sequentially display from right to left includes the event-based dense point clouds with texture information and intensity images, at selected viewpoints. (click the image to open video demo)

Evaluation using Traditional Image-based Dense Mapping

  • We also use this dataset to evaluate some traditional image-based dense mapping methods using monocular, stereo, and RGB-D setup.
  • A project of monocular dense mapping using RGB camera with VINS-MONO as pose estimation can be seen in Github Repository.

Evaluation using NeRF-based SLAM