Abstract
1. Introduction
1.1. Motivation
Real-world robotic datasets are vital for algorithm development. They provide diverse environments and challenging real-world sequences for training and evaluation of systems. They reduce costs and workforce requirements, such as system integration, calibration, and field operations (Nguyen et al., 2022), promoting broader participation in robotic research and fostering novel algorithm development. As robotics research transitions from traditional handcrafted methods to data-driven and hybrid approaches (Brohan et al., 2022; Shah et al., 2023b), the importance of these datasets continues to grow. Building upon this trend, this paper contributes to the exploration of SLAM dataset’s potential by introducing a diverse multi-sensor dataset. Our goal aims to develop a dataset to address the generalization challenges in robotic perception and navigation across various environments and operational conditions.
Comparing SLAM Datasets: Highlights differences in platforms, sensors, and ground truth methods, focusing on FusionPortableV2’s attributes. It classifies environment scales into small
and
indicate whether dataset satisfies the option or not.
indicates that the dataset misses robot kinematic data.
Despite the advancements in SLAM, a gap persists between the diversity of real-world scenarios and available datasets, especially concerning variations in environments, sensor modalities, and platforms executing diverse motions. This disparity affects the further development and evaluation of SLAM algorithms, potentially limiting their generalization capability and robustness in diverse real-world environments. Drawing on the recent success in the generalized manipulation and navigation models such as RT-2 (Brohan et al., 2023) and GNM (Shah et al., 2023a), it is our conviction that datasets featuring high motion and environmental diversity are crucial for the development of a versatile and generalized SLAM system.
1.2. Contributions
This paper aims to provide diverse, high-quality data using a consistent hardware setup, facilitating the development and benchmarking of robust and generalized SLAM systems. To achieve this objective, we present a comprehensive multi-sensor dataset, along with a detailed description of our data collection methodologies and complete benchmarking tools for evaluation. Building upon our previous FusionPortable dataset (Jiao et al., 2022), we introduce FusionPortableV2, a significant upgrade that expands the dataset in terms of data modalities, scenarios, and motion capabilities. This paper presents two main contributions: 1. 2.
During the platform development and data collection process, we addressed numerous technical challenges and meticulously documented the issues encountered and their solutions. This guidance should be a valuable resource for future researchers in the field. To promote collaborative advancements, we have publicly released all data and implementation details. We believe this will open up a wide range of research opportunities in field robotics and aid in the development of versatile, resilient robotic systems.
1.3. Organization
The remainder of this paper is structured in the following manner: Section 2 discusses related works most of on SLAM datasets and summarizes key contributions of this paper. Section 3 outlines the hardware setup and sensor details. Section 4 covers sensor calibration procedures. Section 5 describes the dataset, including platform characteristics and scenarios. Section 6 introduces details post-processing steps on raw sensor measurements and GT data. Section 7 presents the methodologies used for evaluating localization, mapping, and monocular depth estimation. Known issues of this dataset are also discussed. Finally, Section 8 concludes the paper and suggests directions for future research.
2. Related works
In the last decade, the availability of high-quality datasets has significantly accelerated the development of SOTA SLAM algorithms by reducing the time and cost associated with data acquisition and algorithm evaluation. The rapid progress in sensor and robotics technology has led to the widespread adoption of multiple sensors across various robotic platforms. This evolution has set new benchmarks and hastened the enhancement of SOTA algorithms, spanning both handcrafted and data-driven methods such as VINS-Mono (Qin et al., 2018), FAST-LIO2 (Xu et al., 2022), VILENS (Wisth et al., 2022), DROID-SLAM (Teed and Deng, 2021), and Gaussian Splatting SLAM (Matsuki et al., 2024).
Recent advancements in the SLAM field have also extended to areas such as place recognition and collaborative SLAM. Although these areas are not the primary focus of our work, they contribute significantly to the broader SLAM research community. Datasets such as Wild-Places (Knights et al., 2023) and HeLiPR (Jung et al., 2023) are specifically designed for LiDAR-based place recognition, emphasizing large-scale, long-term localization under varying appearance conditions (Yin et al., 2024). Collaborative SLAM, which enables information sharing and cooperative mapping among multiple robots, has gained attention with datasets such as Kimera-Multi (Tian et al., 2023), GrAco (Zhu et al., 2023), and S3E (Feng et al., 2022). These datasets provide multi-sensor data (e.g., cameras and LiDARs) and benchmarking tools, which are useful for both collaborative and single-robot SLAM evaluation. However, as they focus primarily on multi-robot scenarios with identical platforms, their utility for exploring cross-platform variability and single-robot SLAM challenges is limited.
In contrast, our FusionPortableV2 dataset is primarily designed for short-term odometry and SLAM, with a focus on accurately tracking the robot’s pose and constructing consistent maps within a single session or over a shorter time period. The dataset features diverse platforms, sensor configurations, and environments, making it better suited for studying SLAM generalization across different conditions compared to most related works, as detailed in Table 1.
2.1. Specific-platform datasets
Early SLAM datasets predominantly focused on visual-inertial fusion, targeting specific platforms and environments. This focus was largely due to the ubiquity and convenience of visual and inertial sensors which are cheap and lightweight. They cover sequences which were captured by different platforms ranging from handheld devices (Pfrommer et al., 2017; Schubert et al., 2018; Zuñiga-Noël et al., 2020), drones (Burri et al., 2016; Delmerico et al., 2019; Li et al., 2024; Majdik et al., 2017), unmanned ground vehicles (Pire et al., 2019), and aquatic vehicles (Miller et al., 2018), respectively. Notably, the UZH-FPV dataset (Delmerico et al., 2019) stands out for its integration of event cameras and the inclusion of rapid trajectories from aggressive drone flights.
Concurrently, in the automotive industry, urban environment datasets introduce specific challenges including adverse lighting, weather conditions, and larger scales. Long-range sensors such as LiDARs and Radars are preferred for their capabilities, even though they were initially bulky and costly. The KITTI dataset (Geiger et al., 2013) sets a benchmark in autonomous driving with its rich urban sensor data collection. Further developments in driving-related datasets have expanded across dimensions of duration (Maddern et al., 2017), urban complexity (Jeong et al., 2019), and weather adversity (Agarwal et al., 2020). The DSEC dataset (Gehrig et al., 2021), akin to UZH-FPV, leverages stereo event cameras for extensive driving scenes. Moreover, Radars are essential for outdoor perception, offering advantages in range, velocity measurement via the Doppler effect and weather resilience. Related datasets such as Boreas (Burnett et al., 2023), Oxford Radar RoboCar (Barnes et al., 2020), and OORD (Gadd et al., 2024) collected data under conditions such as fog, rain, and snow.
The trend toward multi-sensor fusion spurred the creation of diverse and complex datasets. Datasets such as NCLT (Carlevaris-Bianco et al., 2016), M2DGR (Yin et al., 2021), NTU-VIRAL (Nguyen et al., 2022), ALITA (Yin et al., 2022), and FusionPortable (Jiao et al., 2022) also pose challenges for SLAM, given their diverse environmental appearance and structure. These datasets feature a variety of environments, including dense vegetation, open spaces, and complex buildings with multiple levels and detailed layouts. The changing lighting, seasonal foliage variations, and movement of pedestrians and vehicles add complexity to campus environments. As highlighted in NCLT dataset, these factors are crucial for life-long SLAM challenges. However, these datasets, collected via specific platforms such as unmanned ground vehicles (UGVs) and drones, fall short in showcasing diverse motion patterns, especially aggressive maneuvers.
Recent advancements in sensor technology have enabled the development of portable multi-sensor suites capable of collecting high-quality, multi-modal data in diverse environments (e.g., multi-floor buildings). Notable examples include the Newer College (Ramezani et al., 2020) and Hilti-Oxford (Zhang et al., 2022) datasets. These datasets also offer dense, high-quality 3D global maps, enabling the generation of high-rate 6-DoF reference poses and the evaluation of mapping algorithms. The RELLIS-3D dataset (Jiang et al., 2021), though it primarily focuses on semantic scene understanding in off-road environments, also provides diverse sensor data that can be valuable for SLAM research.
2.2. Cross-platform datasets
As SLAM research progressed from single-platform or collaborative applications, there has been a growing interest in cross-platform generalization. Building upon the specific-platform datasets discussed in Section 2.1, several datasets explore the generalization of algorithms across different platforms and scales, aiming to integrate motion characteristics from varied platforms with minimal parameter tuning for diverse scenarios. The MVSEC dataset (Zhu et al., 2018) collected multi-sensor data with diverse platforms, excluding UGV sequences. Conversely, the Nebula dataset (Reinke et al., 2022), developed during the DARPA Subterranean Challenge, includes field environments with both wheeled and legged robots, providing precise maps and trajectories. However, it lacks urban data and primarily focuses on LiDAR-based perception. The M3ED dataset (Chaney et al., 2023), although closely aligned with our objectives, lacks indoor data and platform-specific kinematic measurements, underscoring the unique contribution of our dataset.
The sensors used in this dataset and their corresponding specifications. The detailed definition of ROS message type and naming of coordinate frames of each sensor are provided in the dataset website.
3. System overview
This section presents our developed multi-sensor suite, designed for integration with various mobile platforms through plug-and-play functionality. All sensors are securely mounted on an aluminum alloy frame, facilitating a unified installation. Additionally, we detail the devices employed for collecting GT trajectories and maps.
3.1. Suite setup and synchronization
The Multi-Sensor Suite (MSS) integrates exteroceptive and proprioceptive sensors, including a 3D Ouster LiDAR, stereo frame and event cameras, and IMUs, as depicted in its CAD model in Figure 1. We use two PCs that are synchronized via a Network Time Protocol (NTP) server for data collection. The primary PC processes data from the frame cameras and IMU, while the auxiliary PC handles additional data types. Both PCs are equipped with a 1 TB SSD, 64 GB of DDR4 memory, and an Intel i7 processor, running Ubuntu with a real-time kernel patch and employing the Robot Operating System (ROS) for data collection. This distributed architecture reduces the number of ROS nodes running on each individual PC, thereby alleviating the risk of queuing problems during data collection. After the mission, separate data collected by these two PCs are merged and post-processed offline. The subsequent sections will elaborate on the synchronization approach and the features of each sensor. CAD model of the sensor rig where axes are marked: red: 
3.1.1. Synchronization
The synchronization process is illustrated in Figure 2. Generally, the field-programmable gate array (FPGA) board synchronizes with the pulse-per-second (PPS) signal from the external GNSS receiver, producing higher frequency trigger signals for the IMU, stereo frame cameras, and LiDAR clock alignment. In GPS-denied environments, it utilizes its internal low drift oscillator for synchronization, achieving a time accuracy below 1 ms between multiple trigger signals. To synchronize LiDAR and camera data, we phase-lock
1
the LiDAR’s rotation so its forward-facing direction aligns with the camera’s capture timing, accounting for the continuous nature of LiDAR’s spinning data. We use the internal (Master–Slave mode) mechanism to synchronize stereo event cameras, where the left event camera is assigned as the master to send trigger signals to the right camera. (a) Illustration of data collection which shows the data flow and synchronization processes. The red arrow indicate PPS signals for synchronization, green arrows show UTC time synchronization, and blue arrows represent sensor triggering signals, and black arrows depict the flow of raw data. (b) The timing diagram for triggerable (our case) and non-triggerable sensors, illustrating the unknown time offset caused by the delay in starting data capture, the duration of data capture, and the time required for data transmission from the sensor to the PC. Our synchronization solution can reduce the time delay (i.e., 
Figure 2(b) illustrates our synchronization scheme. The FPGA sends trigger signals to connected sensors, which starts capturing data after a small delay (commonly
3.1.2. 3D LiDAR
Our LiDAR choose the OS1-128 Gen5 LiDAR that operates at 10 Hz. It features a built-in IMU capturing gyroscope, acceleration, and magnetometer data and generates four types of images to facilitate the usage of image-based algorithms:
3.1.3. Stereo frame cameras
Our setup includes two FLIR BFS-U3-31S4C global-shutter color cameras for stereo imaging, synchronizing to output images at 1024 × 768 pixels and 20 Hz. The exposure time
3.1.4. Stereo event cameras
Two event cameras, which are known for their high temporal resolution, extensive dynamic range, and energy efficiency, are used for data collection. These cameras output events data, 346 × 260 frame images, and high-rate IMU measurements. Frame images cannot be synchronized, resulting in a 10–20 ms delay. Infrared filters are used to lessen LiDAR light interference. Exposure times are set fixedly, whereas outdoor settings use auto-exposure to maintain image quality under varying light conditions.
3.1.5. Inertial measurement unit
The STIM300 IMU, a tactical-grade sensor 2 , serves as the primary inertial sensor of our system, mounted beneath the LiDAR. It has a bias instability of 0.3°/h for the gyroscope and 0.04 mg for the accelerometer. The sensor outputs angular velocity and acceleration measurements at 200 Hz. Other components, including the LiDAR, event cameras, and the 3DM-GQ7 Inertial Navigation System (INS), are also integrated with IMUs. Further details are provided in subsequent sections.
3.2. Platform-specific sensor setup
Our goal is to create a diverse dataset by capturing sequences with multiple mobile platforms, thereby increasing the dataset’s complexity and challenge compared to those relying on a single platform. Each platform is equipped with a handheld multi-sensor suite and platform-specific sensors, as shown in Figure 3. Figure 4 displays the platforms and exemplifies typical scenes from which data were gathered. Platform-specific sensor settings are introduced in the subsequent sections, while the description of their motion and scenario patterns are presented in Section 5.1. Layouts of the platform-specific sensor setup, including different coordinate systems and their relative translation. More detailed and accurate dimensional data are provided in our calibration files. Platform-specific data samples: 

3.2.1. Legged robot
We have selected the Unitree A1 quadruped robot as our legged platform, as shown in Figure 3(c). This robot is equipped with 12 joint motor encoders and four contact sensors per leg, located at the hip, thigh, calf, and foot. These sensors provide kinematic measurements at a rate of 50 Hz. The MSS is affixed to the robot’s dorsal side and communicates with the kinematic sensors via Ethernet. In addition to the raw sensor measurements, we record metadata for each motor, which includes torque, velocity, position, and temperature, along with kinematic-inertial odometry data.
3.2.2. Unmanned ground vehicle
The MSS is integrated into a four-wheeled Ackerman UGV (see Figure 3(a), 3(b)), originally designed for logistics transportation (Liu et al., 2021). To optimize signal reception, the dual GNSS antennas of the INS are positioned at UGV’s rear side. Kinematic data for the UGV is acquired through two incremental rotary encoders, strategically positioned at the center of the rear wheel. These encoders, featuring 1000 pulses per revolution, produce measurement data at a rate of approximately 100 Hz, which is then recorded.
3.2.3. Vehicle
As depicted in Figure 3(d), we follow the KITTI setup (Geiger et al., 2013) by extending the baseline of both the stereo cameras and the dual antenna, with the stereo frame camera having a baseline of 83 cm and the event camera having a baseline of 73 cm. This extended baseline enhances the accuracy of depth estimation for distant objects, as compared with that in the UGV. The MSS is mounted on the vehicle’s luggage rack using a custom-designed aluminum frame.
3.3. Ground truth provision setup
High-precision, dense RGB point cloud maps and GT trajectories are essential for evaluating SLAM and perception algorithms. This section describes three types of GT devices featured in our dataset, selected to meet the varied needs of the sequences. Through the integration of data from these GT devices, our dataset provides comprehensive support for algorithm benchmarking, not only in localization and mapping but also across diverse applications and requirements.
3.3.1. Dense RGB point cloud map
For creating dense point cloud maps of outdoor scenarios, the Leica RTC360 laser scanner was selected, because of its high scanning rate of up to 2 million points per second and accuracy under 5.3 mm within a 40 m radius. Some indoor areas were scanned with the Leica BLK360, which operates at a rate of 0.68 million points per second and achieves an accuracy of 4 mm within a 10 m range. All scans are registered and merged by the Leica Cyclone software 3 , resulting in a dense and precise RGB point cloud map. This map with the resolution as 8 cm, covering all data collection areas, can be used to evaluate the mapping results of algorithms 4 ranging from model-based (Lin and Zhang, 2022) and learning-based methods (Pan et al., 2024).
3.3.2. 3-DoF GT trajectory
For indoor and small-scale outdoor environments, the Leica MS60 total station 5 was utilized to measure the GT trajectory of the robot at the 3-DoF position. As shown in Figure 4(a), the tracking prism is placed atop the LiDAR. The GT trajectory was captured at a frequency between 5 and 8 Hz, achieving an accuracy of 1 mm. However, due to occasional instability in the measurement rate, the GT trajectory is resampled at 20 Hz using cubic spline interpolation for a more consistent evaluation. For further details on this process, please refer to Section 6.2.1.
3.3.3. 6-DoF GT trajectory
While the stationary Leica MS60 provides accurate measurements, it cannot track the prism when it is occluded or outside the visible range. Consequently, we employ the INS to capture 6-DoF GT trajectories in large-scale and outdoor environments with available GNSS satellites. This sensor integrates data from its internal dual-antenna RTK-GNSS, which provides raw data at a frequency of 2 Hz, and an IMU, to deliver estimated poses with an output rate of up to 30 Hz. Before commencing data collection, we ensure that the GNSS has initialized with a sufficient satellite lock and the RTK is in a fixed status, typically achieving a positioning accuracy of up to 1.4 cm. This initialization process usually takes 1–3 min in outdoor. To maintain the reliability of our ground truth data, we adhere to strict criteria. First, the INS filter must be in a stable navigation status. Second, both dual-antenna GNSS must be in RTK-fixed status, receiving signals from at least 20 satellites. Lastly, the positional error covariance must have converged to an optimal state.
4. Sensor calibration
Description of intrinsic and extrinsic parameter calibration.
detailed in Section 4.2 of the paper.
4.1. Intrinsic calibration
We calibrate IMUs and cameras using the off-the-shelf Kalibr toolbox (Furgale et al., 2013; Rehder et al., 2016a). For wheel encoder intrinsics, such as wheel radius and axle track, we implement the motion-based calibration algorithm outlined in (Jeong et al., 2019). This involves manually maneuver the UGV through significant transformations, as depicted in Figure 5. We calculate the UGV’s planar motion for each interval Comparison of trajectories: estimated motion by the INS (3DM-GQ7) (red), integration of encoders’ measurements before calibration (green), and after calibration (blue) using the sequence 
4.2. Extrinsic calibration
Extrinsic calibrations, encompassing 6-DoF transformations and time offsets for IMU–IMU, IMU–camera, IMU-prism, camera–camera, and camera–LiDAR pairs, are typically obtained with off-the-shelf toolboxes. We specifically describe the calibration between the IMU and the prism that defines the reference frame of GT measurements relative to the total station (Leica MS60). We design the indirect calibration method since the total station provides only 3-DoF and low-rate trajectories. We observed that the prism is visible to infrared cameras in the motion capture room. We place and adjust three tracking markers around the prism to approximate its center, as shown in Figure 6. We move the handheld device to perform the “8”-shape trajectory. Both the motion capture system and LiDAR-inertial odometry (Xu et al., 2022) can estimate trajectories of the prism and STIM300 IMU ( Sensor placement for the IMU–Prism calibration. Reflective balls for motion capture cameras (MCC) and the prism are marked in red and blue, respectively. We use MCC’s measurements to infer high-rate motion of the prism.
5. Dataset description
Statistics and key challenges of each sequence are reported. Abbreviations: T: Total time. D: Total distance traveled. L: Large. M: Medium. S: Small.
5.1. Analysis of platforms characteristics
Each platform has its motion patterns (e.g., speed, angular velocity, dynamic frequency) and working ranges. Figure 7 visualize typical motion patterns of different platforms on some example sequences. We can clearly observe that the legged robot's motion is highly dynamic and the vehicle's motion is fast but smooth. Drawing from this observation, we meticulously design sequences to highlight the unique features of each platform. Motion analysis with four mobile platforms in terms of linear acceleration [m/s2], angular velocity [rad/s], and velocity [m/s]. We use measurements from STIM-300 to get linear acceleration (including gravity) and angular velocity as well as SLAM results to get rough velocity. (a)–(d): The left two columns of each row illustrate time-domain data, revealing immediate dynamic behaviors, while the right two columns display frequency-domain data, highlighting motion features of different platforms. Each line in the legend represents linear acceleration and angular velocity of each axis: 
5.1.1. Handheld
Since the handheld MSS is commonly held by a user, it offers flexibility for data collection scenarios. The handheld multi-sensor device provides adaptable data collection across diverse settings, akin to market counterparts like the Leica BLK2GO mobile scanning device, which excels in precision scanning and motion estimates. Therefore, we collect data in scenarios including a
5.1.2. Legged robot
The quadruped robot carries a sensor suite and commonly operates in
5.1.3. Unmanned ground vehicle
The UGV is typically designed for last-mile delivery and navigates middle-scale areas like campuses and factories. Constrained by Ackermann steering geometry, the UGV executes planar and smooth movements in response to the operator’s inputs. Data collection is conducted in various environments, including an
5.1.4. Vehicle
The vehicle collects data across diverse urban environments in Hong Kong, navigating through
5.2. Challenging factors
Prior to data collection, we acknowledge that practical factors contribute to sensor degradation and potential algorithmic failure. Our data sequences, integrated with the platforms described, aim to comprehensively evaluate algorithm performance in terms of accuracy, efficiency, and robustness. Additionally, we anticipate these sequences will draw the development of novel algorithms.
5.2.1. Illumination conditions
Different illumination conditions, such as bright sunlight, shadows, and low light, affect the quality of visual sensors and pose challenges for visual perception algorithms. For example, in bright sunlight, cameras are sometimes overexposed, resulting in a loss of appearance information. On the contrary, cameras are sometimes underexposed in low light conditions, leading to image noise and poor visibility.
5.2.2. Richness of texture and structure
Structured environments (e.g., offices or buildings) can mainly be explained using geometric primitives, while semi-structured environments have both geometric and complex elements like trees and sundries. Scenarios like narrow corridors are structured but may challenge state estimators. Additionally, texture-rich scenes facilitate visual algorithms to extract stable features (e.g., points and lines), while texture-less may negatively affect the performance. Also, in texture-less environments, only a small amount of events is triggered.
5.2.3. Dynamic objects
In dynamic environments, several elements (e.g., pedestrians or cars) are moving when the data are captured. This is in contrast to static environments. For instance, moving cars cause noisy reflections and occlusions to LiDAR data, while pedestrians cause motion blur to images. Overall, dynamic objects induce negative effects from several aspects such as incorrect data association, occlusion, and “ghost” points remaining on the map.
5.2.4. Intermittent GNSS
The intermittent GNSS signal issue typically arises in environments like places where dense and towering urban clusters are presented, overpasses, and indoor–outdoor transition areas. A special example is the city center of Hong Kong. In such scenarios, GNSS signals are often obstructed, leading to sporadic reception and significant uncertainty.
5.2.5. Scale variability
Developing SLAM and perception algorithms for large-scale environments may encounter challenges such as an increased computational load and a heightened risk of perceptual aliasing. The former necessitates stricter demands on algorithm latency and memory usage, whereas the latter requires more accurate long-term associations for place recognition (Yin et al., 2024), given the potential for environments to include geographically distant yet visually similar locations.
5.2.6. Viewpoint change
We consider the viewpoint change from two perspectives: “yaw-change” and “roll-pitch-change.” The former often occurs in sequences with loops. Several sequences in our dataset feature at least one loop, posing challenges for place recognition and image matching algorithms. For instance:
The latter aspect is particularly relevant to the cross-view localization (CVL) problem (Shi et al., 2023), which involves viewpoint changes between down-facing satellite images and ground-level images. Open-source satellite images can be obtained from tools such as Google Earth. Since several sequences include accurate GT geo-localization information and covers diverse scenarios (e.g., campus, urban roads, downhill mountain roads), it can serve as a challenging benchmark for CVL.
5.3. Sequence description
Table 4 summarizes the characteristics of our proposed sequences, detailing aspects such as temporal and spatial dimensions, motion patterns, locations, textural and structural richness, and whether GT poses and maps cover. Figures 8 and 9 illustrate the coverage areas of the sequences from a satellite view perspective. Trajectories of several sequences collected using the low-speed UGV in the campus, where environments with different structures and texture including room, escalator, grassland, parking lots are presented. Trajectories of several sequences collected using the high-speed vehicle in Hong Kong.

5.4. Dataset organization
Figure 10 outlines our dataset’s organization. Sensor data were captured using the ROS bag tool
12
. ROS bags are used due to their numerous advantages, such as mature tools for debugging, visualization of the tf tree, especially with multiple sensors present, a broadcasting mechanism for batching output messages of different types, and the ability to convert them into individual files. To facilitate download, ROS bags were compressed with 7-Zip. Each bag follows the naming convention The dataset organization.
5.5. Development tools
We release a set of tools that enable users to tailor our dataset to their specific application needs. Components are introduced as follows:
5.5.1. Software development kit (SDK)
We present a Python-only SDK that is both extensible and user-friendly. The kit includes foundational functions such as loading calibration parameters and visualizing them using a TF tree, parsing ROS messages into discrete files, data post-processing, and basic data manipulation. Figure 11 shows the point cloud projection function provided by the package. The projected point cloud onto the left frame image using our SDK shows points’ colors indicating relative distances. This involves a basic implementation, including a data loader, calibration loader, point cloud manipulation, and camera model.
5.5.2. Evaluation
We provide a set of scripts and tools for algorithm evaluation including localization and mapping.
5.5.3. Application
We provide open-source repositories for users to try different applications with our dataset covering localization, mapping, monocular depth estimation, and anonymization of specific objects. All can be found on the dataset website.
6. Data post-processing
The raw data captured by sensors and GT devices undergo post-processing before public release. The specifics are outlined as follows.
6.1. Privacy management
Data collection in public spaces such as the campus and urban roads was conducted with strict adherence to privacy regulations. We employed the anonymization technique 13 introduced in (Burnett et al., 2023) to obscure all human faces and license plates in images from our stereo frame cameras. Building upon the original implementation, we enhanced the algorithm’s efficiency using the ONNX Runtime Library 14 . This upgraded version is now ROS-compatible, offering a valuable resource to the community, and is included in our development tools.
6.2. GT data processing
Due to diverse sources of GT trajectories and maps, it is necessary to standardize the GT data through processing and conversion and then verify them. These steps should be executed sequentially.
6.2.1. 3-DoF GT poses of total station
The preprocessing initiates with temporal alignment, crucial for synchronizing Leica MS60 total station measurements with sensor data, following the approach proposed in (Nguyen et al., 2022). This synchronization finds the optimal time offset that minimizes the Absolute Trajectory Error (ATE) between the MS60’s recorded poses Alignment process for 
6.2.2. 6-DoF GT poses of INS
Each 6-DoF pose provided by the INS is accompanied by a variance value, indicating the measurement’s uncertainty. This uncertainty increases when the GNSS signal is obstructed. For a fair comparison, we manually removed data points with excessive uncertainty to serve as ground truth. We also provide the original data along with relevant tools for users to process the data according to their needs. For the sequences
6.2.3. Accuracy of GT maps
In the construction of our ground truth (GT) maps, we employed a high-precision Leica scanner to capture detailed environments both indoors and outdoors. Figure 13 displays the complete RGB point cloud map of the entire campus and a section of the underground parking garage. The fine details observable in the depicted areas highlight the high quality of both the point cloud and its color fidelity. According to the Leica Cyclone software report, the GT map exhibits an average error of less than 15 mm across all pairwise scans, the average error is 3.4 mm and with 90% of these scans maintaining an error margin of 10 mm or less. The precision of our GT map exceeds that of point cloud maps constructed via current LiDAR SLAM technologies by nearly two orders of magnitude, making it suitable for algorithm evaluation. GT RGB point cloud map of the (a) campus 
7. Experiment
We select eight representative sequences (two sequences from each platform) from the dataset to conduct a series algorithm evaluation and verification. Experiments include localization, mapping, and monocular depth estimation.
7.1. Evaluation of localization
7.1.1. Experiment setting
As one of the main applications, this dataset can be used to benchmark SOTA SLAM algorithms. Here in, for evaluation of localization systems with different input modalities, we select four SOTA SLAM algorithms (including a learning-based method): DROID-SLAM (left frame camera) (Teed and Deng, 2021), VINS-Fusion (LC) (IMU + stereo frame cameras, with loop closure enabled) (Qin et al., 2018), FAST-LIO2 (IMU+LiDAR) (Xu et al., 2022), and R3LIVE (IMU+LiDAR+left frame camera) (Lin and Zhang, 2022). The customized data loaders of each method are publicly released to foster research. This figure illustrates the comparative performance of leading SLAM algorithms operating across four distinct platforms and environmental contexts, from indoor spaces to a university campus. It is evident that LiDAR-based methods such as FAST-LIO2 and R3LIVE consistently outperform their vision-based counterparts across all scenarios, maintaining a higher trajectory accuracy. On the other hand, the performance of vision-based algorithms, particularly DROID-SLAM, deteriorates as the environment scale increases, with significant scale recovery issues observed in the expansive 
Marching from traditional to deep learning-based SLAM methods, we evaluate DROID-SLAM, an end-to-end deep visual SLAM algorithm. We employ DROID-SLAM on the monocular image stream with a pre-trained model 15 without fine-tuning to present a fair comparison to model-based methods. Meanwhile, testing the limits of its generalization ability is crucial for SLAM. The pre-trained model is trained by supervision from optical flow and poses on the synthetic dataset TartanAir (Wang et al., 2020), covering various conditions (e.g., appearance and viewpoint changes) and environments (e.g., from small-scale indoor to large-scale suburban). All the experiments are conducted on NVIDIA GPU GeForce RTX 3090 with a downsampled image resolution of 320 × 240. The average runtime is 16 FPS with a global bundle adjustment (BA) layer. The average GPU memory consumption is below 11 GB.
7.1.2. Evaluation
Localization accuracy: We calculate translation ATE [m] for each sequence. The best result among all methods is shown in
Our evaluation of SOTA SLAM systems, as summarized in Table 5, demonstrates that each system’s performance varies across different environments, depending on its sensor configuration and algorithmic approach. Due to the precise geometric information inherent in LiDAR raw data, methods incorporating LiDAR generally exhibit higher accuracy. However, as scene scale increases and becomes more complex (like the highway), segments lacking visual texture or structural features become challenging. FAST-LIO2, which utilizes IMU and LiDAR data, showcased robust performance across a diverse array of environments. This highlights the inherent strength of LiDAR-based systems in tackling various and complex scenarios. In contrast, R3LIVE, which integrates IMU, LiDAR, and visual data, consistently demonstrated superior accuracy in different settings, particularly outperforming FAST-LIO2 in scenarios where LiDAR degradation and jerky motion pattern are present (e.g.,
For vision-based methods, VINS-Fusion outperforms DROID-SLAM on average, demonstrating robustness and generalization ability over learning-based methods. However, it is important to note that DROID-SLAM, using only monocular input, surpasses VINS-Fusion in three specific sequences:
7.2. Evaluation of mapping
Localization and mapping represent the foundational tasks for robotic navigation, and evaluating trajectory accuracy alone does not suffice to encapsulate the efficacy of such processes comprehensively. Within the framework of SLAM algorithms predicated on Gaussian models, the map serves as a crucial output, and its accuracy assessment indirectly mirrors the precision of localization. For the broader spectrum of mapping tasks, whether conducted online or offline, sparse or dense, direct evaluation of map accuracy remains crucial. Hence, a module dedicated to assessing map accuracy has been developed to address this need, ensuring a holistic appraisal of navigational competencies.
7.2.1. Experiment setting
For evaluating point cloud maps estimated by SOTA SLAM algorithms, we first downsampled the estimated maps using a 0.1 m grid. Initial alignment with the GT map was performed using CloudCompare software. We set the maximum threshold distance for corresponding points at 0.2 m. Thus, after evaluation, point pairs with a distance less than 0.2 m were considered as the same point for distance calculation.
7.2.2. Evaluation
We use the mapping evaluation metrics in PALoc (Hu et al., 2024b) to complement our localization evaluation. After initial alignment, the error metrics were then calculated using our map evaluation library as introduced in Section 3.3.1. In the presence of high-precision RGB point cloud map ground truth, the accuracy of the maps reconstructed by the algorithm can be evaluated. We register the estimated point cloud map • • •
Figure 15 employs the map evaluation module to present the assessment outcomes of the FAST-LIO2 algorithm across six indoor and outdoor data sequences, with varying colors representing the accuracy levels across different map regions. The map accuracy estimated by FAST-LIO2 notably decreases in outdoor large-scale scenes (Figure 15(a)) or areas with dense vegetation (Figure 15(c) and Figure 15(f)), attributable to significant measurement noise from trees or overall z-axis drift in outdoor LiDAR odometry and mapping applications. Conversely, indoor settings, barring the effects introduced by dynamic obstacles (Figure 15(f)), predominantly exhibit high map quality (Figure 15(e)). This figure presents the mapping performance of FAST-LIO2 in various environments: (a) 
Mapping accuracy: We calculate four metrics to evaluate

This figure demonstrates the map evaluation results within the
7.3. Evaluation of depth estimation
The diversity of sensors, mobile platforms, and scenarios make our dataset appealing for algorithm verification not limited to localization and mapping. In this section, we demonstrate that our dataset can serve for the evaluation of advanced perception algorithms. Due to the easily accessible GT, we set the benchmark for measuring the generalization ability of unsupervised monocular depth prediction. The benchmark measures how unsupervised monocular depth prediction networks could perform on scenes collected from different data collection platforms.
7.3.1. Data preparation
Each frame image is accompanied by a GT depth image of identical size for evaluation. Depth images are produced by projecting point clouds, generated by FAST-LIO2 (Xu et al., 2022) through IMU interpolation, onto these frames:
7.3.2. Experiment setting
Monocular depth estimation tests are essential for evaluating a system’s ability to perceive relative object distances from a single camera view, a key aspect of understanding spatial relationships. Based on the FSNet that is a self-supervised depth estimation model (Liu et al., 2023), we fine-tune this model with our dataset using GT poses. We organize the train-validation data into two groups. The first group allocates 70% of handheld indoor sequence data (i.e.,
7.3.3. Evaluation
We assess models’ performance of unsupervised monocular depth prediction models use the proposed scale-invariant metrics in (Eigen et al., 2014): Results of the unsupervised depth prediction method Liu et al. (2023) which can generalize over different environments. Performance of FSNet MonoDepth (Liu et al., 2023) on FusionPortableV2. Results are categorized by the training domain and testing domain. The left four metrics: ARD, SRD, RMSE-linear, and RMSE-log are error metrics (the lower the better). The right three metric:

7.4. Known issues and limitations
Creating a comprehensive dataset spanning multiple platforms, sensors, and scenes is labor-intensive. Despite our efforts to resolve many issues, we acknowledge the presence of several imperfections within the dataset. We detail these common challenges in the subsequent sections and present our technical solutions. We hope this discussion will provide valuable insights and lessons for future researchers.
7.4.1. Calibration
Achieving the life-long sensor calibration poses significant challenges (Maddern et al., 2017). We try our best to provide the best estimate of calibration parameters. Calibration was performed each time when the data collection platform changed, employing SOTA methods for parameter adjustments, which were also manually verified and fine-tuned. For extrinsic parameters difficult to estimate, such as the relative transformation between specific components, we refer to the CAD model. Efforts were made to reinforce the mechanical structure and minimize external disturbances during the data collection process. Nevertheless, it is acknowledged that high accuracy for specific traversals cannot be assured. We encourage users to use our calibration estimates as initial values and to explore innovative approaches for long-term extrinsic calibration, as suggested in these studies (Luo et al., 2024; Ulrich and Hillemann, 2023; Wu et al., 2021). To aid in these endeavors, we provide raw calibration data and reports, allowing users to develop their methodologies and consider our estimates as a foundational benchmark.
7.4.2. Synchronization
Section 3.1.1 presents our hardware synchronization solution that guarantees the IMU, frame cameras, and LiDAR are triggered by the same clock source. However, the timestamp of the ROS message of each sensor data has minor differences since the time of data transmission and decode varies. For vehicle-related sequences, the average relative time latency (ARTL) among stereo frame images is smaller than 20 ms. This is mainly caused by the long connection between the camera and the signal trigger. For other sequences, the ARTL is smaller than 5 ms. Due to the special design of the event cameras, the ARTL between the left event camera and the LiDAR is unstable and sometimes smaller than 15 ms.
7.4.3. Partial loss of information
In the construction of the dataset, real-world challenges have led to partial loss of sensor information in certain sequences, reflecting practical issues encountered during robotic deployment. Specifically, in
7.4.4. Camera exposure setting
To ensure image consistency during the whole sequence, we fixed the camera exposure time with a specific value before collecting each sequence, mitigating color varies from illumination changes. This scheme is also important to stereo matching since consistent brightness is commonly desirable. However, this scheme can darken images in significantly different lighting conditions, such as entering a tunnel. The darker appearance can be a challenge for most visual perception algorithms.
7.4.5. Limited diversity and volume
Drawing inspiration from foundational models in computer vision and natural language processing, developing a robot-agnostic general model for robotics could be an exciting avenue for future research (Khazatsky et al., 2024; Padalkar et al., 2023; Shah et al., 2023a). However, acquiring a large, diverse real-world dataset poses significant challenges. We address key problems in data collection for field robots, including system integration and data post-processing. But we acknowledge the limitations in our dataset’s scale and diversity: (1)
8. Conclusion and future work
This paper presents the FusionPortableV2 dataset, a comprehensive multi-sensor collection designed to advance research in SLAM and mobile robot navigation. The dataset is built around a compact, multi-sensor device that integrates IMUs, stereo cameras (both frame-based and event-based), LiDAR, and INS, all carefully calibrated and synchronized. This primary device is deployed on various platforms, including a legged robot, a low-speed UGV, and a high-speed vehicle, each equipped with additional platform-specific sensors such as wheel encoders and legged sensors. The FusionPortableV2 dataset features a diverse range of environments, spanning indoor spaces, grasslands, campuses, parking lots, tunnels, downhill roads, and highways. This environmental diversity challenges existing SLAM and navigation technologies with realistic scenarios involving dynamic objects and variable lighting conditions. To ensure the dataset’s utility for the research community, we have meticulously designed 27 sequences, totaling 2.5 hours of data, and provided ground truth data for the objective evaluation of SOTA methods in localization, mapping, and monocular depth estimation.
As we explore future directions, we aim to enhance this dataset’s applicability beyond SLAM by developing novel navigation methods based on the proposed dataset. We will continue to improve the quality of the data and the integration of the system to facilitate easier use by non-expert users. Alongside this dataset, we also release our implementation details and tools to encourage further research advancements. Future update will be provided on the dataset's website.
Supplemental Material
Footnotes
Acknowledgments
Declaration of conflicting interests
Funding
Supplemental Material
Notes
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
