Lidar maps a city

The German Aero Space Center integrates Outsight to fuse GNSS data with LiDAR

The Institute of Traffic System Technology at DLR explores the fusion of GNSS and SLAM methods to achieve robust and geo-referenced localization in dynamic urban traffic.

In the quest for precise perception amidst modern cities’ bustling streets, self-driving cars face a unique challenge.

Traditional GNSS solutions fall short in urban areas, lacking the needed precision. Meanwhile, visual odometry and SLAM offer relative motion data but lack geo-referenced information. 

This article showcases the successful fusion of GNSS and LiDAR solutions thanks to Outsight's processing software integrated into the work of the Institute of Traffic System Technology at the German Aero Space Center, enabling safe navigation through even the tightest construction sites, a feat unattainable with GNSS alone.


In the realm of self-driving cars, two primary approaches emerge: the full perception-based method, striving to replicate human-like awareness by constructing a comprehensive digital representation of the environment, and the map-based approach, reliant on high-precision geo-referenced maps for localization and trajectory planning.

Full perception systems offer flexibility but pose challenges in validation, while map-based systems simplify validation but demand precise localization.

GNSS solutions excel in accuracy but falter in urban areas due to signal interference, while SLAM relies on local sensor data to create virtual maps. To address this, this research combines GNSS and SLAM, achieving accurate geo-referenced localization with varied SLAM map transformations.

Developed in collaboration with Outsight, this approach is part of the Automated Driving test project in the city of Düsseldorf led by the German Aero Space Center.

LiDAR and GNSS Fusion

The localization strategy involves three coordinate frames: odom [for odometry], map, and utm.

The odom frame may drift due to sensor or SLAM inaccuracies, while the map frame provides continuity without drift. The frames are bridged using a static transformation into the utm frame that provides universal transverse mercator (UTM) coordinates.

Control actions occur in the odom frame for smooth car operation, while planning relies on the drift-free map and utm frames.

Outsight’s LiDAR-based SLAM map and Relocalization is used to address GNSS's limitations, offering robustness through environmental perception.

Step 1: Map generation

Outsight's map generation algorithm employs LiDAR-based SLAM (no IMU required).

During LiDAR data recording, the algorithm is supplemented with WGS 84 coordinates of the sensor's position. This creates a trajectory in the SLAM map, preserving world coordinates at each point.

Only trajectory points with precise covariance matrix norms are considered valid, ensuring accurate mapping from local SLAM to WGS84 coordinates while disregarding imprecise sections.

LiDAR maps by Outsight, integrating GNSS data with LiDAR.
Full LiDAR Map provided by Outsight

Step 2: LiDAR-based localization

During autonomous drives, Outsight's SLAM algorithm was used for car localization within the existing SLAM map. This approach also yields robust WGS 84 coordinates.

Assuming the initial LiDAR-based localization in the SLAM map is robust, the algorithm selects nearby points from the trajectory in the SLAM map and calculates barycentric coordinates for the current hypothesis.

These barycentric coordinates, when applied to the corresponding nearby WGS 84 points, determine the LiDAR sensor's current WGS 84 position.

Step 3: uncertainty correction

In the final step, a Kalman filter combines odometry, GNSS localization input, and the LiDAR-based GNSS solution, requiring covariance information. Since the LiDAR localization in the SLAM map lacks this data, an estimation is made.

A minimal covariance matrix is assigned to the LiDAR-based solution to ensure its error ellipse covers the GNSS-based solution without altering the ellipse's aspect ratio. This addition results in a covariance matrix with an error ellipse likely to encompass the true real-world position.

Overview of Outsight algorithms
An overview of some of the most important real-time LiDAR algorithms provided by Outsight’s Software.

Demonstration Setup

The test route, located in Düsseldorf, follows a counterclockwise path around a block of both narrow and wide streets. It encompasses junctions, and is surronded by multi-story buildings limit the sky view, diminishing GNSS precision.

LiDAR map cut-out of a city with a red line drawn on the ground, accompanied by a GNSS track for orientation
LiDAR map cut-out of demo relevant area with one GNSS track as orientation aid

The test vehicle utilized was the ViewCar2, a VW Passat from the German Aerospace Center's Institute of Transportation Research. It came equipped with various sensors for autonomous driving, with key components for localization including:

  • GNSS: Comprising a Novatel SPAN OEM6 with ProPak6 receiver, a Litef μIMU-IC IMU, dual GPS antennas, and RTK correction.
  • LiDAR Sensor: A Velodyne VLP32-C with 32 layers, capable of providing around 600,000 measurements per second and operating at 20 Hz.
  • ALB (LiDAR-based GNSS): An external embedded solution by Outsight. During the demo, it was running alpha firmware version 5.3.0, with all relevant features available in firmware version 5.5.0 and later.
The autonomous driving software comprises multiple components, including the perception system and the controlling/navigation software known as ADORe (Automated Driving Open Research). 

ADORe operates based on map-oriented principles, allowing for empirical assessment of localization accuracy and errors by the driver and co-driver:

  • Lateral errors can be observed as the vehicle follows the virtual lane's centerline.
  • Abrupt steering maneuvers without visible obstacles often signal localization errors in the opposite direction.
  • When the vehicle is at a standstill, a drifting localization solution results in errors in the estimated trajectory.


Acquiring ground truth data for localization in this intricate test site proves costly. Traditional methods, such as employing a total station with a 360° prism, require an unobstructed line of sight, which is hindered by adjacent buildings.

Alternatively, utilizing infrastructure sensors to track the positions of traffic participants is an expensive and unfeasible option for this project.

However, empirical results regarding GNSS and LiDAR-based GNSS accuracy are evident in the recordings. These include fluctuations in GNSS accuracy, contrasting with the stability of LiDAR-based GNSS accuracy, and instances of low GNSS accuracy during specific driving phases.

Rapid improvements in GNSS accuracy result in a noticeable shift in the solution, resembling a drift with an offset from the true position. Subsequently, the GNSS corrects this during a short interval.

Conversely, the fusion of GNSS and LiDAR-based GNSS maintains lateral stability, as seen on the picture below.

Even when GNSS accuracy remains consistently low throughout the drive, the fusion with LiDAR-based GNSS compensates effectively. This situation was intentionally induced by toggling RTK for the GNSS on and off. GNSS inaccuracies misplace the vehicle, while the fused solution provides a plausible trajectory, nearly centered in the lane.

Red Circle: GNSS only Yellow Circle: LiDAR-based GNSS only Blue Circle: Fusion of GNSS
Red Circle – GNSS only, Yellow Circle – LiDAR-based GNSS only and Blue Circle – Fusion of GNSS

Encountered Problems and Suggested Solutions

While not all encountered, issues were addressed in this project. The following insights highlight post-demo improvements:

  • Carefully selecting valid GNSS points for trajectory generation is crucial for accurate LiDAR-based GNSS localization. Large sections lacking valid GNSS data can negatively impact LiDAR-based GNSS solutions due to SLAM algorithm drift. Choosing an appropriate threshold to distinguish valid from invalid trajectory points is essential. In cases where no suitable threshold exists, offline map generation allows manual correction of corresponding GNSS information.
  • The fusion of different localization hypotheses using Kalman filters requires each hypothesis to provide a covariance matrix. Initially, a fixed covariance matrix was assumed for the demo, tailored to the specific test site. However, these fixed values proved insufficient in other test areas, such as Brunswig or Berlin. A heuristic was developed to estimate the covariance of the LiDAR-based GNSS solution (see Section 2.3), allowing for autonomous driving in various areas (both rural and urban) without parameter adjustments.
  • Enhancing the robustness of LiDAR-based localization and reducing the LiDAR map's size involves removing potentially moving objects from the map. A neural network on point clouds (e.g., 3D-MiniNet) is considered for identifying and eliminating points corresponding to dynamic objects like vehicles, pedestrians, and bicycles.
German Aero Space Center integrates Outsight to fuse GNSS data with LiDAR for enhanced spatial awareness.


In this project, Outsight's SLAM-based localization solution was effectively integrated into the Kalman filter-based framework and incorporated into world coordinates.

The demo demonstrates that a localization solution designed to overcome typical GNSS problems plays a vital role in ensuring safe navigation through dense urban traffic.

Additionally, the heuristic for supplying covariance information enhances the robust fusion of GNSS and LiDAR-based localization in world coordinates.

Read the full Research Paper on this application here.

Would you like to find out more about LiDAR applications for the ITS market and see how easy would be to make your own?

Cookbook download for building your ITS, intelligent transportation system, applications

Related Articles