Modeling of Motion Distortion Effect of Scanning LiDAR Sensors for
Simulation-based Testing
Abstract
Automated vehicles use light detection and ranging (LiDAR) sensors for
environmental scanning. However, the relative motion between the
scanning LiDAR sensor and objects leads to a distortion of the point
cloud. This phenomenon is known as the motion distortion effect,
significantly degrading the sensor’s object detection capabilities and
generating false negative or false positive errors. In this work, we
have introduced ray tracing-based deterministic and analytical
approaches to model the motion distortion effect on the scanning LiDAR
sensor’s performance for simulation-based testing. In addition, we have
performed dynamic test drives at a proving ground to compare real LiDAR
data with the motion distortion effect simulation data. The real-world
scenarios, the environmental conditions, the digital twin of the
scenery, and the object of interest (OOI) are replicated in the virtual
environment of commercial software to obtain the synthetic LiDAR data.
The real and the virtual test drives are compared frame by frame to
validate the motion distortion effect modeling. The mean absolute
percentage error (MAPE), the occupied cell ratio (OCR), and the Barons
cross-correlation coefficient (BCC) are used to quantify the correlation
between the virtual and the real LiDAR point cloud data. The results
show that the deterministic approach matches the real measurements
better than the analytical approach for the scenarios in which the yaw
rate of the ego vehicle changes rapidly.