3D Coordinate Systems for Lidar Sensor Explained

Time to brush up on some calculus. Thomas Paul explains the basics of 3D coordinate systems in a Hackernoon blog post so that everyone can understand how it works.

“When a laser pulse is emitted, its time-of-shooting and direction are registered. The laser pulse travels through the air until it hits an obstacle which reflects some of the energy. The time-of-acquisition and power received are registered by the sensor after receiving the portion of energy. The spherical coordinates of the obstacle are calculated using time-of-acquisition which is returned by the sensor along with power received(as reflectance) after each scan.

As the LiDAR sensor returns reading in spherical coordinates, let’s brush up with the spherical coordinate system.

Spherical Coordinate System

In a spherical coordinate system, a point is defined by a distance and two angles. To represent the two angles we use azimuth(θ) and polar angle(ϕ) convention. Thus a point is defined by (r,θ,ϕ).

3D Coordinate Systems defined

3D Coordinate Systems – Spherical

As you can see from the above diagram, the azimuth angle is in X-Y plane measured from X-axis and polar angle is in Z-Y plane measured from Z axis.

From the above diagram, we can get the following equations for converting a Cartesian coordinate to spherical coordinates.

3D Coordinate Systems EquationsWe can derive Cartesian coordinates from spherical coordinates using below equations.

3D Coordinate Systems EquationsLiDAR Coordinate System

LiDAR returns reading in spherical coordinates. As you can see from the below diagram, there is a slight difference from the above-discussed convention.

In the sensor coordinate system, a point is defined by (radius r, elevation ω, azimuth α). Elevation angle, ω is in Z-Y plane measured from Y-axis. Azimuth angle, α is in X-Y plane measured from Y-axis.

Azimuth angle depends upon the position at when a laser is fired and is registered at the time of firing. Elevation angle for a laser emitter is fixed in a sensor. The radius is calculated using the time taken by the beam to come back.

3D Coordinate Systems for scanner

3D Coordinate Systems – Scanner

Cartesian coordinates can be derived from the following equations.

The Cartesian coordinate system is easy to manipulate and hence most of the time we need to convert spherical coordinates to a Cartesian system using the above equations.

So a computation is necessary to convert the spherical data from the sensor to Cartesian coordinates using the above equations. Drivers of LiDAR sensors usually do that for us. For example, Velodyne LiDARs sensors provide a ROS package- Velodyne pointcloud for converting the coordinate system.

For more information click here.

References for further reading:

•A story on LiDAR sensor is changing self-driving car blog by Oliver Cameron is a good read.
•Read more about Spherical coordinate system at Wikipedia.
•Velodyne VLP-16 manual is a good start to know more about how LiDAR works.”

Note – If you liked this post click here to stay informed of all of the 3D laser scanning, UAS, autonomous vehicle and Lidar News. If you have an informative 3D video that you would like us to promote, please forward to editor@lidarnews.com.

This entry was posted in 3D Modeling, Education, Laser Scanning, Lidar, point clouds and tagged , , , . Bookmark the permalink.

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.