Time to brush up on some calculus. Thomas Paul explains the basics of 3D coordinate systems in a Hackernoon blog post so that everyone can understand how it works.
“When a laser pulse is emitted, its time-of-shooting and direction are registered. The laser pulse travels through the air until it hits an obstacle which reflects some of the energy. The time-of-acquisition and power received are registered by the sensor after receiving the portion of energy. The spherical coordinates of the obstacle are calculated using time-of-acquisition which is returned by the sensor along with power received(as reflectance) after each scan.
As LiDAR sensor returns reading in spherical coordinates, let’s brush up with the spherical coordinate system.
Spherical Coordinate System
In a spherical coordinate system, a point is defined by a distance and two angles. To represent the two angles we use azimuth(θ) and polar angle(ϕ) convention. Thus a point is defined by (r,θ,ϕ).
As you can see from the above diagram, the azimuth angle is in X-Y plane measured from X-axis and polar angle is in Z-Y plane measured from Z axis.
From the above diagram, we can get the following equations for converting a Cartesian coordinate to spherical coordinates.
LiDAR returns reading in spherical coordinates. As you can see from the below diagram, there is a slight difference from the above-discussed convention.
In the sensor coordinate system, a point is defined by (radius r, elevation ω, azimuth α). Elevation angle, ω is in Z-Y plane measured from Y-axis. Azimuth angle, α is in X-Y plane measured from Y-axis.
Azimuth angle depends upon the position at when a laser is fired and is registered at the time of firing. Elevation angle for a laser emitter is fixed in a sensor. The radius is calculated using the time taken by the beam to come back.
Cartesian coordinates can be derived from the following equations.
So a computation is necessary to convert the spherical data from the sensor to Cartesian coordinates using the above equations. Drivers of LiDAR sensors usually do that for us. For example, Velodyne LiDARs sensors provide a ROS package- Velodyne pointcloud for converting the coordinate system.
For more information click here.
References for further reading:
•A story on LiDAR sensor is changing self-driving car blog by Oliver Cameron is a good read.
•Read more about Spherical coordinate system at Wikipedia.
•Velodyne VLP-16 manual is a good start to know more about how LiDAR works.”
Note – If you liked this post click here to stay informed of all of the 3D laser scanning, UAS, autonomous vehicle and Lidar News. If you have an informative 3D video that you would like us to promote, please forward to firstname.lastname@example.org.