The building block of a human nervous system is a neuron and millions of them form a neural network in the body’s central nervous system. To make autonomous cars a reality, computer scientists need to create artificial neural networks (ANNs) that can do the same job as a human’s biological neural network.
From an article in Autocar by Jesse Crosse.
So assuming that really is achievable, the other thing an autonomous car needs is the ability to see, and this is where opinions in the industry are split. Until recently, conventional wisdom had it that as well as the cameras, radars and ultrasonic sensors cars already have for cruise control and advanced driver assistance systems, lidar (light detection and ranging) is essential. Lidar is like high-definition radar, using laser light instead of radio waves to scan a scene and create an accurate HD image of it.
One stumbling block has been the high cost of lidar sensors, which only two years ago cost more than £60,000. Lower-cost versions on the way should bring the price down to around £4000 but that’s still a lot for a single component. Not everyone believes lidar is even necessary or desirable, though, and both Tesla and research scientists at Cornell University have independently arrived at that conclusion.
Cornell found that processing by artificially intelligent (AI) computers can distort camera images viewed from the front. But by changing the perspective in the software to more of a bird’s-eye view, scientists were able to achieve a similar positioning accuracy to lidar using stereo cameras costing a few pounds, placed either side of the windscreen.
Tesla reasons that no human is equipped with laser projectors for eyes and that the secret lies in better understanding the way neural networks identify objects and how to teach them. Whereas a human can identify an object from a single image at a glance, what the computer sees is a matrix of numbers identifying the location and brightness of each pixel in an image.
Because of that, the neural network needs thousands of images to learn the identity of an object, each one labelled to identify it in any situation. Tesla says no chip has yet been produced specifically with neural networking and autonomous driving in mind, so it has spent the past three years designing one.
The new computer can be retro-fitted and has been incorporated in new Teslas since March 2019. The Tesla fleet is already gathering the hundreds of thousands of images needed to train the neural network ‘brains’ in ‘shadow mode’ but without autonomous functions being turned on at this stage. Tesla boss Elon Musk expects to have a complete suite of self-driving software features installed in its cars this year and working robotaxis under test in 2020.
For the complete article click here.
Note – If you liked this post click here to stay informed of all of the 3D laser scanning, geomatics, UAS, autonomous vehicle, Lidar News and more. If you have an informative 3D video that you would like us to promote, please forward to email@example.com and if you would like to join the Younger Geospatial Professional movement click here.