Affiliate Disclosure
If you buy through our links, we may get a commission. Read our ethics policy.

Apple research paper details LiDAR-based 3D object recognition for autonomous vehicle navigation

Last updated

Apple researchers are pushing forward with efforts to bring autonomous vehicle systems to public roads, and last week published an academic paper outlining a method of detecting objects in 3D point clouds using trainable neural networks. While still in its early stages, the technology could mature to improve accuracy in LiDAR navigation solutions.

Like other recent scholarly articles published by Apple engineers, the latest entry, "VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection" by AI researcher Yin Zhou and machine learning specialist Oncel Tuzel, was made public (PDF link) through the arXiv archive of scientific papers.

In its article, Apple notes accurate detection of objects in 3D point clouds, like those generated by LiDAR arrays, is a sticking point in a number of burgeoning real-world applications. From autonomous cars to robotic vacuums, machines that navigate the world around them without the assistance of human operators need to detect critical object with speed and precision.

Compared to 2D image-based detection, LiDAR technology proves to be a more reliable alternative as it provides depth information to better localize objects in space, Apple says. However, LiDAR point clouds, generated by emitting laser pulses and logging the time it takes for the light to return after bouncing off a solid surface, are sparse and have highly variable point density, thus causing a host of problems.

Current state-of-the-art techniques designed to manage data interpretation involve manually creating feature representations for said point clouds. Some methods project point clouds into a bird's eye perspective view, while others transform the data into 3D voxel grids and encode each voxel with certain features. Manually crafting feature representations introduce an "information bottleneck" that restricts such systems from efficiently leveraging 3D shape information, according to Apple.

Instead, Zhou and Tuzel propose the implementation of a trainable deep architecture for point cloud based 3D detection. The framework, called VoxelNet, uses voxel feature encoding (VFE) layers to learn complex features for characterizing 3D shapes. In particular, the technique breaks down the point cloud into 3D voxels, encodes the voxels via stacked VFE layers and renders a volumetric representation.

In tests, Apple's methodology showed promise, outperforming current LiDAR based detection algorithms and image-based approaches "by a large margin." This is according to evaluations run through the KITTI 3D object detection benchmark, which Apple used to assess its process. VoxelNet was trained to detect three basic objects — car, pedestrian and cyclist — in a variety of tests.

Aside from theoretical research, Apple is currently evaluating a self-driving vehicle testbed on the streets of Cupertino, Calif. The company's efforts in autonomous technology began under the "Project Titan" initiative, which sought to build a branded self-driving car from the ground up. After significant investment and multiple employee reassignments, Titan hit a number of snags and was ultimately put on ice in late 2016, though remnants of the initiative, like supporting software and hardware, remain active.

A report in August claimed Apple is looking to parlay the technology into an autonomous shuttle that will ferry employees between its Silicon Valley campuses.

While Apple's research paper focuses heavily on autonomous vehicle navigation, the tech described can also be applied to augmented reality systems that use depth mapping hardware to detect real-world objects. The new iPhone X sports equipment similar to LiDAR arrays in its front-facing TrueDepth camera, which incorporates a miniaturized dot projector for accurate depth mapping operations. If TrueDepth's range is extended, and mounted on the rear of a portable device, it could potentially be paired with advanced software to power an entirely new consumer AR experience.



15 Comments

jd_in_sb 14 Years · 1599 comments

Those images remind me of The Terminator

foggyhill 10 Years · 4767 comments

Funny how the rumor it was "put on ice" is seemingly notworthy even though it was pretty sketchy and there is very little info on the actual scope of project Titan.
Note that Apple spent 1B on didi investment after this thing was "put on ice".... Wonder what thati is all about huh.

Doubtful that this thing is in any way put on ice, they just shifted their focus on public transport (and automated fleets) (especially in asia) for now cause it will be an easier target and you can gather a hell of a lot more data there than anywhere else. Building Apple Transport fleet would be a much more interesting endeavor than building doing a self-driving car sold to individuals.

BTW, analysts just throw things at wall and hope the tech press will be a PR agent for their turd throwing feats. The whole "news" about Apple's foray in cars fit in that mold.

Many of the people that were hired in their buildup for Titan, especially people like mechanical engineers, are STILL AT APPLE. What are they doing there?

radarthekat 12 Years · 3904 comments

I’m just happy the article doesn’t include that Motorweek iCar design render.  Whew!

charlesgres 8 Years · 71 comments

Am I missing something, but if all vehicles are equipped with LIDAR, how does each vehicle distinguish its own dots from the others'?
Can only work if you're the only one on the road, it seems to me..

Rayz2016 8 Years · 6957 comments

Am I missing something, but if all vehicles are equipped with LIDAR, how does each vehicle distinguish its own dots from the others'?
Can only work if you're the only one on the road, it seems to me..

Your dots are the ones zero feet away from the LIDAR?  

Or am I missing something.