New AI algorithm transforms 2D photos into 3D maps

It might allow self-driving cars, drones, and robots to use regular cameras to move around our world.

A new method named MonoXiver uses AI to build up reliable 3D maps of a camera’s surroundings based only on 2D photos. The technology could be especially useful for navigation in self-driving cars. 

The challenge: As 2D representations of a 3D world, photos are missing a huge amount of information about the actual sizes of the objects they portray, as well as their distance from the camera (and each other). These issues with depth and perspective sometimes make for weird, fun optical illusions, but it becomes a serious problem if you need to navigate the real world using 2D cameras.

The problem is especially restricting for self-driving cars, which must constantly scan the road to keep track of potential objects, and the paths and positions of other cars. 

LIDAR is costly, and the hardware can be difficult to integrate with modern car designs.

Without a reliable 3D map of their surroundings, the software driving the vehicles are prone to making mistakes. So far, this has been one of the main barriers to the widespread adoption of self-driving technology.

There are already a few ways around the problem. Today, one of the most widely used techniques involves combining photos with LIDAR, which builds up 3D maps by sending out laser beams in multiple directions, and measuring how long it takes for the light to bounce off surrounding objects. Yet this technology is costly, and the hardware can be difficult to integrate with modern car designs. 

Introducing MonoXiver: A team of researchers led by Xianpeng Liu at North Carolina State University have now developed a smarter solution. Named MonoXiver, their method uses AI to extract 3D information directly from 2D photos.

MonoXiver’s multi-step approach begins with a simple, off-the-shelf monocular camera. These devices use a single lens to roughly estimate 3D geometries in 2D images, based on image cues including shading and lighting, patterns and texture, and the apparent sizes of similar-looking objects.

Based on these estimations, MonoXiver then identifies and encloses important objects in the image – like cars on a street – with a virtual 3D “bounding box.” The boxes help visualise the various scales, aspect ratios, and orientations of 3D objects in a scene. 

An example of bounding boxes and 3D map predicted by the algorithm (highlighted in red) vs. the ground truth data of car locations (in green and blue)
An example of bounding boxes and 3D map predicted by the algorithm (in red) vs. the ground truth data of car locations (green and blue). Credit: Liu et al. 2023. “Monocular 3D Object Detection with Bounding Box Denoising in 3D by Perceiver.”

At first, these bounding box positions are based on the crude estimates of the monocular camera – but using them as a starting point, MonoXiver the re-analyses the area contained inside them, and constructs a secondary group of smaller boxes, which capture many of the smaller details which weren’t indicated by the initial box. 

Using specialised models, MonoXiver can also determine when these small details overlap, and distinguish between them when they are crowded or close together.

Finally, Liu’s team checked whether the details indicated by MonoXiver’s secondary boundary boxes matched what they actually saw in 2D photos. This involved testing whether the smaller boxes contained shapes, colours, and textures consistent with those in the larger initial box.

Improving navigation: Having examined MonoXiver’s performance on two large datasets of 2D photos, the researchers consistently demonstrated its ability to build accurate 3D maps of its surroundings.

Based on this performance, Liu’s team are now hopeful that their approach could transform the ability of AI systems like self-driving cars to navigate 3D space, finding safe and efficient routes even in cluttered urban environments.

It’s possible a similar AI approach could be applied across a diverse array of areas beyond self-driving cars.

On top of this, MonoXiver is highly adaptable, and the researchers say it could be easily modified to work with any monocular camera – making it far more accessible to everyday users than more expensive systems like LIDAR. 

The system isn’t perfect — when challenged with more difficult photos, like cloudy and foggy conditions, it did fail sometimes.

For now, Liu’s team will continue to test and fine-tune their approach, with a view to bring MonoXiver a step closer to commercial rollout. With further adaptation, they also hope a similar AI approach could be applied across a diverse array of areas beyond self-driving cars: from robotics and VR to environmental monitoring and medical imaging. 

We’d love to hear from you! If you have a comment about this article or if you have a tip for a future Freethink story, please email us at [email protected].

Related
Meet the humanoids: 8 robots ready to revolutionize work
Everything you need to know about the humanoids that will soon enter the workforce — or are in it already.
How to build the skills needed for the age of AI
Knowledge-based workers already need to skill-up to coexist with sophisticated artificial intelligence technologies.
See how Moderna is using OpenAI tech across its workforce
A partnership between Moderna and OpenAI provides a real-world example of what can happen when a company leans into generative AI.
Shining a light on oil fields to make them more sustainable
Sensors and analytics give oil well operators real-time alerts when things go wrong, so they can respond before they become disasters.
OpenAI’s GPT-4 outperforms doctors in another new study
OpenAI’s most powerful AI model, GPT-4, outperformed junior doctors in deciding how to treat patients with eye problems.
Up Next
A man with dreadlocks in front of a vibrant background.
Subscribe to Freethink for more great stories