NeRF Google Brain AI 3D Models Photos
Always wanted to relive some of your vacation photos, but only have photos? Well, Google researchers may have the tool for you. They’ve managed to reconstruct detailed 3D scenes of famous landmarks, like the Trevi Fountain in Rome, using regular photographs and machine learning. These aren’t basic models, but rather 3D renderings that allows you to move the camera around like you’re actually there.



NeRF-W, which builds upon NeRF (Google Brain’s Neural Radiance Fields), is the tool they used and it can also reconstruct 3D scenes from a collection of photographs taken in a controlled setting. In other words, it may have trouble with different lighting, image exposure, post processing, random objects, etc.

Sale
Google Pixel 7-5G Android Phone - Unlocked Smartphone with Wide Angle Lens and 24-Hour Battery - 128GB -...
  • Google Tensor G2 Processor: Faster and more efficient than previous Pixel models
  • 5G Connectivity: Allows you to take advantage of faster data speeds on major carriers
  • Adaptive Battery: Can last up to 24 hours on a single charge with extreme battery saver mode

The Neural Radiance Fields (NeRF) approach implicitly models the radiance field and density of a scene within the weights of a neural network. Direct volume rendering is then used to synthesize new views, demonstrating a heretofore unprecedented level of fidelity on a range of challenging scenes,” said the researchers.

Author

A technology, gadget and video game enthusiast that loves covering the latest industry news. Favorite trade show? Mobile World Congress in Barcelona.