1 / 4

NeRF-ing the world around us: An overview of how Avataar is using NeRFs to make

In "NeRF-ing the world around us," Avataar's innovative use of nerf 3D technology is explored as a means of making 3D models globally accessible. This PDF provides an overview of Avataar's pioneering work in utilizing nerf 3D models to create immersive and interactive environments that can be accessed from anywhere in the world.<br>

AvataarAI
Télécharger la présentation

NeRF-ing the world around us: An overview of how Avataar is using NeRFs to make

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Integrations  Resources  Company  Schedule Demo Home Blog NeRF-ing the world around us: An overview of how Avataar is using NeRFs to make 3D globally accessible Technology NeRF-ing the world around us: An overview of how Avataar is using NeRFs to make 3D globally accessible Table of Contents Making 3D Creation Easy and Available to All Neural Radiance Fields for Quality and Scale Making NeRF’s hyper realistic for real Manasi Mahadik | January 6, 2023 | 5 minutes read | world objects, across categories and surfaces   Every now and again, there comes a technology so indistinguishable from magic that crowds just sit back and wonder Creating explicit Mesh whether civilization around us will ever be the same again. representations with NeRF Extended reality is unquestionably the "closest thing to magic" in this decade, and the good news is that it has ?nally passed its pre-diluvian phase. It is now entering the mainstream and inundating shopping, gaming, media, robotics, and possibly any Extending the NeRF universe interactive medium - with its indomitable in?uence. For more than seven years, Avataar has been at the forefront of this revolution, witnessing its unprecedented impact. An animated illustration of ray tracing and radiance ?eld modeling   Making 3D Creation Easy and Available to All At Avataar, we have been working to establish a new paradigm of visual discovery to provide shoppers with a more seamless shopping experience. A persistent challenge in making 3D universally accessible is that although manually created 3D assets are photorealistic, they are arduous to create and require uncompromised expertise from professional 3D designers. This makes it challenging to scale for all SKUs, encouraging shoppers to make 3D purchase decisions mostly based on 2D images. At Avataar, we have been working at the intersection of deep learning and computer graphics to make 3D assets easy and available for everyone. Developing a 3D camera, and extending it far beyond by making it nimble while retaining its agility is a part of the plan. This blog skims over the breadth of technical problems that we’re actively pursuing. Neural Radiance Fields for Quality and Scale We represent 3D geometry as an implicit neural function. It has recently emerged as a powerful representation for 3D, encoding a high-quality surface with intricate details into the parameters of a deep neural network. This leaves us with an optimization problem to solve - ?tting an implicit neural surface to input observations. This problem is fundamentally tackled by the NeRF (Neural Radiance Fields) based methods, ?rst introduced by the seminal work of Milendhall et al (2020). Technically, we use a set of input images taken from 3 circular trajectories of varying heights from our capture app, the NeRF algorithm shoots rays corresponding to pixels from our input images, sampling 5D coordinates (A 3D X, Y, Z point and a 2D viewing direction) on the rays, passing through a coordinate-based neural network, obtaining density values, and colors for each point in the ray.

  2. Integrations  Resources  Company  Schedule Demo Finally, borrowing long-standing principles of volumetric rendering, the method integrates color values over the ray and gets a ?nal pixel value for each ray. This entire process is iteratively learned via gradient descent, forming a continuous representation and allowing us to do novel view synthesis. Making NeRF’s hyper realistic for real world objects, across categories and surfaces   While this learned 3D representation of the object is robust, potent, and vastly game-changing in itself, it is not the messiah. Training times are sluggish, and real-world captures come with their own idiosyncrasies, capture trajectories, object surfaces, room lighting, the opacity of the object, and the size of the object - all need to be modeled and accounted for. For consumers to make purchase decisions, the 3D representations must be accurate and photorealistic. Our model has now been used on thousands of objects in the wild across a variety of real-world conditions, helping the model get better with time. We have also worked meticulously to create an omni-?exible capture app that guides the user while capturing custom objects in the real world. Creating explicit Mesh representations with NeRF We have also successfully been able to convert these implicit surfaces into high-quality surface representations to accurately reconstruct discrete triangular meshes, along with their corresponding texture patches. Our hybrid architecture consists of an explicit voxel representation with an implicit surface representation. This architecture combines the bene?ts of an explicit representation, which allows for manipulation of the scene, with the representational power of an implicit network. Having mesh outputs allows us to split and edit scenes in traditional 3D software like Blender or Maya. Our focus has been to make the output meshes editable by 3d artists so that they can be used for downstream tasks such as enhancement, animation, etc.

  3. Integrations  Resources  Company  Schedule Demo To model surface lightning, we use BRDF measurements, which describe the angle-dependent optical scatter from these surfaces and can roughly illustrate what the surface may look like under different lighting conditions, as well as disentangle these surfaces into diffused and specular forms. While we are able to generate photorealistic 3D representations, we also need to render them in real-time on smartphones and web devices. We have built our renderer using custom compute shaders combined with our ef?cient ray casting-based rendering algorithm, enabling high FPS neural rendering in 3D and AR. We are actively researching in this area to render large scenes in real-time. Extending the NeRF universe Finally, we are also creating an editing/creator interface to help sellers enhance and build over the generated 3d representations. This includes two major tasks. 1. Disentangling the neural implicit space to enable users to edit elements, including color and texture change, ?oor mapping, part selection, and scene manipulation. 2. Building a set of smart tools using generative models (eg-diffusion models) to go from single/sparse images to 3d. Plus also covers aspects such as human-in-the-loop feedback and online learning to help sellers build better products and virtual room experiences. Our lab is currently a brimming hotbed at the intersection of highly challenging graphics and deep learning problems, coupled with our vision. It’s a special place from where we aim to shape what the creative future around us will look like. It is our ultimate endeavor to empower creators to replace existing 2D content with much more exciting and realistic AR- based 3D visuals. The potential of technology to signi?cantly improve customer shopping experiences excites us. This can enable customers to view more realistic product visualizations from eCommerce stores by simply using their phone's camera. Imagine shopping from the comfort of your home without having to worry about size, texture, look and feel, and, most importantly, without having to deal with the hassle of returns. eCommerce retailers, on the other hand, will be able to create 3D models on their own without going for photo shoots. These 3D outputs have the potential to drastically reduce photography costs. We are excited about the prospects and will keep you updated. Watch this space for updates. Twitter Facebook LinkedIn Continue Reading

  4. Technology Webinar Podcast Resources  Integrations  Company  Schedule Demo Democratizing AR: Making 3D Scalable and Accessible WebAR and its role as a building block for the next stage of the internet Are we at an in?ection point of 3D content creation & AR experiences? Read More Read More Read More Integrations Resources Company Shopify Blogs About Contact Us BigCommerce Knowledge Center Press Enterprise Careers ©Avataar 2023 Privacy Policy Terms & Conditions

More Related