NVIDIA's New Move: Targeting the Industrial Arena

Advertisements

Recently, NVIDIA captured significant attention within the industry with the launch of its generative AI service designed for 3D modelingWhereas earlier generative AI applications primarily focused on creating content in two dimensions—text, images, and videos—NVIDIA is now pioneering the use of generative AI to assist companies in building 3D assetsThis innovation is poised to accelerate the development of digital twins and simulation industries, thereby enhancing the application of AI in the physical world.

Targeting Industrial Needs with "CUDA Native"

NVIDIA, a global leader in accelerated computing, is hinting at new directions in AI's industrial applicationsDuring two fireside chats at the 2024 SIGGRAPH conference, founder and CEO Jensen Huang shared insights on how generative AI and accelerated computing can transform industries such as manufacturing through visualization

Alongside these discussions, NVIDIA introduced a new set of NIM microservices.

The SIGGRAPH conference is renowned for discussing the latest innovations in computer graphicsAt this event, NVIDIA unveiled generative AI models and NIM microservices that are compatible with OpenUSD, geometry, physics, and materialsOpenUSD is an open-source software that facilitates data interchange within 3D scenes, gradually setting the standard across numerous sectors including 3D visualization, architecture, design, and manufacturing.

These models and services equip developers to expedite application development across industries including manufacturing, automotive, and robotics.

During the discussions, Huang emphasized the importance of creating digital twins and virtual worldsHe stated that industries are leveraging large-scale digital twins, akin to entire cities, to boost efficiency and reduce costs

“For example, before deploying next-generation humanoid robots, AI can train within these virtual realms,” he elaborated.

Why has Huang prioritized the discourse around industrial visualization, virtual worlds, and digital twins? What prompted NVIDIA to introduce its new NIM microservices within the CUDA ecosystem at this juncture?

Rev Lebaredian, NVIDIA's Vice President of Omniverse and Simulation Technology, remarked that the wave of generative AI catering to heavy industries has officially arrivedReports indicate that generative AI is advancing from simplistic applications to intricate production processesThis technical ecosystem heralds an expedited transition.

“Until recently, the primary users of the digital realm were within creative industries; with the latest enhancements and accessibility of NVIDIA NIM microservices for OpenUSD, any sector can now create physics-based virtual worlds and digital twins, preparing for this new wave of AI-driven technologies,” Lebaredian explained.

In the automotive sector, domestic firms are heavily engaged in the push for digital twins

A representative from a major Chinese automotive manufacturer remarked, “Tesla is soon set to release version FSD 12.5 and is actively promoting the FSD rollout in China.” Emphasizing that Tesla considers simulation a strategic priority, the source noted that they are also venturing into the metaverse to resolve issues surrounding autonomous driving data loopsPreviously, the collection of “ghost probe” data was both challenging and costly; however, it is now feasible for companies to address rare scenarios through metaverse simulation environments.

Within the robotics industry, a company specializing in electric inspection robots is training AI in simulated environments, enabling robots to navigate the complexities of power plant infrastructure, plan movement paths, and analyze readings from thousands of meters across different devices.

Architectural design, a multifaceted and time-consuming endeavor, fundamentally relies on 3D models—essential deliverables in the field

alefox

The intricate reconstruction of models representing complex geometries and unconventional structures poses a considerable challengeCurrently, several design firms are collaborating with AI organizations to create models utilizing merely a few photographs, sketches, or textual descriptionsThey’re also able to apply varied materials to refine designs.

In the metallurgical sector, metallographic analysis—a method for assessing internal defects and structures of material slices through microscopy to understand inherent material properties—has historically been inefficient and heavily reliant on human expertiseNowadays, many steel enterprises share a common aspiration: to utilize previous knowledge repositories, training specialized AI to perform comprehensive material analysis.

NVIDIA’s novel NIM microservices allow application companies to build upon existing services without starting from scratch, integrating their data efficiently to realize applications

As such, some firms have characterized this initiative as “CUDA native.”

As generative AI transitions from peripheral scenarios to more central applications, Huang proclaimed, “Everyone will possess their own AI assistant.” The integration of AI with imaging technologies is deepening, and “virtually every industry will feel the impact of this technology—whether it is in scientific calculations that yield improved weather predictions with lower energy consumption, collaborating with creators to generate images, or establishing virtual scenarios for industrial visualization,” Huang assertedHe also pointed out, “Generative AI is set to revolutionize areas such as robotics and driverless vehicles.”

New NIM Microservices: Unleashing Potential

At the core of these industry applications lies the reliance on 3D modeling and simulation technologies.

Historically, constructing 3D content and scenes has posed significant challenges, involving complex workflows of modeling, texturing, animation, lighting, and rendering.

Over the past few decades, animation studios, visual effects companies, and game creators have sought to enhance interoperability between various tools in their workflows, yet with limited success

Transferring data from one location to another proved cumbersome, leading studios to develop convoluted workflows to manage data interoperability.

Moreover, traditional 3D production processes have been linear in collaboration, necessitating multiple formats and edits across various departments, which is time-consuming and resource-intensive.

OpenUSD, an open-source universal 3D data exchange framework, was established in 2023 by NVIDIA, Pixar, Apple, and others to facilitate the seamless construction of virtual worlds via compatible software tools and data types, characterized by high interoperability and compatibility—effectively addressing several challenges encountered during 3D scene creation.

OpenUSD serves as the foundation for NVIDIA’s Omniverse platformIn conversations, Huang referred to OpenUSD as the first format to merge multimodal expressions from nearly all tools

Ideally, over time, individuals could introduce almost any format into the fold, enabling widespread collaboration and the permanent preservation of contentGenerative AI is expected to further enhance the simulation realism within Omniverse.

NVIDIA’s introduction of the NIM microservices specifically designed for OpenUSD marks the world’s first generative AI model tailored for OpenUSD developmentThis service integrates generative AI capabilities into the USD workflow via NIM microservices, substantially lowering the barrier for users to leverage OpenUSDAdditionally, NVIDIA also unveiled a series of new USD connectors applicable to robotic data formats and Apple Vision Pro streaming.

Currently, three NIM microservices have been released: First is the USD Code NIM microservice, capable of answering general OpenUSD questions and generating Python code based on textual prompts.

The second is the USD Search NIM microservice, which enables developers to conduct searches across vast databases of OpenUSD, 3D models, and images using natural language or image inputs, greatly enhancing the speed of material retrieval and processing within companies.

The third is the USD Validate NIM microservice, which checks the compatibility of uploaded files with OpenUSD release versions and generates entirely NVIDIA Omniverse Cloud API-driven RTX rendered path-traced images.

In addition to NVIDIA’s native NIM microservices, ecosystem partners are actively building a variety of popular AI models leveraging these microservices, providing users with optimized inference.

Renowned global creative content platform Shutterstock has launched a new text-to-3D service based on NVIDIA’s latest Edify visual generation model, enabling the creation of 3D prototypes or enriching virtual environments.

For instance, creating accurate lighting that reflects within a virtual scene used to be a complicated endeavor

Previously, creators had to utilize expensive 360-degree camera equipment, physically traveling to shoot scenes from scratch or scouring enormous libraries for similar content.

Now, with 3D generation services, users simply need to describe the environment they need using text or images, yielding high dynamic range panoramic images (360 HDRi) at resolutions up to 16KFurthermore, these scenes and components can be seamlessly switched, allowing for a sports car to appear in diverse backdrops, such as deserts, tropical beaches, or winding mountain roads.

Beyond lighting creation, users can quickly incorporate various rendering materials like concrete, wood, or leather into their 3D assets—these AI-generated assets can now also be edited and provided in various popular file formats.

NVIDIA’s Edify AI model bolstered Getty Images' capabilities, enabling artists to execute precise control over the composition and style of their images

For example, one could place a red beach ball on a picture of a pristine coral reef while retaining precise compositional elementsMoreover, creators can fine-tune existing models using company-specific data to generate images that align with particular branding styles.

These model microservices and tools are significantly accelerating brand creation of 3D assets, thus making the development of digital twins more ubiquitous and accessible.

Early Adopters Are Experimenting

As the creation of 3D content and assets becomes more user-friendly and precise, sectors such as industrial manufacturing, autonomous driving, engineering, and robotics are beginning to reap the benefits of generative AI technologiesParticularly in manufacturing and advertising, a host of pioneering enterprises are actively utilizing the NVIDIA Omniverse platform to hasten the implementation of digital twins and simulations.

Coca-Cola is the first brand to deploy the generative AI capabilities of Omniverse and NIM microservices within its marketing endeavors

In a demonstration video, it needed only to input the natural language request, “Build me a table with tacos and salsa, bathed in morning light.”

Swiftly, the USD Search NIM microservice searched the extensive 3D asset library for suitable models, retrieving them through API calls, while USD Code NIM generated the Python code necessary to combine these models into a scene, thereby significantly enhancing their creative capabilitiesWith generative AI, Coca-Cola can customize its brand imagery across over 100 global markets, achieving localized marketing strategies.

As Coca-Cola's advertising partner, WPP introduced its smart marketing operating system, capitalizing on the Omniverse development platform and OpenUSD to streamline and automate the creation of multilingual text, images, and videos—thus simplifying the material creation processes for advertisers and marketers

By leveraging generative AI to serve its clients, WPP has turned innovative ideas into tangible realities.

As WPP’s Chief Technology Officer articulated, “The beauty of these innovations lies in their high compatibility with our work processes, fully leveraging open standardsNot only does this accelerate future work, but it also solidifies and expands our previous investments in standards like OpenUSDThrough the implementation of NVIDIA NIM microservices and NVIDIA Omniverse, we can launch creative new production tools in partnership with clients like Coca-Cola at unprecedented speeds.”

Foxconn, the world's largest consumer electronics contract manufacturer, has created a virtual digital twin factory for a new facility in Mexico, allowing engineers to define processes and train robots within a virtual environment—enhancing automation levels and production efficiency while saving time, cost, and energy.

Utilizing the Omniverse platform, Foxconn integrated all 3D CAD components into a single virtual factory, employing NVIDIA Isaac Sim—developed on Omniverse and OpenUSD—for scalable robot simulation to train robots, achieving physics-level accuracy and realistic visual presentations for its digital twin.

In addition to Foxconn, electronic manufacturers such as Delta Electronics, MediaTek, and Pegatron are embracing NVIDIA AI and Omniverse to construct their factory digital twins.

Meanwhile, XPeng Motors has utilized the Omniverse platform in the design process of its MPV model, the XPeng X9, transitioning their vehicle development workflow into a virtual space to circumvent bottlenecks faced with traditional workflows.

For instance, the Omniverse platform’s strong interoperability eliminates the need for convoluted conversions of files and data used for industrial modeling, rendering, and 3D effects, accelerating communication and collaboration within XPeng's design team

Post Your Comment