Browsing: Tech

tech news

NVIDIA Honors Americas Partners Advancing Agentic and Physical AI​on March 19, 2025 at 3:00 pm

NVIDIA this week recognized 14 partners leading the way across the Americas for their work advancing agentic and physical AI across industries. The 2025 Americas NVIDIA Partner Network awards — announced at the GTC 2025 global AI conference — represent key efforts by industry leaders to help customers become experts in using AI to solve
Read ArticleNVIDIA this week recognized 14 partners leading the way across the Americas for their work advancing agentic and physical AI across industries. The 2025 Americas NVIDIA Partner Network awards — announced at the GTC 2025 global AI conference — represent key efforts by industry leaders to help customers become experts in using AI to solve
Read Article  

 

NVIDIA this week recognized 14 partners leading the way across the Americas for their work advancing agentic and physical AI across industries.

The 2025 Americas NVIDIA Partner Network awards — announced at the GTC 2025 global AI conference — represent key efforts by industry leaders to help customers become experts in using AI to solve many of today’s greatest challenges. The awards honor the diverse contributions of NPN members fostering AI-driven innovation and growth.

This year, NPN introduced three new award categories that reflect how AI is driving economic growth and opportunities, including:

  • Trailblazer, which honors a visionary partner spearheading AI adoption and setting new industry standards.
  • Rising Star, which celebrates an emerging talent helping industries harness AI to drive transformation.
  • Innovation, which recognizes a partner that’s demonstrated exceptional creativity and forward thinking.

This year’s NPN ecosystem winners have helped companies across industries use AI to adapt to new challenges and prioritize energy-efficient accelerated computing. NPN partners help customers implement a broad range of AI technologies, including NVIDIA-accelerated AI factories, as well as large language models and generative AI chatbots, to transform business operations.

The 2025 NPN award winners for the Americas are:

  • Global Consulting Partner of the Year — Accenture is recognized for its impact and depth of engineering with its AI Refinery platform for industries, simulation and robotics, marketing and sovereignty, which helps organizations enhance innovation and growth with custom-built approaches to AI-driven enterprise reinvention.
  • Trailblazer Partner of the Year — Advizex is recognized for its commitment to driving innovation in AI and high-performance computing, helping industries like healthcare, manufacturing, retail and government seamlessly integrate advanced AI technologies into existing business frameworks. This enables organizations to achieve significant operations efficiencies, enhanced decision-making, and accelerated digital transformation.
  • Rising Star Partner of the Year — AHEAD is recognized for its leadership, technical expertise and deployment of NVIDIA software, NVIDIA DGX systems, NVIDIA HGX and networking technologies to advance AI, benefitting customers across healthcare, financial services, life sciences and higher education.
  • Networking Partner of the Year — Computacenter is recognized for advancing high-performance computing and data centers with NVIDIA networking technologies. The company achieved this by using the NVIDIA AI Enterprise software platform, DGX platforms and NVIDIA networking to drive innovation and growth throughout industries with efficient, accelerated data centers.
  • Solution Integration Partner of the Year — EXXACT is recognized for its efforts in helping research institutions and businesses tap into generative AI, large language models and high-performance computing. The company harnesses NVIDIA GPUs and networking technologies to deliver powerful computing platforms that accelerate innovation and tackle complex computational challenges across various industries.
  • Enterprise Partner of the Year — World Wide Technology (WWT) is recognized for its leadership in advancing AI adoption of customers across industry verticals worldwide. The company expanded its end-to-end AI capabilities by integrating NVIDIA Blueprints into its AI Proving Ground and has made a $500 million commitment to AI development over three years to help speed enterprise generative AI deployments.
  • Software Partner of the Year — Mark III is recognized for the work of its cross-functional team spanning data scientists, developers, 3D artists, systems engineers, and HPC and AI architects, as well as its close collaborations with enterprises and institutions, to deploy NVIDIA software, including NVIDIA AI Enterprise and NVIDIA Omniverse, across industries. These efforts have helped many customers build software-powered pipelines and data flywheels with machine learning, generative AI, high-performance computing and digital twins.
  • Higher Education Research Partner of the Year — Mark III is recognized for its close engagement with universities, academic institutions and research organizations to cultivate the next generation of leaders across AI, machine learning, generative AI, high-performance computing and digital twins.
  • Healthcare Partner of the Year — Lambda is recognized for empowering healthcare and biotech organizations with AI training, fine-tuning and inferencing solutions to speed innovation and drive breakthroughs in AI-driven drug discovery. The company provides AI training, fine-tuning and inferencing solutions at every scale — from individual workstations to comprehensive AI factories — that help healthcare providers seamlessly integrate NVIDIA accelerated computing and software into their infrastructure.
  • Financial Services Partner of the Year — WWT is recognized for driving the digital transformation of the world’s largest banks and financial institutions. The company harnesses NVIDIA AI technologies to optimize data management, enhance cybersecurity and deliver transformative generative AI solutions, helping financial services clients navigate rapid technological changes and evolving customer expectations.
  • Innovation Partner of the Year — Cambridge Computer is recognized for supporting customers deploying transformative technologies, including NVIDIA Grace Hopper, NVIDIA Blackwell and the NVIDIA Omniverse platform for physical AI.
  • Service Delivery Partner of the Year — SoftServe is recognized for its impact in driving enterprise adoption of NVIDIA AI and Omniverse with custom NVIDIA Blueprints that tap into NVIDIA NIM microservices and NVIDIA NeMo and Riva software. SoftServe helps customers create generative AI services for industries spanning manufacturing, retail, financial services, auto, healthcare and life sciences.
  • Distribution Partner of the Year — TD SYNNEX has been recognized for the second consecutive year for supporting customers in accelerating AI growth through rapid delivery of NVIDIA accelerated computing and software, as part of its Destination AI initiative.
  • Rising Star Consulting Partner of the Year Tata Consultancy Services (TCS) is recognized for its growth and commitment to providing industry-specific solutions  that help customers adopt AI faster and at scale. Through its recently launched business unit and center of excellence built on NVIDIA AI Enterprise and Omniverse, TCS is poised to accelerate adoption of agentic AI and physical AI solutions to speed innovation for customers worldwide.
  • Canadian Partner of the Year — Hypertec is recognized for its advancement of high-performance computing and generative AI across Canada. The company has employed the full-stack NVIDIA platform to accelerate AI for financial services, higher education and research.
  • Public Sector Partner of the Year — Government Acquisitions (GAI) is recognized for its rapid AI deployment and robust customer relationships, helping serve the unique needs of the federal government by adding AI to operations to improve public safety and efficiency.

Learn more about the NPN program.

 

NVIDIA this week recognized 14 partners leading the way across the Americas for their work advancing agentic and physical AI across industries. The 2025 Americas NVIDIA Partner Network awards — announced at the GTC 2025 global AI conference — represent key efforts by industry leaders to help customers become experts in using AI to solve
Read Article

Innovation to Impact: How NVIDIA Research Fuels Transformative Work in AI, Graphics and Beyond​on March 20, 2025 at 12:00 am

The roots of many of NVIDIA’s landmark innovations — the foundational technology that powers AI, accelerated computing, real-time ray tracing and seamlessly connected data centers — can be found in the company’s research organization, a global team of around 400 experts in fields including computer architecture, generative AI, graphics and robotics. Established in 2006 and
Read ArticleThe roots of many of NVIDIA’s landmark innovations — the foundational technology that powers AI, accelerated computing, real-time ray tracing and seamlessly connected data centers — can be found in the company’s research organization, a global team of around 400 experts in fields including computer architecture, generative AI, graphics and robotics. Established in 2006 and
Read Article  

 

The roots of many of NVIDIA’s landmark innovations — the foundational technology that powers AI, accelerated computing, real-time ray tracing and seamlessly connected data centers — can be found in the company’s research organization, a global team of around 400 experts in fields including computer architecture, generative AI, graphics and robotics.

Established in 2006 and led since 2009 by Bill Dally, former chair of Stanford University’s computer science department, NVIDIA Research is unique among corporate research organizations — set up with a mission to pursue complex technological challenges while having a profound impact on the company and the world.

“We make a deliberate effort to do great research while being relevant to the company,” said Dally, chief scientist and senior vice president of NVIDIA Research. “It’s easy to do one or the other. It’s hard to do both.”

Dally is among NVIDIA Research leaders sharing the group’s innovations at NVIDIA GTC, the premier developer conference at the heart of AI, taking place this week in San Jose, California.

“We make a deliberate effort to do great research while being relevant to the company.” — Bill Dally, chief scientist and senior vice president

While many research organizations may describe their mission as pursuing projects with a longer time horizon than those of a product team, NVIDIA researchers seek out projects with a larger “risk horizon” — and a huge potential payoff if they succeed.

“Our mission is to do the right thing for the company. It’s not about building a trophy case of best paper awards or a museum of famous researchers,” said David Luebke, vice president of graphics research and NVIDIA’s first researcher. “We are a small group of people who are privileged to be able to work on ideas that could fail. And so it is incumbent upon us to not waste that opportunity and to do our best on projects that, if they succeed, will make a big difference.”

Innovating as One Team

One of NVIDIA’s core values is “one team” — a deep commitment to collaboration that helps researchers work closely with product teams and industry stakeholders to transform their ideas into real-world impact.

“Everybody at NVIDIA is incentivized to figure out how to work together because the accelerated computing work that NVIDIA does requires full-stack optimization,” said Bryan Catanzaro, vice president of applied deep learning research at NVIDIA. “You can’t do that if each piece of technology exists in isolation and everybody’s staying in silos. You have to work together as one team to achieve acceleration.”

When evaluating potential projects, NVIDIA researchers consider whether the challenge is a better fit for a research or product team, whether the work merits publication at a top conference, and whether there’s a clear potential benefit to NVIDIA. If they decide to pursue the project, they do so while engaging with key stakeholders.

“We are a small group of people who are privileged to be able to work on ideas that could fail. And so it is incumbent upon us to not waste that opportunity.” — David Luebke, vice president of graphics research

“We work with people to make something real, and often, in the process, we discover that the great ideas we had in the lab don’t actually work in the real world,” Catanzaro said. “It’s a tight collaboration where the research team needs to be humble enough to learn from the rest of the company what they need to do to make their ideas work.”

The team shares much of its work through papers, technical conferences and open-source platforms like GitHub and Hugging Face. But its focus remains on industry impact.

“We think of publishing as a really important side effect of what we do, but it’s not the point of what we do,” Luebke said.

NVIDIA Research’s first effort was focused on ray tracing, which after a decade of sustained work led directly to the launch of NVIDIA RTX and redefined real-time computer graphics. The organization now includes teams specializing in chip design, networking, programming systems, large language models, physics-based simulation, climate science, humanoid robotics and self-driving cars — and continues expanding to tackle additional areas of study and tap expertise across the globe.

“You have to work together as one team to achieve acceleration.” — Bryan Catanzaro, vice president of applied deep learning research

Transforming NVIDIA — and the Industry

NVIDIA Research didn’t just lay the groundwork for some of the company’s most well-known products — its innovations have propelled and enabled today’s era of AI and accelerated computing.

It began with CUDA, a parallel computing software platform and programming model that enables researchers to tap GPU acceleration for myriad applications. Launched in 2006, CUDA made it easy for developers to harness the parallel processing power of GPUs to speed up scientific simulations, gaming applications and the creation of AI models.

“Developing CUDA was the single most transformative thing for NVIDIA,” Luebke said. “It happened before we had a formal research group, but it happened because we hired top researchers and had them work with top architects.”

Making Ray Tracing a Reality

Once NVIDIA Research was founded, its members began working on GPU-accelerated ray tracing, spending years developing the algorithms and the hardware to make it possible. In 2009, the project — led by the late Steven Parker, a real-time ray tracing pioneer who was vice president of professional graphics at NVIDIA — reached the product stage with the NVIDIA OptiX application framework, detailed in a 2010 SIGGRAPH paper.

The researchers’ work expanded and, in collaboration with NVIDIA’s architecture group, eventually led to the development of NVIDIA RTX ray-tracing technology, including RT Cores that enabled real-time ray tracing for gamers and professional creators.

Unveiled in 2018, NVIDIA RTX also marked the launch of another NVIDIA Research innovation: NVIDIA DLSS, or Deep Learning Super Sampling. With DLSS, the graphics pipeline no longer needs to draw all the pixels in a video. Instead, it draws a fraction of the pixels and gives an AI pipeline the information needed to create the image in crisp, high resolution.

Accelerating AI for Virtually Any Application

NVIDIA’s research contributions in AI software kicked off with the NVIDIA cuDNN library for GPU-accelerated neural networks, which was developed as a research project when the deep learning field was still in its initial stages — then released as a product in 2014.

As deep learning soared in popularity and evolved into generative AI, NVIDIA Research was at the forefront — exemplified by NVIDIA StyleGAN, a groundbreaking visual generative AI model that demonstrated how neural networks could rapidly generate photorealistic imagery.

While generative adversarial networks, or GANs, were first introduced in 2014, “StyleGAN was the first model to generate visuals that could completely pass muster as a photograph,” Luebke said. “It was a watershed moment.”

NVIDIA StyleGAN
NVIDIA StyleGAN

NVIDIA researchers introduced a slew of popular GAN models such as the AI painting tool GauGAN, which later developed into the NVIDIA Canvas application. And with the rise of diffusion models, neural radiance fields and Gaussian splatting, they’re still advancing visual generative AI — including in 3D with recent models like Edify 3D and 3DGUT.

NVIDIA GauGAN
NVIDIA GauGAN

In the field of large language models, Megatron-LM was an applied research initiative that enabled the efficient training and inference of massive LLMs for language-based tasks such as content generation, translation and conversational AI. It’s integrated into the NVIDIA NeMo platform for developing custom generative AI, which also features speech recognition and speech synthesis models that originated in NVIDIA Research.

Achieving Breakthroughs in Chip Design, Networking, Quantum and More

AI and graphics are only some of the fields NVIDIA Research tackles — several teams are achieving breakthroughs in chip architecture, electronic design automation, programming systems, quantum computing and more.

In 2012, Dally submitted a research proposal to the U.S. Department of Energy for a project that would become NVIDIA NVLink and NVSwitch, the high-speed interconnect that enables rapid communication between GPU and CPU processors in accelerated computing systems.

NVLink Switch tray
NVLink Switch tray

In 2013, the circuit research team published work on chip-to-chip links that introduced a signaling system co-designed with the interconnect to enable a high-speed, low-area and low-power link between dies. The project eventually became the link between the NVIDIA Grace CPU and NVIDIA Hopper GPU.

In 2021, the ASIC and VLSI Research group developed a software-hardware codesign technique for AI accelerators called VS-Quant that enabled many machine learning models to run with 4-bit weights and 4-bit activations at high accuracy. Their work influenced the development of FP4 precision support in the NVIDIA Blackwell architecture.

And unveiled this year at the CES trade show was NVIDIA Cosmos, a platform created by NVIDIA Research to accelerate the development of physical AI for next-generation robots and autonomous vehicles. Read the research paper and check out the AI Podcast episode on Cosmos for details.

Learn more about NVIDIA Research at GTC. Watch the keynote by NVIDIA founder and CEO Jensen Huang below:

See notice regarding software product information.

 

The roots of many of NVIDIA’s landmark innovations — the foundational technology that powers AI, accelerated computing, real-time ray tracing and seamlessly connected data centers — can be found in the company’s research organization, a global team of around 400 experts in fields including computer architecture, generative AI, graphics and robotics. Established in 2006 and
Read Article

EPRI, NVIDIA and Collaborators Launch Open Power AI Consortium to Transform the Future of Energy​on March 20, 2025 at 12:00 pm

The power and utilities sector keeps the lights on for the world’s populations and industries. As the global energy landscape evolves, so must the tools it relies on. To advance the next generation of electricity generation and distribution, many of the industry’s members are joining forces through the creation of the Open Power AI Consortium.
Read ArticleThe power and utilities sector keeps the lights on for the world’s populations and industries. As the global energy landscape evolves, so must the tools it relies on. To advance the next generation of electricity generation and distribution, many of the industry’s members are joining forces through the creation of the Open Power AI Consortium.
Read Article  

 

The power and utilities sector keeps the lights on for the world’s populations and industries. As the global energy landscape evolves, so must the tools it relies on.

To advance the next generation of electricity generation and distribution, many of the industry’s members are joining forces through the creation of the Open Power AI Consortium. The consortium includes energy companies, technology companies and researchers developing AI applications to tackle domain-specific challenges, such as adapting to an increased deployment of distributed energy resources and significant load growth on electric grids.

Led by independent, nonprofit energy R&D organization EPRI, the consortium aims to spur AI adoption in the power sector through a collaborative effort to build open models using curated, industry-specific data. The initiative was launched today at NVIDIA GTC, a global AI conference taking place through Friday, March 21, in San Jose, California.

“Over the next decade, AI has the great potential to revolutionize the power sector by delivering the capability to enhance grid reliability, optimize asset performance, and enable more efficient energy management,” said Arshad Mansoor, EPRI’s president and CEO. “With the Open Power AI Consortium, EPRI and its collaborators will lead this transformation, driving innovation toward a more resilient and affordable energy future.”

As part of the consortium, EPRI, NVIDIA and Articul8, a member of the NVIDIA Inception program for cutting-edge startups, are developing a set of domain-specific, multimodal large language models trained on massive libraries of proprietary energy and electrical engineering data from EPRI that can help utilities streamline operations, boost energy efficiency and improve grid resiliency.

The first version of an industry-first open AI model for electric and power systems was developed using hundreds of NVIDIA H100 GPUs and is expected to soon be available in early access as an NVIDIA NIM microservice.

“Working with EPRI, we aim to leverage advanced AI tools to address today’s unique industry challenges, positioning us at the forefront of innovation and operational excellence,” said Vincent Sorgi, CEO of PPL Corporation and EPRI board chair.

PPL is a leading U.S. energy company that provides electricity and natural gas to more than 3.6 million customers in Pennsylvania, Kentucky, Rhode Island and Virginia.

The Open AI Consortium’s Executive Advisory Committee includes executives from over 20 energy companies such as Duke Energy, Exelon, Pacific Gas & Electric Company and Portland General Electric, as well as leading tech companies such as AWS, Oracle and Microsoft. The consortium plans to further expand its global member base.

Powering Up AI to Energize Operations, Drive Innovation

Global energy consumption is projected to grow by nearly 4% annually through 2027, according to the International Energy Agency. To support this surge in demand, electricity providers are looking to enhance the resiliency of power infrastructure, balance diverse energy sources and expand the grid’s capacity.

AI agents trained on thousands of documents specific to this sector — including academic research, industry regulations and standards, and technical documents — can enable utility and energy companies to more quickly assess energy needs and prepare the studies and permits required to improve infrastructure.

“We can bring AI to the global power sector in a much more accelerated way by working together to develop foundation models for the industry, and collaborating with the power sector to y apply solutions tailored to its unique needs,” Mansoor said.

Utilities could tap the consortium’s model to help accelerate interconnection studies, which analyze the feasibility and potential impact of connecting new generators to the existing electric grid. The process varies by region but can take up to four years to complete. By introducing AI agents that can support the analysis, the consortium aims to cut this timeline down by at least 5x.

The AI model could also be used to support the preparation of licenses, permits, environmental studies and utility rate cases, where energy companies seek regulatory approval and public comment on proposed changes to electricity rates.

Beyond releasing datasets and models, the consortium also aims to develop a standardized framework of benchmarks to help utilities, researchers and other energy sector stakeholders evaluate the performance and reliability of AI technologies.

Learn more about the Open Power AI Consortium online and in EPRI’s sessions at GTC:

To learn more about advancements in AI across industries, watch the GTC keynote by NVIDIA founder and CEO Jensen Huang:

See notice regarding software product information.

 

The power and utilities sector keeps the lights on for the world’s populations and industries. As the global energy landscape evolves, so must the tools it relies on. To advance the next generation of electricity generation and distribution, many of the industry’s members are joining forces through the creation of the Open Power AI Consortium.
Read Article

‘Assassin’s Creed Shadows’ Emerges From the Mist on GeForce NOW​on March 20, 2025 at 1:00 pm

Time to sharpen the blade. GeForce NOW brings a legendary addition to the cloud: Ubisoft’s highly anticipated Assassin’s Creed Shadows is now available for members to stream. Plus, dive into the updated version of the iconic Fable Anniversary — part of 11 games joining the cloud this week. Silent as a Shadow Explore 16th-century Japan,
Read ArticleTime to sharpen the blade. GeForce NOW brings a legendary addition to the cloud: Ubisoft’s highly anticipated Assassin’s Creed Shadows is now available for members to stream. Plus, dive into the updated version of the iconic Fable Anniversary — part of 11 games joining the cloud this week. Silent as a Shadow Explore 16th-century Japan,
Read Article  

 

Time to sharpen the blade. GeForce NOW brings a legendary addition to the cloud: Ubisoft’s highly anticipated Assassin’s Creed Shadows is now available for members to stream.

Plus, dive into the updated version of the iconic Fable Anniversary — part of 11 games joining the cloud this week.

Silent as a Shadow

Assassin's Creed Shadows on GeForce NOW
Take the Leap of Faith from the cloud.

Explore 16th-century Japan, uncover conspiracies and shape the destiny of a nation — all from the cloud.

Assassin’s Creed Shadows unfolds in 1579, during the turbulent Azuchi-Momoyama period of feudal Japan, a time of civil war and cultural exchange.

Step into the roles of Naoe, a fictional shinobi assassin and daughter of Fujibayashi Nagato, and Yasuke, a character based on the historical African samurai. Their stories intertwine as they find themselves on opposite sides of a conflict.

The game’s dynamic stealth system enables players to hide in shadows and use a new “Observe” mechanic to identify targets, tag enemies and highlight objectives. Yasuke and Naoe each have unique abilities and playstyles: Naoe excels in stealth, equipped with classic Assassin techniques and shinobi skills, while Yasuke offers a more combat-focused approach.

Navigate the turbulent Sengoku period on GeForce NOW, and experience the game’s breathtaking landscapes and intense combat at up to 4K resolution and 120 frames per second with an Ultimate membership. Every sword clash and sweeping vista is delivered with exceptional smoothness and clarity.

A Classic Reborn

Fable Anniversary revitalizes the original Fable: The Lost Chapters with enhanced graphics, a new save system and Xbox achievements. This action role-playing game invites players to shape their heroes’ destinies in the whimsical world of Albion.

Fable Anniversary on GeForce NOW
Make every choice from the cloud.

Fable Anniversary weaves an epic tale of destiny and choice, following the journey of a young boy whose life is forever changed when bandits raid his peaceful village of Oakvale. Recruited to the Heroes’ Guild, he embarks on a quest to uncover the truth about his family and confront the mysterious Jack of Blades.

Players shape their hero’s destiny through a series of moral choices. These decisions influence the story’s progression and even manifest physically on the character.

Stream the title with a GeForce NOW membership across PCs that may not be game-ready, Macs, mobile devices, and Samsung and LG smart TVs. GeForce NOW transforms these devices into powerful gaming rigs, with up to eight-hour gaming sessions for Ultimate members.

Unleash the Games

Wreckfest 2 on GeForce NOW
Crash, smash, repeat.

Wreckfest 2, the highly anticipated sequel by Bugbear Entertainment to the original demolition derby racing game, promises an even more intense and chaotic experience. The game features a range of customizable cars, from muscle cars to novelty vehicles, each with a story to tell.

Play around with multiple modes, including traditional racing with physics-driven handling, and explore demolition derby arenas where the goal is to cause maximum destruction. With enhanced multiplayer features, including skills-based matchmaking and split-screen mode, Wreckfest 2 is the ultimate playground for destruction-racing enthusiasts.

Look for the following games available to stream in the cloud this week:

What are you planning to play this weekend? Let us know on X or in the comments below.

 

Time to sharpen the blade. GeForce NOW brings a legendary addition to the cloud: Ubisoft’s highly anticipated Assassin’s Creed Shadows is now available for members to stream. Plus, dive into the updated version of the iconic Fable Anniversary — part of 11 games joining the cloud this week. Silent as a Shadow Explore 16th-century Japan,
Read Article

NVIDIA NIM Microservices Now Available to Streamline Agentic Workflows on RTX AI PCs and Workstations​on March 25, 2025 at 1:00 pm

Generative AI is unlocking new capabilities for PCs and workstations, including game assistants, enhanced content-creation and productivity tools and more. NVIDIA NIM microservices, available now, and AI Blueprints, in the coming weeks, accelerate AI development and improve its accessibility. Announced at the CES trade show in January, NVIDIA NIM provides prepackaged, state-of-the-art AI models optimized
Read ArticleGenerative AI is unlocking new capabilities for PCs and workstations, including game assistants, enhanced content-creation and productivity tools and more. NVIDIA NIM microservices, available now, and AI Blueprints, in the coming weeks, accelerate AI development and improve its accessibility. Announced at the CES trade show in January, NVIDIA NIM provides prepackaged, state-of-the-art AI models optimized
Read Article  

 

Generative AI is unlocking new capabilities for PCs and workstations, including game assistants, enhanced content-creation and productivity tools and more.

NVIDIA NIM microservices, available now, and AI Blueprints, coming in April, accelerate AI development and improve its accessibility. Announced at the CES trade show in January, NVIDIA NIM provides prepackaged, state-of-the-art AI models optimized for the NVIDIA RTX platform, including the NVIDIA GeForce RTX 50 Series and, now, the new NVIDIA Blackwell RTX PRO GPUs. The microservices are easy to download and run. They span the top modalities for PC development and are compatible with top ecosystem applications and tools.

The experimental System Assistant feature of Project G-Assist was also released today. Project G-Assist showcases how AI assistants can enhance apps and games. The System Assistant allows users to run real-time diagnostics, get recommendations on performance optimizations, or control system software and peripherals — all via simple voice or text commands. Developers and enthusiasts can extend its capabilities with a simple plug-in architecture and new plug-in builder.

Amid a pivotal moment in computing — where groundbreaking AI models and a global developer community are driving an explosion in AI-powered tools and workflows — NIM microservices, AI Blueprints and G-Assist are helping bring key innovations to PCs. This RTX AI Garage blog series will continue to deliver updates, insights and resources to help developers and enthusiasts build the next wave of AI on RTX AI PCs and workstations.

Ready, Set, NIM! 

Though the pace of innovation with AI is incredible, it can still be difficult for the PC developer community to get started with the technology.

Bringing AI models from research to the PC requires curation of model variants, adaptation to manage all of the input and output data, and quantization to optimize resource usage. In addition, models must be converted to work with optimized inference backend software and connected to new AI application programming interfaces (APIs). This takes substantial effort, which can slow AI adoption.

NVIDIA NIM microservices help solve this issue by providing prepackaged, optimized, easily downloadable AI models that connect to industry-standard APIs. They’re optimized for performance on RTX AI PCs and workstations, and include the top AI models from the community, as well as models developed by NVIDIA.

NIM microservices support a range of AI applications, including large language models (LLMs), vision language models, image generation, speech processing, retrieval-augmented generation  (RAG)-based search, PDF extraction and computer vision. Ten NIM microservices for RTX are available, supporting a range of applications, including language and image generation, computer vision, speech AI and more. Get started with these NIM microservices today:

NIM microservices are also available through top AI ecosystem tools and frameworks.

For AI enthusiasts, AnythingLLM and ChatRTX now support NIM, making it easy to chat with LLMs and AI agents through a simple, user-friendly interface. With these tools, users can create personalized AI assistants and integrate their own documents and data, helping automate tasks and enhance productivity.

For developers looking to build, test and integrate AI into their applications, FlowiseAI and Langflow now support NIM and offer low- and no-code solutions with visual interfaces to design AI workflows with minimal coding expertise. Support for ComfyUI is coming soon. With these tools, developers can easily create complex AI applications like chatbots, image generators and data analysis systems.

In addition, Microsoft VS Code AI Toolkit, CrewAI and Langchain now support NIM and provide advanced capabilities for integrating the microservices into application code, helping ensure seamless integration and optimization.

Visit the NVIDIA technical blog and build.nvidia.com to get started.

NVIDIA AI Blueprints Will Offer Pre-Built Workflows

NVIDIA AI Blueprints, coming in April, give AI developers a head start in building generative AI workflows with NVIDIA NIM microservices.

Blueprints are ready-to-use, extensible reference samples that bundle everything needed — source code, sample data, documentation and a demo app — to create and customize advanced AI workflows that run locally. Developers can modify and extend AI Blueprints to tweak their behavior, use different models or implement completely new functionality.

PDF to podcast AI Blueprint coming soon.

The PDF to podcast AI Blueprint will transform documents into audio content so users can learn on the go. By extracting text, images and tables from a PDF, the workflow uses AI to generate an informative podcast. For deeper dives into topics, users can then have an interactive discussion with the AI-powered podcast hosts.

The AI Blueprint for 3D-guided generative AI will give artists finer control over image generation. While AI can generate amazing images from simple text prompts, controlling image composition using only words can be challenging. With this blueprint, creators can use simple 3D objects laid out in a 3D renderer like Blender to guide AI image generation. The artist can create 3D assets by hand or generate them using AI, place them in the scene and set the 3D viewport camera. Then, a prepackaged workflow powered by the FLUX NIM microservice will use the current composition to generate high-quality images that match the 3D scene.

NVIDIA NIM on RTX With Windows Subsystem for Linux

One of the key technologies that enables NIM microservices to run on PCs is Windows Subsystem for Linux (WSL).

Microsoft and NVIDIA collaborated to bring CUDA and RTX acceleration to WSL, making it possible to run optimized, containerized microservices on Windows. This allows the same NIM microservice to run anywhere, from PCs and workstations to the data center and cloud.

Get started with NVIDIA NIM on RTX AI PCs at build.nvidia.com.

Project G-Assist Expands PC AI Features With Custom Plug-Ins

As part of Project G-Assist, an experimental version of the System Assistant feature for GeForce RTX desktop users is now available via the NVIDIA App, with laptop support coming soon.

G-Assist helps users control a broad range of PC settings — including optimizing game and system settings, charting frame rates and other key performance statistics, and controlling select peripherals settings such as lighting — all via basic voice or text commands.

G-Assist is built on NVIDIA ACE — the same AI technology suite game developers use to breathe life into non-player characters. Unlike AI tools that use massive cloud-hosted AI models that require online access and paid subscriptions, G-Assist runs locally on a GeForce RTX GPU. This means it’s responsive, free and can run without an internet connection. Manufacturers and software providers are already using ACE to create custom AI Assistants like G-Assist, including MSI’s AI Robot engine, the Streamlabs Intelligent AI Assistant and upcoming capabilities in HP’s Omen Gaming hub.

G-Assist was built for community-driven expansion. Get started with this NVIDIA GitHub repository, including samples and instructions for creating plug-ins that add new functionality. Developers can define functions in simple JSON formats and drop configuration files into a designated directory, allowing G-Assist to automatically load and interpret them. Developers can even submit plug-ins to NVIDIA for review and potential inclusion.

Currently available sample plug-ins include Spotify, to enable hands-free music and volume control, and Google Gemini — allowing G-Assist to invoke a much larger cloud-based AI for more complex conversations, brainstorming sessions and web searches using a free Google AI Studio API key.

In the clip below, you’ll see G-Assist ask Gemini about which Legend to pick in Apex Legends when solo queueing, and whether it’s wise to jump into Nightmare mode at level 25 in Diablo IV:

For even more customization, follow the instructions in the GitHub repository to generate G-Assist plug-ins using a ChatGPT-based “Plug-in Builder.” With this tool, users can write and export code, then integrate it into G-Assist — enabling quick, AI-assisted functionality that responds to text and voice commands.

Watch how a developer used the Plug-in Builder to create a Twitch plug-in for G-Assist to check if a streamer is live:

More details on how to build, share and load plug-ins are available in the NVIDIA GitHub repository.

Check out the G-Assist article for system requirements and additional information.

Build, Create, Innovate

NVIDIA NIM microservices for RTX are available at build.nvidia.com, providing developers and AI enthusiasts with powerful, ready-to-use tools for building AI applications.

Download Project G-Assist through the NVIDIA App’s “Home” tab, in the “Discovery” section. G-Assist currently supports GeForce RTX desktop GPUs, as well as a variety of voice and text commands in the English language. Future updates will add support for GeForce RTX Laptop GPUs, new and enhanced G-Assist capabilities, as well as support for additional languages. Press “Alt+G” after installation to activate G-Assist.

Each week, RTX AI Garage features community-driven AI innovations and content for those looking to learn more about NIM microservices and AI Blueprints, as well as building AI agents, creative workflows, digital humans, productivity apps and more on AI PCs and workstations.

Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X — and stay informed by subscribing to the RTX AI PC newsletter.

Follow NVIDIA Workstation on LinkedIn and X.

See notice regarding software product information.

 

Generative AI is unlocking new capabilities for PCs and workstations, including game assistants, enhanced content-creation and productivity tools and more. NVIDIA NIM microservices, available now, and AI Blueprints, in the coming weeks, accelerate AI development and improve its accessibility. Announced at the CES trade show in January, NVIDIA NIM provides prepackaged, state-of-the-art AI models optimized
Read Article

Buzz Solutions Uses Vision AI to Supercharge the Electric Grid​on March 26, 2025 at 3:00 pm

The reliability of the electric grid is critical. From handling demand surges and evolving power needs to preventing infrastructure failures that can cause wildfires, utility companies have a lot to keep tabs on. Buzz Solutions — a member of the NVIDIA Inception program for cutting-edge startups — is helping by using AI to improve how
Read ArticleThe reliability of the electric grid is critical. From handling demand surges and evolving power needs to preventing infrastructure failures that can cause wildfires, utility companies have a lot to keep tabs on. Buzz Solutions — a member of the NVIDIA Inception program for cutting-edge startups — is helping by using AI to improve how
Read Article  

 

The reliability of the electric grid is critical.

From handling demand surges and evolving power needs to preventing infrastructure failures that can cause wildfires, utility companies have a lot to keep tabs on.

Buzz Solutions — a member of the NVIDIA Inception program for cutting-edge startups — is helping by using AI to improve how utilities monitor and maintain their infrastructure.

Kaitlyn Albertoli, CEO and cofounder of Buzz Solutions joined the AI Podcast to explain how the company’s vision AI technology helps utilities spot potential problems faster.

Buzz Solutions helps utility companies analyze the massive amounts of inspection data collected by drones and helicopters. The company’s proprietary machine learning algorithms identify potential issues ranging from broken and rusted components to encroaching vegetation and unwelcome wildlife visits — before they cause outages or wildfires.

To help address substation issues, Buzz Solutions built PowerGUARD, a container-based application pipeline that uses AI to analyze video streams from substation cameras in real time. It detects security, safety, fire, smoke and equipment issues, annotates the video, then sends alerts via email or to a dashboard.

PowerGUARD uses the NVIDIA DeepStream software development kit for processing and inference of video streams used in real-time video analytics. DeepStream runs within the NVIDIA Metropolis framework on the NVIDIA Jetson edge AI platform or on cloud-based virtual machines ‌to improve performance, reduce costs and save time.

Albertoli believes AI is just getting started in the utility industry, as it enables workers to take action rather than spend months reviewing images manually. “We are just at the tip of the iceberg of seeing AI enter into the energy sector and start to provide real value,” she said.

Time Stamps

05:15: How Buzz Solutions saw an opportunity in the massive amounts of inspection data utility companies were collecting but not analyzing.

12:25: The importance of modernizing energy infrastructure with actionable intelligence.

16:27: How AI identifies critical risks like rusted components, vegetation encroachment and sparking issues before they cause wildfires.

20:00: Buzz Solutions’ innovative use of synthetic data to train algorithms for rare events.

You Might Also Like…

Telenor Builds Norway’s First AI Factory, Offering Sustainable and Sovereign Data Processing

Telenor opened Norway’s first AI factory in November 2024, enabling organizations to process sensitive data securely on Norwegian soil while prioritizing environmental responsibility. Telenor’s Chief Innovation Officer and Head of the AI Factory Kaaren Hilsen discusses the AI factory’s rapid development, going from concept to reality in under a year.

NVIDIA’s Josh Parker on How AI and Accelerated Computing Drive Sustainability

AI isn’t just about building smarter machines. It’s about building a greener world. AI and accelerated computing are helping industries tackle some of the world’s toughest environmental challenges. Joshua Parker, senior director of corporate sustainability at NVIDIA, explains how these technologies are powering a new era of energy efficiency.

Currents of Change: ITIF’s Daniel Castro on Energy-Efficient AI and Climate Change

AI is everywhere. So, too, are concerns about advanced technology’s environmental impact. Daniel Castro, vice president of the Information Technology and Innovation Foundation and director of its Center for Data Innovation, discusses his AI energy use report that addresses misconceptions about AI’s energy consumption. He also talks about the need for continued development of energy-efficient technology.

Subscribe to the AI Podcast

Get the AI Podcast through Amazon Music, Apple Podcasts, Google Podcasts, Google Play, Castbox, DoggCatcher, Overcast, PlayerFM, Pocket Casts, Podbay, PodBean, PodCruncher, PodKicker, SoundCloud, Spotify, Stitcher and TuneIn.

 

The reliability of the electric grid is critical. From handling demand surges and evolving power needs to preventing infrastructure failures that can cause wildfires, utility companies have a lot to keep tabs on. Buzz Solutions — a member of the NVIDIA Inception program for cutting-edge startups — is helping by using AI to improve how
Read Article

The Dream Life Awaits: Play ‘inZOI’ on GeForce NOW Anytime, Anywhere​on March 27, 2025 at 1:00 pm

A new resident is moving into the cloud — KRAFTON’s inZOI joins the 2,000+ games in the GeForce NOW cloud gaming library. Plus, members can get ready for an exclusive sneak peek as the Sunderfolk First Look Demo comes to the cloud. The demo is exclusively available for players on GeForce NOW until April 7,
Read ArticleA new resident is moving into the cloud — KRAFTON’s inZOI joins the 2,000+ games in the GeForce NOW cloud gaming library. Plus, members can get ready for an exclusive sneak peek as the Sunderfolk First Look Demo comes to the cloud. The demo is exclusively available for players on GeForce NOW until April 7,
Read Article  

 

A new resident is moving into the cloud — KRAFTON’s inZOI joins the 2,000+ games in the GeForce NOW cloud gaming library.

Plus, members can get ready for an exclusive sneak peek as the Sunderfolk First Look Demo comes to the cloud. The demo is exclusively available for players on GeForce NOW until April 7, including Performance and Ultimate members as well as free users.

And explore the world of Atomfall — part of 12 games joining the cloud this week.

Cloud of Possibilities

inZOI on GeForce NOW
Live the life of your dreams in the cloud.

In inZOI — a groundbreaking life simulation game by Krafton that pushes the genre’s boundaries — take on the role of an intern at AR COMPANY, managing virtual beings called “Zois” in a simulated city.

The game features over 400 mental elements influencing Zois’ behaviors. Experience the game’s dynamic weather system, open-world environments inspired by real locations and cinematic cut scenes for key life events — and even create in-game objects. inZOI lets players craft unique stories and live out their dreams in a meticulously designed virtual world.

Dive into the world of Zois without the need for high-end hardware. Members can manage their virtual homes, customize characters and explore the game’s dynamic environments from various devices, streaming its detailed graphics and complex simulations with ease.

A Magical Gateway

Sunderfolk’s First Look Demo has arrived on GeForce NOW, offering a tantalizing look into the magical realm of the Sunderlands. Designed as a TV-first experience, this shared-turn-based tactical role-playing game (RPG) enables using a mobile phone as the gameplay controller. Up to four players can gather around the big screen and embark on a journey filled with strategic battles.

This second-screen approach keeps players engaged in real time, adding new layers of immersion. With all six unique character classes unlocked from the start, players can experience the early hours of the game, experimenting with different team compositions and tactics to overcome the challenges that await.

Sunderfolk First Look Demo on GeForce NOW
Let the magic begin.

Accessing the demo is a breeze — head to the GeForce NOW app, select Sunderfolk and jump right in. Explore the Sunderlands, engage in flexible turn-based combat and help rebuild the village of Arden to get a taste of the full game’s depth and camaraderie.

Gather the gaming squad, grab a phone and prepare to write a completely new legend in this RPG adventure. The First Look Demo is only available on GeForce NOW, where members can enjoy high-quality graphics and seamless gameplay on their phones and tablets, along with the innovative mobile-as-controller mechanic that makes Sunderfolk’s couch co-op experience so engaging.

Epic Adventures Await

Atomfall on GeForce NOW
Enter a world where danger lurks in every shadow.

Blending folk horror and intense combat, Atomfall is a survival-action game set in an alternate 1960s Britain, where the Windscale nuclear disaster has left Northern England a radioactive wasteland. Players explore eerie open zones filled with mutated creatures, cultists and Cold War mysteries while scavenging resources, crafting weapons and uncovering the truth behind the disaster. GeForce NOW members can stream it today across their devices of choice.

Look for the following games available to stream in the cloud this week:

  • Sunderfolk First Look Demo (New release, March 25)
  • Atomfall (New release on Steam and Xbox available on PC Game Pass, March 27)
  • The First Berserker: Khazan (New release on Steam, March 27)
  • inZOI (New release on Steam, March 27)
  • Beholder (Epic Games Store)
  • Bus Simulator 21 (Epic Games Store)
  • Galacticare (Xbox, available on PC Game Pass)
  • Half-Life 2 RTX Demo (Steam)
  • The Legend of Heroes: Trails through Daybreak II (Steam)
  • One Lonely Outpost (Xbox, available on PC Game Pass)
  • Psychonauts (Xbox, available on PC Game Pass)
  • Undying (Epic Games Store)

What are you planning to play this weekend? Let us know on X or in the comments below.

 

A new resident is moving into the cloud — KRAFTON’s inZOI joins the 2,000+ games in the GeForce NOW cloud gaming library. Plus, members can get ready for an exclusive sneak peek as the Sunderfolk First Look Demo comes to the cloud. The demo is exclusively available for players on GeForce NOW until April 7,
Read Article

Industrial Ecosystem Adopts Mega NVIDIA Omniverse Blueprint to Train Physical AI in Digital Twins​on March 31, 2025 at 4:00 pm

Advances in physical AI are enabling organizations to embrace embodied AI across their operations, bringing unprecedented intelligence, automation and productivity to the world’s factories, warehouses and industrial facilities.  Humanoid robots can work alongside human teams, autonomous mobile robots (AMRs) can navigate complex warehouse environments, and intelligent cameras and visual AI agents can monitor and optimize
Read ArticleAdvances in physical AI are enabling organizations to embrace embodied AI across their operations, bringing unprecedented intelligence, automation and productivity to the world’s factories, warehouses and industrial facilities.  Humanoid robots can work alongside human teams, autonomous mobile robots (AMRs) can navigate complex warehouse environments, and intelligent cameras and visual AI agents can monitor and optimize
Read Article  

 

Advances in physical AI are enabling organizations to embrace embodied AI across their operations, bringing unprecedented intelligence, automation and productivity to the world’s factories, warehouses and industrial facilities.

Humanoid robots can work alongside human teams, autonomous mobile robots (AMRs) can navigate complex warehouse environments, and intelligent cameras and visual AI agents can monitor and optimize entire facilities. In these ways, physical AI is becoming integral to today’s industrial operations.

Helping industrial enterprises accelerate the development, testing and deployment of physical AI, the Mega NVIDIA Omniverse Blueprint for testing multi-robot fleets in digital twins is now available in preview on build.nvidia.com.

At Hannover Messe — a trade show on industrial development running through April 4 in Germany — manufacturing, warehousing and supply chain leaders such as Accenture and Schaeffler are showcasing their adoption of the blueprint to simulate Digit, a humanoid robot from Agility Robotics, and discussing how they use industrial AI and digital twins to optimize facility layouts, material flow and collaboration between humans and robots inside complex production environments.

In addition, NVIDIA  ecosystem partners — including Delta Electronics, Rockwell Automation and Siemens — are announcing further integrations with NVIDIA Omniverse and NVIDIA AI technologies at the event.

Digital Twins — the Training Ground for Physical AI

Industrial facility digital twins are physically accurate virtual replicas of real-world facilities that serve as critical testing grounds for simulating and validating physical AI and how robots and autonomous fleets interact, collaborate and tackle complex tasks before  deployment.

Developers can use NVIDIA Omniverse platform technologies and the Universal Scene Description (OpenUSD) framework to develop digital twins of their facilities and processes. This simulation-first approach dramatically accelerates development cycles while reducing the costs and risks associated with real-world testing.

Built for a Diversity of Robots and AI Agents

The Mega blueprint equips industrial enterprises with a reference workflow for combining sensor simulation and synthetic data generation to simulate complex human-robot interactions and verify the performance of autonomous systems in industrial digital twins.

Enterprises can use Mega to test various robot brains and policies at scale for mobility, navigation, dexterity and spatial reasoning. This enables fleets comprising different types of robots to work together as a coordinated system.

As robot brains execute their missions in simulation, they perceive the results of their actions through sensor simulation and plan their next action. This cycle continues until the policies are refined and ready for deployment.

Once validated, these policies are deployed to real robots, which continue to learn from their environment — sending sensor information back through the entire loop and creating a continuous learning and improvement cycle.

Transforming Industrial Operations With Visual AI Agents

In addition to AMRs and humanoid robots, advanced visual AI agents extract information from live and recorded video data, enabling new levels of intelligence and automation. These visual AI agents bring real-time contextual awareness to robots and help to improve worker safety, maintain warehouse compliance, support visual inspection and maximize space utilization.

To support developers building visual AI agents, which can be integrated with the Mega blueprint, NVIDIA last year announced an AI Blueprint for video search and summarization (VSS). At Hannover Messe, leading partners are featuring how they use the VSS blueprint to improve productivity and operational efficiency.

Accelerating Industrial Digitalization

The industrial world is now experiencing its software-defined moment, with visual AI agents and digital twins as the training ground for physical AI.

Join NVIDIA and its partners at Hannover Messe to discover how AI agents and real-time simulation, powered by NVIDIA’s Three Computer Solution, are reshaping industrial workflows and driving innovation, automation and efficiency in manufacturing.

Read the technical blog to learn more about the Mega blueprint for industrial robot fleets. See the blueprint in action on this interactive demo page.

Stay up to date by subscribing to NVIDIA news, joining the Omniverse community and following NVIDIA Omniverse on Instagram, LinkedIn, Medium and X.

Explore the new self-paced Learn OpenUSD training curriculum that includes free NVIDIA Deep Learning Institute courses for 3D practitioners and developers.

Featured image courtesy of Accenture, Agility Robotics and Schaeffler.

See notice regarding software product information.

 

Advances in physical AI are enabling organizations to embrace embodied AI across their operations, bringing unprecedented intelligence, automation and productivity to the world’s factories, warehouses and industrial facilities.  Humanoid robots can work alongside human teams, autonomous mobile robots (AMRs) can navigate complex warehouse environments, and intelligent cameras and visual AI agents can monitor and optimize
Read Article

NVIDIA GeForce RTX 50 Series Accelerates Adobe Premiere Pro and Media Encoder’s 4:2:2 Color Sampling​on April 2, 2025 at 1:00 pm

Video editing workflows are getting a lot more colorful. Adobe recently announced massive updates to Adobe Premiere Pro (beta) and Adobe Media Encoder, including PC support for 4:2:2 video color editing. The 4:2:2 color format is a game changer for professional video editors, as it retains nearly as much color information as 4:4:4 while greatly
Read ArticleVideo editing workflows are getting a lot more colorful. Adobe recently announced massive updates to Adobe Premiere Pro (beta) and Adobe Media Encoder, including PC support for 4:2:2 video color editing. The 4:2:2 color format is a game changer for professional video editors, as it retains nearly as much color information as 4:4:4 while greatly
Read Article  

 

Video editing workflows are getting a lot more colorful.

Adobe recently announced massive updates to Adobe Premiere Pro (beta) and Adobe Media Encoder, including PC support for 4:2:2 video color editing.

The 4:2:2 color format is a game changer for professional video editors, as it retains nearly as much color information as 4:4:4 while greatly reducing file size. This improves color grading and chroma keying — using color information to isolate a specific range of hues — while maximizing efficiency and quality.

In addition, new NVIDIA GeForce RTX 5090 and 5080 laptops — built on the NVIDIA Blackwell architecture — are out now, accelerating 4:2:2 and advanced AI-powered features across video-editing workflows.

Adobe and other industry partners are attending NAB Show — a premier gathering of over 100,000 leaders in the broadcast, media and entertainment industries — running April 5-9 in Las Vegas. Professionals in these fields will come together for education, networking and exploring the latest technologies and trends.

Shed Some Color on 4:2:2

Consumer cameras that are limited to 4:2:0 color compression capture a limited amount of color information. 4:2:0 is acceptable for video playback on browsers, but professional video editors often rely on cameras that capture 4:2:2 color depth with precise color accuracy to ensure higher color fidelity.

Adobe Premiere Pro’s beta with 4:2:2 means video data can now provide double the color information with just a 1.3x increase in raw file size over 4:2:0. This unlocks several key benefits within professional video-production workflows:

Increased Color Accuracy: 10-bit 4:2:2 retains more color information compared with 8-bit 4:2:0, leading to more accurate color representation and better color grading results.

4:2:2 offers more accurate color representation for better color grading results.

More Flexibility: The extra color data allows for increased flexibility during color correction and grading, enabling more nuanced adjustments and corrections.

Improved Keying: 4:2:2 is particularly beneficial for keying — including green screening — as it enables cleaner, more accurate extraction of the subject from the background, as well as cleaner edges of small keyed objects like hair.

4:2:2 enables cleaner green screen video content.

Smaller File Sizes: Compared with 4:4:4, 4:2:2 reduces file sizes without significantly impacting picture quality, offering an optimal balance between quality and storage.

Combining 4:2:2 support with NVIDIA hardware increases creative possibilities.

Advanced Video Editing

Prosumer-grade cameras from most major brands support HEVC and H.264 10-bit 4:2:2 formats to deliver superior image quality, manageable file sizes and the flexibility needed for professional video production.

GeForce RTX 50 Series GPUs paired with Microsoft Windows 11 come with GPU-powered decode acceleration in HEVC and H.264 10-bit 4:2:2 formats.

GPU-powered decode enables faster-than-real-time playback without stuttering, the ability to work with original camera media instead of proxies, smoother timeline responsiveness and reduced CPU load — freeing system resources for multi-app workflows and creative tasks.

RTX 50 Series’ 4:2:2 hardware can decode up to six 4K 60 frames-per-second video sources on an RTX 5090-enabled studio PC, enabling smooth multi-camera video-editing workflows on Adobe Premiere Pro.

Video exports are also accelerated with NVIDIA’s ninth-generation encoder and sixth-generation decoder.

NVIDIA and GeForce RTX Laptop GPU encoders and decoders.

In GeForce RTX 50 Series GPUs, the ninth-generation NVIDIA video encoder, NVENC, offers an 8% BD-BR upgrade in video encoding efficiency when exporting to HEVC on Premiere Pro.

Adobe AI Accelerated

Adobe delivers an impressive array of advanced AI features for idea generation, enabling streamlined processes, improved productivity and opportunities to explore new artistic avenues — all accelerated by NVIDIA RTX GPUs.

For example, Adobe Media Intelligence, a feature in Premiere Pro (beta) and After Effects (beta), uses AI to analyze footage and apply semantic tags to clips. This lets users more easily and quickly find specific footage by describing its content, including objects, locations, camera angles and even transcribed spoken words.

Media Intelligence runs 30% faster on the GeForce RTX 5090 Laptop GPU compared with the GeForce RTX 4090 Laptop GPU.

In addition, the Enhance Speech feature in Premiere Pro (beta) improves the quality of recorded speech by filtering out unwanted noise and making the audio sound clearer and more professional. Enhance Speech runs 7x faster on GeForce RTX 5090 Laptop GPUs compared to the MacBook Pro M4 Max.

Visit Adobe’s Premiere Pro page to download a free trial of the beta and explore the slew of AI-powered features across the Adobe Creative Cloud and Substance 3D apps.

Unleash (AI)nfinite Possibilities

GeForce RTX 5090 and 5080 Series laptops deliver the largest-ever generational leap in portable performance for creating, gaming and all things AI.

They can run creative generative AI models such as Flux up to 2x faster in a smaller memory footprint, compared with the previous generation.

The previously mentioned ninth-generation NVIDIA encoders elevate video editing and livestreaming workflows, and come with NVIDIA DLSS 4 technology and up to 24GB of VRAM to tackle massive 3D projects.

NVIDIA Max-Q hardware technologies use AI to optimize every aspect of a laptop — the GPU, CPU, memory, thermals, software, display and more — to deliver incredible performance and battery life in thin and quiet devices.

All GeForce RTX 50 Series laptops include NVIDIA Studio platform optimizations, with over 130 GPU-accelerated content creation apps and exclusive Studio tools including NVIDIA Studio Drivers, tested extensively to enhance performance and maximize stability in popular creative apps.

The game-changing NVIDIA GeForce RTX 5090 and 5080 GPU laptops are available now.

Adobe will participate in the Creator Lab at NAB Show, offering hands-on training for editors to elevate their skills with Adobe tools. Attend a 30-minute section and try out Puget Systems laptops equipped with GeForce RTX 5080 Laptop GPUs to experience blazing-fast performance and demo new generative AI features.

Use NVIDIA’s product finder to explore available GeForce RTX 50 Series laptops with complete specifications.

New creative app updates and optimizations are powered by the NVIDIA Studio platform. Follow NVIDIA Studio on Instagram, X and Facebook. Access tutorials on the Studio YouTube channel and get updates directly in your inbox by subscribing to the Studio newsletter

See notice regarding software product information.

 

Video editing workflows are getting a lot more colorful. Adobe recently announced massive updates to Adobe Premiere Pro (beta) and Adobe Media Encoder, including PC support for 4:2:2 video color editing. The 4:2:2 color format is a game changer for professional video editors, as it retains nearly as much color information as 4:4:4 while greatly
Read Article

Speed Demon: NVIDIA Blackwell Takes Pole Position in Latest MLPerf Inference Results​on April 2, 2025 at 3:00 pm

In the latest MLPerf Inference V5.0 benchmarks, which reflect some of the most challenging inference scenarios, the NVIDIA Blackwell platform set records — and marked NVIDIA’s first MLPerf submission using the NVIDIA GB200 NVL72 system, a rack-scale solution designed for AI reasoning. Delivering on the promise of cutting-edge AI takes a new kind of compute
Read ArticleIn the latest MLPerf Inference V5.0 benchmarks, which reflect some of the most challenging inference scenarios, the NVIDIA Blackwell platform set records — and marked NVIDIA’s first MLPerf submission using the NVIDIA GB200 NVL72 system, a rack-scale solution designed for AI reasoning. Delivering on the promise of cutting-edge AI takes a new kind of compute
Read Article  

 

In the latest MLPerf Inference V5.0 benchmarks, which reflect some of the most challenging inference scenarios, the NVIDIA Blackwell platform set records — and marked NVIDIA’s first MLPerf submission using the NVIDIA GB200 NVL72 system, a rack-scale solution designed for AI reasoning.

Delivering on the promise of cutting-edge AI takes a new kind of compute infrastructure, called AI factories. Unlike traditional data centers, AI factories do more than store and process data — they manufacture intelligence at scale by transforming raw data into real-time insights. The goal for AI factories is simple: deliver accurate answers to queries quickly, at the lowest cost and to as many users as possible.

The complexity of pulling this off is significant and takes place behind the scenes. As AI models grow to billions and trillions of parameters to deliver smarter replies, the compute required to generate each token increases. This requirement reduces the number of tokens that an AI factory can generate and increases cost per token. Keeping inference throughput high and cost per token low requires rapid innovation across every layer of the technology stack, spanning silicon, network systems and software.

The latest updates to MLPerf Inference, a peer-reviewed industry benchmark of inference performance, include the addition of Llama 3.1 405B, one of the largest and most challenging-to-run open-weight models. The new Llama 2 70B Interactive benchmark features much stricter latency requirements compared with the original Llama 2 70B benchmark, better reflecting the constraints of production deployments in delivering the best possible user experiences.

In addition to the Blackwell platform, the NVIDIA Hopper platform demonstrated exceptional performance across the board, with performance increasing significantly over the last year on Llama 2 70B thanks to full-stack optimizations.

NVIDIA Blackwell Sets New Records

The GB200 NVL72 system — connecting 72 NVIDIA Blackwell GPUs to act as a single, massive GPU — delivered up to 30x higher throughput on the Llama 3.1 405B benchmark over the NVIDIA H200 NVL8 submission this round. This feat was achieved through more than triple the performance per GPU and a 9x larger NVIDIA NVLink interconnect domain.

While many companies run MLPerf benchmarks on their hardware to gauge performance, only NVIDIA and its partners submitted and published results on the Llama 3.1 405B benchmark.

Production inference deployments often have latency constraints on two key metrics. The first is time to first token (TTFT), or how long it takes for a user to begin seeing a response to a query given to a large language model. The second is time per output token (TPOT), or how quickly tokens are delivered to the user.

The new Llama 2 70B Interactive benchmark has a 5x shorter TPOT and 4.4x lower TTFT — modeling a more responsive user experience. On this test, NVIDIA’s submission using an NVIDIA DGX B200 system with eight Blackwell GPUs tripled performance over using eight NVIDIA H200 GPUs, setting a high bar for this more challenging version of the Llama 2 70B benchmark.

Combining the Blackwell architecture and its optimized software stack delivers new levels of inference performance, paving the way for AI factories to deliver higher intelligence, increased throughput and faster token rates.

NVIDIA Hopper AI Factory Value Continues Increasing

The NVIDIA Hopper architecture, introduced in 2022, powers many of today’s AI inference factories, and continues to power model training. Through ongoing software optimization, NVIDIA increases the throughput of Hopper-based AI factories, leading to greater value.

On the Llama 2 70B benchmark, first introduced a year ago in MLPerf Inference v4.0, H100 GPU throughput has increased by 1.5x. The H200 GPU, based on the same Hopper GPU architecture with larger and faster GPU memory, extends that increase to 1.6x.

Hopper also ran every benchmark, including the newly added Llama 3.1 405B, Llama 2 70B Interactive and graph neural network tests. This versatility means Hopper can run a wide range of workloads and keep pace as models and usage scenarios grow more challenging.

It Takes an Ecosystem

This MLPerf round, 15 partners submitted stellar results on the NVIDIA platform, including ASUS, Cisco, CoreWeave, Dell Technologies, Fujitsu, Giga Computing, Google Cloud, Hewlett Packard Enterprise, Lambda, Lenovo, Oracle Cloud Infrastructure, Quanta Cloud Technology, Supermicro, Sustainable Metal Cloud and VMware.

The breadth of submissions reflects the reach of the NVIDIA platform, which is available across all cloud service providers and server makers worldwide.

MLCommons’ work to continuously evolve the MLPerf Inference benchmark suite to keep pace with the latest AI developments and provide the ecosystem with rigorous, peer-reviewed performance data is vital to helping IT decision makers select optimal AI infrastructure.

Learn more about MLPerf

Images and video taken at an Equinix data center in the Silicon Valley.

 

In the latest MLPerf Inference V5.0 benchmarks, which reflect some of the most challenging inference scenarios, the NVIDIA Blackwell platform set records — and marked NVIDIA’s first MLPerf submission using the NVIDIA GB200 NVL72 system, a rack-scale solution designed for AI reasoning. Delivering on the promise of cutting-edge AI takes a new kind of compute
Read Article

NVIDIA’s Jacob Liberman on Bringing Agentic AI to Enterprises​on April 2, 2025 at 4:00 pm

AI is rapidly transforming how organizations solve complex challenges. The early stages of enterprise AI adoption focused on using large language models to create chatbots. Now, enterprises are using agentic AI to create intelligent systems that reason, act and execute complex tasks with a degree of autonomy. Jacob Liberman, director of product management at NVIDIA,
Read ArticleAI is rapidly transforming how organizations solve complex challenges. The early stages of enterprise AI adoption focused on using large language models to create chatbots. Now, enterprises are using agentic AI to create intelligent systems that reason, act and execute complex tasks with a degree of autonomy. Jacob Liberman, director of product management at NVIDIA,
Read Article  

 

AI is rapidly transforming how organizations solve complex challenges.

The early stages of enterprise AI adoption focused on using large language models to create chatbots. Now, enterprises are using agentic AI to create intelligent multi-agent systems that reason, act and execute complex tasks with a degree of autonomy.

Jacob Liberman, director of product management at NVIDIA, joined the NVIDIA AI Podcast to explain how agentic AI bridges the gap between powerful AI models and practical enterprise applications.

Enterprises are deploying AI agents to free human workers from time-consuming and error-prone tasks. This allows people to spend more time on high-value work that requires creativity and strategic thinking.

Liberman anticipates it won’t be long before teams of AI agents and human workers collaborate to tackle complex tasks requiring reasoning, intuition and judgement. For example, enterprise software developers will work with AI agents to develop more efficient algorithms. And medical researchers will collaborate with AI agents to design and test new drugs.

NVIDIA AI Blueprints help enterprises build their own AI agents – including many of the use cases listed above.

“Blueprints are reference architectures implemented in code that show you how to take NVIDIA software and apply it to some productive task in an enterprise to solve a real business problem,” Liberman said.

The blueprints are entirely open source. A developer or service provider can deploy a blueprint directly, or customize it by integrating their own technology.

Liberman highlighted the versatility of the AI Blueprint for customer service, for example, which features digital humans.

“The digital human can be made into a bedside digital nurse, a sportscaster or a bank teller with just some verticalization,” he said.

Other popular NVIDIA Blueprints include a video search and summarization agent, an enterprise multimodal PDF chatbot and a generative virtual screening pipeline for drug discovery.

Time Stamps: 

1:14 – What is an AI agent?

17:25 – How software developers are early adopters of agentic AI.

19:50 – Explanation of test-time compute and reasoning models.

23:05 – Using AI agents in cybersecurity and risk management applications.

You Might Also Like…

Imbue CEO Kanjun Que on Transforming AI Agents Into Personal Collaborators

Kanjun Qiu, CEO of Imbue, discusses the emerging era of personal AI agents, drawing a parallel to the PC revolution and explaining how modern AI systems are evolving to enhance user capabilities through collaboration.

Telenor’s Kaaren Hilsen on Launching Norway’s First AI Factory

Kaaren Hilsen, chief innovation officer and head of the AI factory at Telenor, highlights Norway’s first AI factory, which securely processes sensitive data within the country while promoting data sovereignty and environmental sustainability through green computing initiatives, including a renewable energy-powered data center in Oslo.

Firsthand’s Jon Heller Shares How AI Agents Enhance Consumer Journeys in Retail 

Jon Heller of Firsthand explains how the company’s AI Brand Agents are boosting retail and digital marketing by personalizing customer experiences and converting marketing interactions into valuable research data.

 

AI is rapidly transforming how organizations solve complex challenges. The early stages of enterprise AI adoption focused on using large language models to create chatbots. Now, enterprises are using agentic AI to create intelligent systems that reason, act and execute complex tasks with a degree of autonomy. Jacob Liberman, director of product management at NVIDIA,
Read Article

No Foolin’: GeForce NOW Gets 21 Games in April​on April 3, 2025 at 1:00 pm

GeForce NOW isn’t fooling around. This month, 21 games are joining the cloud gaming library of over 2,000 titles. Whether chasing epic adventures, testing skills in competitive battles or diving into immersive worlds, members can dive into April’s adventures arrivals, which are truly no joke. Get ready to stream, play and conquer the eight games
Read ArticleGeForce NOW isn’t fooling around. This month, 21 games are joining the cloud gaming library of over 2,000 titles. Whether chasing epic adventures, testing skills in competitive battles or diving into immersive worlds, members can dive into April’s adventures arrivals, which are truly no joke. Get ready to stream, play and conquer the eight games
Read Article  

 

GeForce NOW isn’t fooling around.

This month, 21 games are joining the cloud gaming library of over 2,000 titles. Whether chasing epic adventures, testing skills in competitive battles or diving into immersive worlds, members can dive into April’s adventures arrivals, which are truly no joke.

Get ready to stream, play and conquer the eight games available this week. Members can also get ahead of the pack with advanced access to South of Midnight, streaming soon before launch.

Unleash the Magic

South of Midnight, an action-adventure game developed by Compulsion Games, offers advanced access for gamers who purchase its Premium Edition. Dive into the title’s hauntingly beautiful world before launch, exploring its rich Southern gothic setting and unique magical combat system while balancing magic with melee attacks.

South of Midnight Advanced Access on GeForce NOW
Step into the shadows.

Set in a mystical version of the American South, the game combines elements of magic, mystery and adventure, weaving a compelling story that draws players in. The endless opportunities for exploration and combat, along with deep lore and engaging characters, make the game a must-play for fans of the action-adventure genre.

With its blend of dark fantasy and historical influences, South of Midnight is poised to deliver a unique gaming experience that will leave players spellbound.

GeForce NOW members can be among the first to get advanced access to the game without the hassle of downloads or updates. With an Ultimate or Performance membership, experience the game’s haunting landscapes and cryptid encounters with the highest frame rates and lowest latency — no need for the latest hardware.

April Is Calling

Call of Duty Warzone Season 3 on GeForce NOW
Verdansk is back! Catch it in the cloud.

Verdansk, the original and iconic map from Call of Duty: Warzone, is making its highly anticipated return in the game’s third season, and available to stream on GeForce NOW. Known for its sprawling urban areas, rugged wilderness and points of interest like Dam and Superstore, Verdansk offers a dynamic battleground for intense combat. The map has been rebuilt from the ground up with key enhancements across audio, visuals and gameplay, getting back to basics and delivering nostalgia for fans.

Look for the following games available to stream in the cloud this week:

Here’s what to expect for April: 

  • South of Midnight (New release on Steam and Xbox, available on PC Game Pass, April 8)
  • Commandos Origins (New release on Steam and Xbox, available on PC Game Pass, April 9)
  • The Talos Principle: Reawakened (New release on Steam, April 10)
  • Night Is Coming (New release on Steam, April 14)
  • Mandragora: Whispers of the Witch Tree (New release on Steam, April 17)
  • Sunderfolk (New release on Steam, April 23)
  • Clair Obscur: Expedition 33 (New release on Steam and Xbox, available on PC Game Pass, April 24)
  • Tempest Rising (New release on Steam, April 24)
  • Aimlabs (Steam)
  • Backrooms: Escape Together (Steam)
  • Blood Strike (Steam) 
  • ContractVille (Steam)
  • EXFIL (Steam)

March Madness

In addition to the 14 games announced last month, 26 more joined the GeForce NOW library:

What are you planning to play this weekend? Let us know on X or in the comments below.

 

GeForce NOW isn’t fooling around. This month, 21 games are joining the cloud gaming library of over 2,000 titles. Whether chasing epic adventures, testing skills in competitive battles or diving into immersive worlds, members can dive into April’s adventures arrivals, which are truly no joke. Get ready to stream, play and conquer the eight games
Read Article

Nintendo Switch 2 Leveled Up With NVIDIA AI-Powered DLSS and 4K Gaming​on April 3, 2025 at 1:00 pm

The Nintendo Switch 2, unveiled April 2, takes performance to the next level, powered by a custom NVIDIA processor featuring an NVIDIA GPU with dedicated RT Cores and Tensor Cores for stunning visuals and AI-driven enhancements. With 1,000 engineer-years of effort across every element — from system and chip design to a custom GPU, APIs
Read ArticleThe Nintendo Switch 2, unveiled April 2, takes performance to the next level, powered by a custom NVIDIA processor featuring an NVIDIA GPU with dedicated RT Cores and Tensor Cores for stunning visuals and AI-driven enhancements. With 1,000 engineer-years of effort across every element — from system and chip design to a custom GPU, APIs
Read Article  

 

The Nintendo Switch 2, unveiled April 2, takes performance to the next level, powered by a custom NVIDIA processor featuring an NVIDIA GPU with dedicated RT Cores and Tensor Cores for stunning visuals and AI-driven enhancements.

With 1,000 engineer-years of effort across every element — from system and chip design to a custom GPU, application programming interfaces (APIs) and world-class development tools — the Nintendo Switch 2 brings major upgrades.

The new console enables up to 4K gaming in TV mode and up to 120 frames per second at 1080p in handheld mode. Nintendo Switch 2 also supports high dynamic range and AI upscaling to sharpen visuals and smooth gameplay.

AI and Ray Tracing for Next-Level Visuals

The new RT Cores bring real-time ray tracing, delivering lifelike lighting, reflections and shadows for more immersive worlds.

Tensor Cores power AI-driven features like Deep Learning Super Sampling (DLSS), boosting resolution for sharper details without sacrificing image quality.

Tensor Cores also enable AI-powered face tracking and background removal in video chat use cases, enhancing social gaming and streaming.

With millions of players worldwide, the Nintendo Switch has become a gaming powerhouse and home to Nintendo’s storied franchises. Its hybrid design redefined console gaming, bridging TV and handheld play.

More Power, Smoother Gameplay

With 10x the graphics performance of the Nintendo Switch, the Nintendo Switch 2 delivers smoother gameplay and sharper visuals.

  • Tensor Cores boost AI-powered graphics while keeping power consumption efficient.
  • RT Cores enhance in-game realism with dynamic lighting and natural reflections.
  • Variable refresh rate via NVIDIA G-SYNC in handheld mode ensures ultra-smooth, tear-free gameplay.

Tools for Developers, Upgrades for Players

Developers get improved game engines, better physics and optimized APIs for faster, more efficient game creation.

Powered by NVIDIA technologies, Nintendo Switch 2 delivers for both players and developers.

 

The Nintendo Switch 2, unveiled April 2, takes performance to the next level, powered by a custom NVIDIA processor featuring an NVIDIA GPU with dedicated RT Cores and Tensor Cores for stunning visuals and AI-driven enhancements. With 1,000 engineer-years of effort across every element — from system and chip design to a custom GPU, APIs
Read Article

NVIDIA Showcases Real-Time AI and Intelligent Media Workflows at NAB​on April 3, 2025 at 3:00 pm

Real-time AI is unlocking new possibilities in media and entertainment, improving viewer engagement and advancing intelligent content creation.  At NAB Show, a premier conference for media and entertainment running April 5-9 in Las Vegas, NVIDIA will showcase how emerging AI tools and the technologies underpinning them help streamline workflows for streamers, content creators, sports leagues
Read ArticleReal-time AI is unlocking new possibilities in media and entertainment, improving viewer engagement and advancing intelligent content creation.  At NAB Show, a premier conference for media and entertainment running April 5-9 in Las Vegas, NVIDIA will showcase how emerging AI tools and the technologies underpinning them help streamline workflows for streamers, content creators, sports leagues
Read Article  

 

Real-time AI is unlocking new possibilities in media and entertainment, improving viewer engagement and advancing intelligent content creation. 

At NAB Show, a premier conference for media and entertainment running April 5-9 in Las Vegas, NVIDIA will showcase how emerging AI tools and the technologies underpinning them help streamline workflows for streamers, content creators, sports leagues and broadcasters.  

Attendees can experience the power of the NVIDIA Blackwell platform, which serves as the foundation of NVIDIA Media2 — a collection of NVIDIA technologies including NVIDIA NIM microservices and NVIDIA AI Blueprints for live video analysis, accelerated computing platforms and generative AI software.   

Attendees can also see NVIDIA Holoscan for Media — an advanced real-time AI platform designed for live media workflows and applications — in action at the Dell booth, as well as experience the NVIDIA AI Blueprint for video search and summarization, which makes it easy to build and customize video analytics AI agents.  

NVIDIA will also present in these sessions: 

Driving Innovation With Partners  

Partners across the industry are showcasing innovative solutions using NVIDIA technologies to accelerate live media. 

Amazon Web Services (booth W1701) will collaborate with NVIDIA to showcase an esport racing challenge through a live cloud production. The professional-grade racing simulator allows users to analyze their performance through cutting-edge AI-powered insights and step into the spotlight for their own post-race interview. Other demos will offer a peek into the future of live cloud production and generative AI in sports broadcasting. 

Beamr (booth SL1730MR) will demonstrate how it’s driving AV1 adoption with GPU-accelerated video processing. Beamr’s technology, powered by the NVIDIA NVENC encoder, enables cost-efficient, high-quality and scalable AV1 transformation. 

Dell (booth SL4616) is collaborating with a wide range of partners to highlight their latest innovations in the media industry. Autodesk will feature its Flame visual effects software for AI-driven compositing; Avid will demonstrate real-time editing and AI metadata tagging on Dell Pro Max high-performance PCs; and Boris FX and RE:Vision Effects will showcase their motion-tracking, slow-motion interpolation and object-removal technologies — all running on NVIDIA accelerated computing. In addition, Speed Read AI will showcase the use of NVIDIA RTX-powered workstations to analyze scripts in seconds, while Arcitecta and Elements will demonstrate high-speed media collaboration and post-production workflows on Dell PowerScale storage.  

HP (booth SL3723) will showcase its desktop and mobile workstation portfolio with NVIDIA RTX PRO Blackwell GPUs, delivering cutting-edge AI performance in a variety of use cases. Attendees can also find HP’s newly announced AI solutions, the HP ZGX Nano AI Station G1n and HP ZGX Fury AI Station G1n, developed in collaboration with NVIDIA.  

Qvest (booth W2055) will spotlight two new AI solutions that help clients increase audience engagement, simplify insight gathering and streamline workflows. The Agentic Live Multi-Camera Video Event Extractor identifies, detects and extracts near-real-time events into structured outputs in an easily configurable, natural language, no-code interface, and the No-Code Media-Centric AI Agent Builder extracts meaningful structured data from unstructured media formats including video, images and complex documents. Both use NVIDIA NIM microservices, NVIDIA NeMo, NVIDIA Holoscan for Media, the NVIDIA AI Blueprint for video search and summarization and more. 

Monks (booth W2530) will announce its complete suite of products and services for the media and entertainment industry, designed to drive innovation, monetization and efficiency. Monks uses tools under NVIDIA Media2, such as NIDIA NIM microservices and Holoscan for Media, to enable real-time audience feedback, AI-powered selective encoding and contextual content analysis for large archives. The company will also launch a new suite of vision language model service offerings with its strategic partner TwelveLabs.  

Supermicro (booth W3713) will demonstrate the ease of setting up and running a complete AI video pipeline with WAN 2.1 and Adobe Premiere Pro, all running on the new high-performance Supermicro AS -531AW-TC workstation with an NVIDIA RTX PRO 6000 Blackwell Workstation Edition GPU. With RAVEL Orchestrate handling workstation and AI cluster orchestration, everything can run smoothly — from setup and deployment to user access and workload management.  

Speechmatics (booth W2317) will demonstrate its speech-to-text technology, which taps into NVIDIA accelerated computing to deliver highly accurate, real-time transcription across multiple languages and use cases, from media production to broadcast captioning. 

Telestream (booth W1501) will showcase its waveform monitoring solution, which seeks to bridge the gap for cloud-native workflows with a microservices architecture that taps into NVIDIA Holoscan for Media. In collaboration with NVIDIA, Telestream will demonstrate the ability to introduce cloud-native waveform monitoring to replicate broadcast center and master control room capabilities for engineering and creative teams. 

TwelveLabs (booth W3921) will showcase its newest models, which are being trained in part on NVIDIA DGX Cloud, to bring state-of-the-art video understanding to the world’s largest sports teams, clubs and leagues. The company is currently developing models based on NVIDIA NIM microservices to bring media and entertainment customers highly efficient inference and easy integration with leading software frameworks and agentic applications. 

VAST Data (booth SL9213) will spotlight the VAST InsightEngine — a solution that securely ingests, processes, and retrieves all enterprise data in real-time —– in a demo powered by the NVIDIA AI Enterprise software platform. Developed in collaboration with the National Hockey League, the demo showcases instant access to an archive of over 550,000 hours of hockey game footage. The work is set to redefine sponsorship analytics and empower video producers to instantly search, edit and deliver dynamic broadcast clips — fueling hyper-personalized fan experiences. 

Vizrt (booth W3031) will present its solution portfolio, which when matched with NVIDIA accelerated computing and NVIDIA Maxine technology, simplifies complex processes to support the immersive talent reflections, shadow casting and 3D pose tracking of Reality Connect, in addition to Particle Effects, Talent Gesture Control, XR Draw and the AI Gaze Correction feature available in the TriCaster Vizion. 

 V-Nova (booth W1252 and W1454) will spotlight its 6DoF virtual-reality experiences with new immersive content — Sharkarma and Weightless in booth W1252 — and AI-accelerated optimization in booth W1454, demonstrating how NVIDIA NVENC and NVIDIA GPUs unlock incredible video quality, efficiency and performance for critical video, AI and VR streaming cloud applications. 

Join NVIDIA at NAB Show 2025

 

Real-time AI is unlocking new possibilities in media and entertainment, improving viewer engagement and advancing intelligent content creation.  At NAB Show, a premier conference for media and entertainment running April 5-9 in Las Vegas, NVIDIA will showcase how emerging AI tools and the technologies underpinning them help streamline workflows for streamers, content creators, sports leagues
Read Article

From Browsing to Buying: How AI Agents Enhance Online Shopping​on April 3, 2025 at 3:00 pm

Editor’s note: This post is part of the AI On blog series, which explores the latest techniques and real-world applications of agentic AI, chatbots and copilots. The series also highlights the NVIDIA software and hardware powering advanced AI agents, which form the foundation of AI query engines that gather insights and perform tasks to transform
Read ArticleEditor’s note: This post is part of the AI On blog series, which explores the latest techniques and real-world applications of agentic AI, chatbots and copilots. The series also highlights the NVIDIA software and hardware powering advanced AI agents, which form the foundation of AI query engines that gather insights and perform tasks to transform
Read Article  

 

Editor’s note: This post is part of the AI On blog series, which explores the latest techniques and real-world applications of agentic AI, chatbots and copilots. The series also highlights the NVIDIA software and hardware powering advanced AI agents, which form the foundation of AI query engines that gather insights and perform tasks to transform everyday experiences and reshape industries.

Online shopping puts a world of choices at people’s fingertips, making it convenient for them to purchase and receive orders — all from the comfort of their homes.

But too many choices can turn experiences from exciting to exhausting, leaving shoppers struggling to cut through the noise and find exactly what they need.

By tapping into AI agents, retailers can deepen their customer engagement, enhance their offerings and maintain a competitive edge in a rapidly shifting digital marketplace.

Every digital interaction results in new data being captured. This valuable customer data can be used to fuel generative AI and agentic AI tools that provide personalized recommendations and boost online sales. According to NVIDIA’s latest State of AI in Retail and Consumer-Packaged Goods report, 64% of respondents investing in AI for digital retail are prioritizing hyper-personalized recommendations.

Smart, Seamless and Personalized: The Future of Customer Experience

AI agents offer a range of benefits that significantly improve the retail customer experience, including:

  • Personalized Experiences: Using customer insights and product information, these digital assistants can deliver the expertise of a company’s best sales associate, stylist or designer — providing tailored product recommendations, enhancing decision-making, and boosting conversion rates and customer satisfaction.
  • Product Knowledge: AI agents enrich product catalogs with explanatory titles, enhanced descriptions and detailed attributes like size, warranty, sustainability and lifestyle uses. This makes products more discoverable and recommendations more personalized and informative, which increases consumer confidence.
  • Omnichannel Support: AI provides seamless integration of online and offline experiences, facilitating smooth transitions between digital and physical retail environments.
  • Virtual Try-On Capabilities: Customers can easily visualize products on themselves or in their homes in real time, helping improve product expectations and potentially lowering return rates.
  • 24/7 Availability: AI agents offer around-the-clock customer support across time zones and languages.

Real-World Applications of AI Agents in Retail

AI is redefining digital commerce, empowering retailers to deliver richer, more intuitive shopping experiences. From enhancing product catalogs with accurate, high-quality data to improving search relevance and offering personalized shopping assistance, AI agents are transforming how customers discover, engage with and purchase products online.

AI agents for catalog enrichment automatically enhance product information with consumer-focused attributes. These attributes can range from basic details like size, color and material to technical details such as warranty information and compatibility.

They also include contextual attributes, like sustainability, and lifestyle attributes, such as “for hiking.” AI agents can also integrate service attributes — including delivery times and return policies — making items more discoverable and relevant to customers while addressing common concerns to improve purchase results.

Amazon faced the challenge of ensuring complete and accurate product information for shoppers while reducing the effort and time required for sellers to create product listings. To address this, the company implemented generative AI using the NVIDIA TensorRT-LLM library. This technology allows sellers to input a product description or URL, and the system automatically generates a complete, enriched listing. The work helps sellers reach more customers and expand their businesses effectively while making the catalog more responsive and energy efficient.

AI agents for search tap into enriched data to deliver more accurate and contextually relevant search results. By employing semantic understanding and personalization, these agents better match customer queries with the right products, making the overall search experience faster and more intuitive.

Amazon Music has optimized its search capabilities using the Amazon SageMaker platform with NVIDIA Triton Inference Server and the NVIDIA TensorRT software development kit. This includes implementing vector search and transformer-based spell-correction models.

As a result, when users search for music — even with typos or vague terms — they can quickly find what they’re looking for. These optimizations, which make the search bar more effective and user friendly, have led to faster search times and 73% lower costs for Amazon Music.

AI agents for shopping assistants build on the enriched catalog and improved search functionality. They offer personalized recommendations and answer queries in a detailed, relevant, conversational manner, guiding shoppers through their buying journeys with a comprehensive understanding of products and user intent.

SoftServe, a leading IT advisor, has launched the SoftServe Gen AI Shopping Assistant, developed using the NVIDIA AI Blueprint for retail shopping assistants. SoftServe’s shopping assistant offers seamless and engaging shopping experiences by helping customers discover products and access detailed product information quickly and efficiently. One of its standout features is the virtual try-on capability, which allows customers to visualize how clothing and accessories look on them in real time.

Defining the Essential Traits of a Powerful AI Shopping Agent

Highly skilled AI shopping assistants are designed to be multimodal, understanding text- and image-based prompts, voice and more through large language models (LLMs) and vision language models. These AI agents can search for multiple items simultaneously, complete complicated tasks — such as creating a travel wardrobe — and answer contextual questions, like whether a product is waterproof or requires drycleaning.

This high level of sophistication offers experiences akin to engaging with a company’s best sales associate, delivering information to customers in a natural, intuitive way.

Diagram showing NVIDIA technologies used to build agentic AI applications, such as NVIDIA AI Blueprints (top), NVIDIA NeMo (middle) and NVIDIA NIM microservices (bottom).
With software building blocks, developers can design an AI agent with various features.

The building blocks of a powerful retail shopping agent include:

  • Multimodal and Multi-Query Capabilities: These agents can process and respond to queries that combine text and images, making search processes more versatile and user friendly. They can also easily be extended to support other modalities such as voice.
  • Integration With LLMs: Advanced LLMs, such as the NVIDIA Llama Nemotron family, bring reasoning capabilities to AI shopping assistants, enabling them to engage in natural, humanlike interactions. NVIDIA NIM microservices provide industry-standard application programming interfaces for simple integration into AI applications, development frameworks and workflows.
  • Management of Structured and Unstructured Data: NVIDIA NeMo Retriever microservices provide the ability to ingest, embed and understand retailers’ suites of relevant data sources, such as customer preferences and purchases, product catalog text and image data, and more, helping ensure AI agent responses are relevant, accurate and context-aware.
  • Guardrails for Brand Safe, On-Topic Conversations: NVIDIA NeMo Guardrails are implemented to help ensure that conversations with the shopping assistant remain safe and on topic, ultimately protecting brand values and bolstering customer trust.
  • State-of-the-Art Simulation Tools: The NVIDIA Omniverse platform and partner simulation technologies can help visualize products in physically accurate spaces. For example, customers looking to buy a couch could preview how the furniture would look in their own living room.

By using these key technologies, retailers can design AI shopping agents that exceed customer expectations, driving higher satisfaction and improved operational efficiency.

Retail organizations that harness AI agents are poised to experience evolving capabilities, such as enhanced predictive analytics for further personalized recommendations.

And integrating AI with augmented- and virtual-reality technologies is expected to create even more immersive and engaging shopping environments — delivering a future where shopping experiences are more immersive, convenient and customer-focused than ever.

Learn more about the AI Blueprint for retail shopping assistants.

 

Editor’s note: This post is part of the AI On blog series, which explores the latest techniques and real-world applications of agentic AI, chatbots and copilots. The series also highlights the NVIDIA software and hardware powering advanced AI agents, which form the foundation of AI query engines that gather insights and perform tasks to transform
Read Article

National Robotics Week — Latest Physical AI Research, Breakthroughs and Resources​on April 6, 2025 at 4:00 pm

This National Robotics Week, running through April 12, NVIDIA is highlighting the pioneering technologies that are shaping the future of intelligent machines and driving progress across manufacturing, healthcare, logistics and more. Check back here throughout the week to learn the latest on physical AI, which enables machines to perceive, plan and act with greater autonomy
Read ArticleThis National Robotics Week, running through April 12, NVIDIA is highlighting the pioneering technologies that are shaping the future of intelligent machines and driving progress across manufacturing, healthcare, logistics and more. Check back here throughout the week to learn the latest on physical AI, which enables machines to perceive, plan and act with greater autonomy
Read Article  

 

Check back here throughout the week to learn the latest on physical AI, which enables machines to perceive, plan and act with greater autonomy and intelligence in real-world environments.

This National Robotics Week, running through April 12, NVIDIA is highlighting the pioneering technologies that are shaping the future of intelligent machines and driving progress across manufacturing, healthcare, logistics and more.

Advancements in robotics simulation and robot learning are driving this fundamental shift in the industry. Plus, the emergence of world foundation models is accelerating the evolution of AI-enabled robots capable of adapting to dynamic and complex scenarios.

For example, by providing robot foundation models like NVIDIA GR00T N1, frameworks such as NVIDIA Isaac Sim and Isaac Lab for robot simulation and training, and synthetic data generation pipelines to help train robots for diverse tasks, the NVIDIA Isaac and GR00T platforms are empowering researchers and developers to push the boundaries of robotics.

Teaching Robots to Think: Nicklas Hansen’s AI Breakthroughs 🔗

What does it take to teach robots complex decision-making in the real world? For Nicklas Hansen, a doctoral candidate at UC San Diego and an NVIDIA Graduate Research Fellow, the answer lies in scalable, robust machine learning algorithms.

With experience from the University of California, Berkeley, Meta AI (FAIR) and the Technical University of Denmark, Hansen is pushing the boundaries of how robots perceive, plan and act in dynamic environments. Their research sits at the intersection of robotics, reinforcement learning and computer vision — bridging the gap between simulation and real-world deployment.

Nicklas Hansen, a doctoral candidate at UC San Diego and an NVIDIA Graduate Research Fellow.

Hansen’s recent work tackles one of robotics’ toughest challenges: long-horizon manipulation. Their paper, Multi-Stage Manipulation With Demonstration-Augmented Reward, Policy and World Model Learning, introduces a framework that enhances data efficiency in sparse-reward environments by using multistage task structures.

Left: A simulated Franka robot solves a peg insertion manipulation task. Right: Hansen’s method, DEMO3, infers task progress directly from raw visual observations.

Another key project of Hansen’s, Hierarchical World Models as Visual Whole-Body Humanoid Controllers, advances control strategies for humanoid robots, enabling more adaptive and humanlike movements.

Beyond their own research, Hansen advocates for making AI-driven robotics more accessible.

“My advice to anyone looking to get started with AI for robotics is to simply play around with the many open-source tools available and gradually start contributing to projects that align with your goals and interests,” they said. “With the availability of free simulation tools like MuJoCo, NVIDIA Isaac Lab and ManiSkill, you can make a profound impact on the field without owning a real robot.”

Hansen is the lead author of TD-MPC2, a model-based reinforcement learning algorithm capable of learning a variety of control tasks without any domain knowledge. The algorithm is open source and can be run on a single consumer-grade GPU.

Learn more about Hansen and other NVIDIA Graduate Fellowship recipients driving innovation in AI and robotics. Watch a replay of the “Graduate Program Fast Forward” session from the NVIDIA GTC AI conference, where doctoral students in the NVIDIA Graduate Fellowship showcased their groundbreaking research.

Hackathon Features Robots Powered by NVIDIA Isaac GR00T N1 🔗

The Seeed Studio Embodied AI Hackathon, which took place last month, brought together the robotics community to showcase innovative projects using the LeRobot SO-100ARM motor kit.

The event highlighted how robot learning is advancing AI-driven robotics, with teams successfully integrating the NVIDIA Isaac GR00T N1 model to speed humanoid robot development. A notable project involved developing leader-follower robot pairs capable of learning pick-and-place tasks by post-training robot foundation models on real-world demonstration data.

How the project worked:

  • Real-World Imitation Learning: Robots observe and mimic human-led demonstrations, recorded through Arducam vision systems and an external camera.
  • Post-Training Pipeline: Captured data is structured into a modality.json dataset for efficient GPU-based training with GR00T N1.
  • Bimanual Manipulation: The model is optimized for controlling two robotic arms simultaneously, enhancing cooperative skills.

The dataset is now publicly available on Hugging Face, with implementation details on GitHub.

Team “Firebreathing Rubber Duckies” celebrating with NVIDIA hosts.

Learn more about the project.

Advancing Robotics: IEEE Robotics and Automation Society Honors Emerging Innovators 🔗

The IEEE Robotics and Automation Society in March announced the recipients of its 2025 Early Academic Career Award, recognizing outstanding contributions to the fields of robotics and automation.

This year’s honorees — including NVIDIA’s Shuran Song, Abhishek Gupta and Yuke Zhu — are pioneering advancements in scalable robot learning, real-world reinforcement learning and embodied AI. Their work is shaping the next generation of intelligent systems, driving innovation that impacts both research and real-world applications.

Learn more about the award winners:

These researchers will be recognized at the International Conference on Robotics and Automation in May.

Stay up to date on NVIDIA’s leading robotics research through the Robotics Research and Development Digest (R2D2) tech blog series, subscribing to this newsletter and following NVIDIA Robotics on YouTube, Discord and developer forums.

 

This National Robotics Week, running through April 12, NVIDIA is highlighting the pioneering technologies that are shaping the future of intelligent machines and driving progress across manufacturing, healthcare, logistics and more. Check back here throughout the week to learn the latest on physical AI, which enables machines to perceive, plan and act with greater autonomy
Read Article

NVIDIA Brings Agentic AI Reasoning to Enterprises With Google Cloud​on April 9, 2025 at 12:00 pm

NVIDIA is collaborating with Google Cloud to bring agentic AI to enterprises seeking to locally harness the Google Gemini family of AI models using the NVIDIA Blackwell HGX and DGX platforms and NVIDIA Confidential Computing for data safety. With the NVIDIA Blackwell platform on Google Distributed Cloud, on-premises data centers can stay aligned with regulatory
Read ArticleNVIDIA is collaborating with Google Cloud to bring agentic AI to enterprises seeking to locally harness the Google Gemini family of AI models using the NVIDIA Blackwell HGX and DGX platforms and NVIDIA Confidential Computing for data safety. With the NVIDIA Blackwell platform on Google Distributed Cloud, on-premises data centers can stay aligned with regulatory
Read Article  

 

NVIDIA is collaborating with Google Cloud to bring agentic AI to enterprises seeking to locally harness the Google Gemini family of AI models using the NVIDIA Blackwell HGX and DGX platforms and NVIDIA Confidential Computing for data safety.

With the NVIDIA Blackwell platform on Google Distributed Cloud, on-premises data centers can stay aligned with regulatory requirements and data sovereignty laws by locking down access to sensitive information, such as patient records, financial transactions and classified government information. NVIDIA Confidential Computing also secures sensitive code in the Gemini models from unauthorized access and data leaks.

“By bringing our Gemini models on premises with NVIDIA Blackwell’s breakthrough performance and confidential computing capabilities, we’re enabling enterprises to unlock the full potential of agentic AI,” said Sachin Gupta, vice president and general manager of infrastructure and solutions at Google Cloud. “This collaboration helps ensure customers can innovate securely without compromising on performance or operational ease.”

Confidential computing with NVIDIA Blackwell provides enterprises with the technical assurance that their user prompts to the Gemini models’ application programming interface — as well as the data they used for fine-tuning — remain secure and cannot be viewed or modified.

At the same time, model owners can protect against unauthorized access or tampering, providing dual-layer protection that enables enterprises to innovate with Gemini models while maintaining data privacy.

AI Agents Driving New Enterprise Applications

This new offering arrives as agentic AI is transforming enterprise technology, offering more advanced problem-solving capabilities.

Unlike AI models that perceive or generate based on learned knowledge, agentic AI systems can reason, adapt and make decisions in dynamic environments. For example, in enterprise IT support, while a knowledge-based AI model can retrieve and present troubleshooting guides, an agentic AI system can diagnose issues, execute fixes and escalate complex problems autonomously.

Similarly, in finance, a traditional AI model could flag potentially fraudulent transactions based on patterns, but an agentic AI system could go even further by investigating anomalies and taking proactive measures such as blocking transactions before they occur or adjusting fraud detection rules in real time.

The On-Premises Dilemma

While many can already use the models with multimodal reasoning — integrating text, images, code and other data types to solve complex problems and build cloud-based agentic AI applications — those with stringent security or data sovereignty requirements have yet been unable to do so.

With this announcement, Google Cloud will be one of the first cloud service providers to offer confidential computing capabilities to secure agentic AI workloads across every environment — whether cloud or hybrid.

Powered by the NVIDIA HGX B200 platform with Blackwell GPUs and NVIDIA Confidential Computing, this solution will enable customers to safeguard AI models and data. This lets users achieve breakthrough performance and energy efficiency without compromising data security or model integrity.

AI Observability and Security for Agentic AI

Scaling agentic AI in production requires robust observability and security to ensure reliable performance and compliance.

Google Cloud today announced a new GKE Inference Gateway built to optimize the deployment of AI inference workloads with advanced routing and scalability. Integrating with NVIDIA Triton Inference Server and NVIDIA NeMo Guardrails, it offers intelligent load balancing that improves performance and reduces serving costs while enabling centralized model security and governance.

Looking ahead, Google Cloud is working to enhance observability for agentic AI workloads by integrating NVIDIA Dynamo, an open-source library built to serve and scale reasoning AI models across AI factories.

At Google Cloud Next, attend NVIDIA’s special address, explore sessions, view demos and talk to NVIDIA experts.

 

NVIDIA is collaborating with Google Cloud to bring agentic AI to enterprises seeking to locally harness the Google Gemini family of AI models using the NVIDIA Blackwell HGX and DGX platforms and NVIDIA Confidential Computing for data safety. With the NVIDIA Blackwell platform on Google Distributed Cloud, on-premises data centers can stay aligned with regulatory
Read Article

‘Black Women in Artificial Intelligence’ Founder Talks AI Education and Empowerment​on April 9, 2025 at 4:00 pm

Necessity is the mother of invention. And sometimes, what a person really needs is hot chocolate served to them by a robot — one named after a pop star, ideally. Angle Bush, founder and CEO of Black Women in Artificial Intelligence (BWIAI), began her AI journey in 2019 with the idea to build a robot
Read ArticleNecessity is the mother of invention. And sometimes, what a person really needs is hot chocolate served to them by a robot — one named after a pop star, ideally. Angle Bush, founder and CEO of Black Women in Artificial Intelligence (BWIAI), began her AI journey in 2019 with the idea to build a robot
Read Article  

 

Necessity is the mother of invention. And sometimes, what a person really needs is hot chocolate served to them by a robot — one named after a pop star, ideally.

Angle Bush, founder and CEO of Black Women in Artificial Intelligence (BWIAI), began her AI journey in 2019 with the idea to build a robot named Usher that could bring her cocoa. As she scoured robotics tutorial videos for ways to bring her vision to life, Bush found herself captivated by something even bigger: artificial intelligence.

“As I’m doing this research, I’m finding more about artificial intelligence, and I’m hearing it’s the fourth industrial revolution,” she said.

But when Angle started attending AI events, a lack of diverse representation became glaringly obvious to her.

“I wasn’t quite seeing a full reflection of myself,” she said. “Surely you can’t have a revolution without Black women.”

From this realization, BWIAI was born.

Bush joined the NVIDIA AI Podcast to share more about the organization’s mission to reshape the AI community by educating, engaging, embracing and empowering Black women in the field.

Not five years after its founding, BWIAI brings together members from five continents and collaborates with key industry leaders and partners — serving as a supportive community and catalyst of change.

BWIAI and its partners offer hands-on learning experiences and online resources to its member community. They also launched a career assessment agent to help members explore how their interests align with emerging career paths in AI, as well as technologies and coursework for getting started.

“We have people in television, we have university professors, we have lawyers, we have doctors,” Bush said. “It runs the gamut because they are an example of what’s happening globally. Every industry is impacted by AI.”

Time Stamps

2:45 – Bush discusses BWIAI’s partnerships and initiatives, including its autonomous hair-braiding machine.

6:30 – The importance of educating, engaging, embracing and empowering Black women in AI.

10:40 – Behind BWIAI’s AI career assessment agent.

12:10 – Bush explains how removing barriers increases innovation.

You Might Also Like… 

NVIDIA’s Louis Stewart on How AI Is Shaping Workforce Development

Louis Stewart, head of strategic initiatives for NVIDIA’s global developer ecosystem, discusses why workforce development is crucial for maximizing AI benefits. He emphasizes the importance of AI education, inclusivity and public-private partnerships in preparing the global workforce for the future. Engaging with AI tools and understanding their impact on the workforce landscape is vital to ensuring these changes benefit everyone.

Tara Chklovksi, Anshita Saini on Technovation Pioneering AI Education for Innovation

Tara Chklovski, founder and CEO of Technovation, returns to discuss the importance of inclusive AI. With Anshita Saini, a Technovation alumna and OpenAI staff member, Chklovski explores how Technovation empowers girls through AI education and enhances real-world problem-solving skills. Saini shares her journey from creating an app that helped combat a vaping crisis at her high school to taking on her current role at OpenAI. She also introduces Wiser AI, an initiative she founded to support women and underrepresented voices in AI.

Imbue CEO Kanjun Qiu on Transforming AI Agents Into Personal Collaborators

Kanjun Qiu, CEO of Imbue, explores the emerging era where individuals can create and use their own AI agents. Drawing a parallel to the personal computer revolution of the late 1970s and ‘80s, Qiu discusses how modern AI systems are evolving to work collaboratively with users, enhancing their capabilities rather than just automating tasks.

 

Necessity is the mother of invention. And sometimes, what a person really needs is hot chocolate served to them by a robot — one named after a pop star, ideally. Angle Bush, founder and CEO of Black Women in Artificial Intelligence (BWIAI), began her AI journey in 2019 with the idea to build a robot
Read Article