NVIDIA Accelerated Quantum Research Center to Bring Quantum Computing Closer

NVIDIA Accelerated Quantum Research Center to Bring Quantum Computing Closer

As quantum computers continue to develop, they will integrate with AI supercomputers to form accelerated quantum supercomputers capable of solving some of the world’s hardest problems.

Integrating quantum processing units (QPUs) into AI supercomputers is key for developing new applications, helping unlock breakthroughs critical to running future quantum hardware and enabling developments in quantum error correction and device control.

The NVIDIA Accelerated Quantum Research Center, or NVAQC, announced today at the NVIDIA GTC global AI conference, is where these developments will happen. With an NVIDIA GB200 NVL72 system and the NVIDIA Quantum-2 InfiniBand networking platform, the facility will house a supercomputer with 576 NVIDIA Blackwell GPUs dedicated to quantum computing research.

“The NVAQC draws on much-needed and long-sought-after tools for scaling quantum computing to next-generation devices,” said Tim Costa, senior director of computer-aided engineering, quantum and CUDA-X at NVIDIA. “The center will be a place for large-scale simulations of quantum algorithms and hardware, tight integration of quantum processors, and both training and deployment of AI models for quantum.”

The NVAQC will host a GB200 NVL72 system.
The NVAQC will host a GB200 NVL72 system.

Quantum computing innovators like Quantinuum, QuEra and Quantum Machines, along with academic partners from the Harvard Quantum Initiative and the Engineering Quantum Systems group at the MIT Center for Quantum Engineering, will work on projects with NVIDIA at the center to explore how AI supercomputing can accelerate the path toward quantum computing.

“The NVAQC is a powerful tool that will be instrumental in ushering in the next generation of research across the entire quantum ecosystem,” said William Oliver, professor of electrical engineering and computer science, and of physics, leader of the EQuS group and director of the MIT Center for Quantum Engineering. “NVIDIA is a critical partner for realizing useful quantum computing.”

There are several key quantum computing challenges where the NVAQC is already set to have a dramatic impact.

Protecting Qubits With AI Supercomputing

Qubit interactions are a double-edged sword. While qubits must interact with their surroundings to be controlled and measured, these same interactions are also a source of noise — unwanted disturbances that affect the accuracy of quantum calculations. Quantum algorithms can only work if the resulting noise is kept in check.

Quantum error correction provides a solution, encoding noiseless, logical qubits within many noisy, physical qubits. By processing the outputs from repeated measurements on these noisy qubits, it’s possible to identify, track and correct qubit errors — all without destroying the delicate quantum information needed by a computation.

The process of figuring out where errors occurred and what corrections to apply is called decoding. Decoding is an extremely difficult task that must be performed by a conventional computer within a narrow time frame to prevent noise from snowballing out of control.

A key goal of the NVAQC will be exploring how AI supercomputing can accelerate decoding. Studying how to collocate quantum hardware within the center will allow the development of low-latency, parallelized and AI-enhanced decoders, running on NVIDIA GB200 Grace Blackwell Superchips.

The NVAQC will also tackle other challenges in quantum error correction. QuEra will work with NVIDIA to accelerate its search for new, improved quantum error correction codes, assessing the performance of candidate codes through demanding simulations of complex quantum circuits.

“The NVAQC will be an essential tool for discovering, testing and refining new quantum error correction codes and decoders capable of bringing the whole industry closer to useful quantum computing,” said Mikhail Lukin, Joshua and Beth Friedman University Professor at Harvard and a codirector of the Harvard Quantum Initiative.

Developing Applications for Accelerated Quantum Supercomputers

The majority of useful quantum algorithms draw equally from classical and quantum computing resources, ultimately requiring an accelerated quantum supercomputer that unifies both kinds of hardware.

For example, the output of classical supercomputers is often needed to prime quantum computations. The NVAQC provides the heterogeneous compute infrastructure needed for research on developing and improving such hybrid algorithms.

A diagram of an accelerated quantum supercomputer connecting classical and quantum processors.
Accelerated quantum supercomputers will connect quantum and classical processors to execute hybrid algorithms.

New AI-based compilation techniques will also be explored at the NVAQC, with the potential to accelerate the runtime of all quantum algorithms, including through work with Quantinuum. Quantinuum will build on its previous integration work with NVIDIA, offering its hardware and emulators through the NVIDIA CUDA-Q platform. Users of CUDA-Q are currently offered unrestricted access to Quantinuum’s QNTM H1-1 hardware and emulator for 90 days.

“We’re excited to deepen our work with NVIDIA via this center,” said Rajeeb Hazra, president and CEO of Quantinuum. “By combining Quantinuum’s powerful quantum systems with NVIDIA’s cutting-edge accelerated computing, we’re pushing the boundaries of hybrid quantum-classical computing and unlocking exciting new possibilities.”

QPU Integration

Integrating quantum hardware with AI supercomputing is the one of the major remaining hurdles on the path to running useful quantum hardware.

The requirements of such an integration can be extremely demanding. The decoding required by quantum error correction can only function if data from millions of qubits can be sent between quantum and classical hardware at ultralow latencies.

Quantum Machines will work with NVIDIA at the NVAQC to develop and hone new controller technologies supporting rapid, high-bandwidth interfaces between quantum processors and GB200 superchips.

“We’re excited to see NVIDIA’s growing commitment to accelerating the realization of useful quantum computers, providing researchers with the most advanced infrastructure to push the boundaries of quantum-classical computing,” said Itamar Sivan, CEO of Quantum Machines.

Depiction of the NVIDIA DGX Quantum system, which comprises an NVIDIA GH200 superchip coupled with Quantum Machines’ OPX1000 control system.
The NVIDIA DGX Quantum system comprises an NVIDIA GH200 superchip and Quantum Machines’ OPX1000 control system.

Key to integrating quantum and classical hardware is a platform that lets researchers and developers quickly shift context between these two disparate computing paradigms within a single application. The NVIDIA CUDA-Q platform will be the entry point for researchers to harness the NVAQC’s quantum-classical integration.

Building on tools like NVIDIA DGX Quantum — a reference architecture for integrating quantum and classical hardware — and CUDA-Q, the NVAQC is set to be an epicenter for next-generation developments in quantum computing, seeding the evolution of qubits into impactful quantum computers.

Learn more about NVIDIA quantum computing.

Read More

Full Steam Ahead: NVIDIA-Certified Program Expands to Enterprise Storage for Faster AI Factory Deployment

Full Steam Ahead: NVIDIA-Certified Program Expands to Enterprise Storage for Faster AI Factory Deployment

AI deployments thrive on speed, data and scale. That’s why NVIDIA is expanding NVIDIA-Certified Systems to include enterprise storage certification — for streamlined AI factory deployments in the enterprise with accelerated computing, networking, software and storage.

As enterprises build AI factories, access to high-quality data is imperative to ensure optimal performance and reliability for AI models. The new NVIDIA-Certified Storage program announced today at the NVIDIA GTC global AI conference validates that enterprise storage systems meet stringent performance and scalability data requirements for AI and high-performance computing workloads.

Leading enterprise data platform and storage providers are already onboard, ensuring businesses have trusted options from day one. These include DDN, Dell Technologies, Hewlett Packard Enterprise, Hitachi Vantara, IBM, NetApp, Nutanix, Pure Storage, VAST Data and WEKA.

Building Blocks for a New Class of Enterprise Infrastructure

At GTC, NVIDIA also announced the NVIDIA AI Data Platform, a customizable reference design to build a new class of enterprise infrastructure for demanding agentic AI workloads.

The NVIDIA-Certified Storage designation is a prerequisite for partners developing agentic AI infrastructure solutions built on the NVIDIA AI Data Platform. Each of these NVIDIA-Certified Storage partners will deliver customized AI data platforms, in collaboration with NVIDIA, that can harness enterprise data to reason and respond to complex queries.

NVIDIA-Certified was created more than four years ago as the industry’s first certification program dedicated to tuning and optimizing AI systems to ensure optimal performance, manageability and scalability. Each NVIDIA-Certified system is rigorously tested and validated to deliver enterprise-grade AI performance.

There are now 50+ partners providing 500+ NVIDIA-Certified systems, helping enterprises reduce time, cost and complexity by giving them a wide selection of performance-optimized systems to power their accelerated computing workloads.

NVIDIA Enterprise Reference Architectures (RAs) were introduced last fall to provide partners with AI infrastructure best practices and configuration guidance for deploying NVIDIA-Certified servers, NVIDIA Spectrum-X networking and NVIDIA AI Enterprise software.

Solutions based on NVIDIA Enterprise RAs are available from the world’s leading systems providers to reduce the time, cost and complexity of enterprise AI deployments. Enterprise RAs are now available for a wide range of NVIDIA Hopper and NVIDIA Blackwell platforms, including NVIDIA HGX B200 systems and the new NVIDIA RTX PRO 6000 Blackwell Server Edition GPU.

These NVIDIA technologies and partner solutions are the building blocks for enterprise AI factories, representing a new class of enterprise infrastructure for high-performance AI deployments at scale.

Enterprise AI Needs Scalable Storage

As the pace of AI innovation and adoption accelerates, secure and reliable access to high-quality enterprise data is becoming more important than ever. Data is the fuel for the AI factory. With enterprise data creation projected to reach 317 zettabytes annually by 2028*, AI workloads require storage architectures built to handle massive, unstructured and multimodal datasets.

NVIDIA’s expanded storage certification program is designed to meet this need and help enterprises build AI factories with a foundation of high-performance, reliable data storage solutions. The program includes performance testing as well as  validation that partner storage systems adhere to design best practices, optimizing performance and scalability for enterprise AI workloads.

NVIDIA-Certified Storage will be incorporated into NVIDIA Enterprise RAs, providing enterprise-grade data storage for AI factory deployments with full-stack solutions from global systems partners.

Certified Storage for Every Deployment

This certification builds on existing NVIDIA DGX systems and NVIDIA Cloud Partner (NCP) storage programs, expanding the data ecosystem for AI infrastructure.

These storage certification programs are aligned with their deployment models and architectures:

  • NVIDIA DGX BasePOD and DGX SuperPOD Storage Certification — designed for enterprise AI factory deployments with NVIDIA DGX systems.
  • NCP Storage Certification — designed for large-scale NCP Reference Architecture AI factory deployments with cloud providers.
  • NVIDIA-Certified Storage Certification — designed for enterprise AI factory deployments with NVIDIA-Certified servers available from global partners, based on NVIDIA Enterprise RA guidelines.

With this framework, organizations of all sizes — from cloud hyperscalers to enterprises — can build AI factories that process massive amounts of data, train models faster and drive more accurate, reliable AI outcomes.

Learn more about how NVIDIA-Certified Systems deliver seamless, high-speed performance and attend these related sessions at GTC:

*Source: IDC, Worldwide IDC Global DataSphere Forecast, 2024–2028: AI Everywhere, But Upsurge in Data Will Take Time, doc #US52076424, May 2024

Read More

From AT&T to the United Nations, AI Agents Redefine Work With NVIDIA AI Enterprise

From AT&T to the United Nations, AI Agents Redefine Work With NVIDIA AI Enterprise

AI agents are transforming work, delivering time and cost savings by helping people resolve complex challenges in new ways.

Whether developed for humanitarian aid, customer service or healthcare, AI agents built with the NVIDIA AI Enterprise software platform make up a new digital workforce helping professionals accomplish their goals faster — at lower costs and for greater impact.

AI Agents Enable Growth and Education

AI can instantly translate, summarize and process multimodal content in hundreds of languages. Integrated into agentic systems, the technology enables international organizations to engage and educate global stakeholders more efficiently.

The United Nations (UN) is working with Accenture to develop a multilingual research agent to support over 150 languages to promote local economic sustainability. The agent will act like a researcher, answering questions about the UN’s Sustainable Development Goals and fostering awareness and engagement toward its agenda of global peace and prosperity.

Mercy Corps, in collaboration with Cloudera, has deployed an AI-driven Methods Matcher tool that supports humanitarian aid experts in more than 40 countries by providing research, summaries, best-practice guidelines and data-driven crisis responses, providing faster aid delivery in disaster situations.

Wikimedia Deutschland, using the DataStax AI Platform, built with NVIDIA AI, can process and embed 10 million Wikidata items in just three days, with 30x faster ingestion performance.

AI Agents Provide Tailored Customer Service Across Industries

Agentic AI enhances customer service with real-time, highly accurate insights for more effective user experiences. AI agents provide 24/7 support, handling common inquiries with more personalized responses while freeing human agents to address more complex issues.

Intelligent-routing capabilities categorize and prioritize requests so customers can be quickly directed to the right specialists. Plus, AI agents’ predictive-analytics capabilities enable proactive support by anticipating issues and empowering human agents with data-driven insights.

Companies across industries including telecommunications, finance, healthcare and sports are already tapping into AI agents to achieve massive benefits.

AT&T, in collaboration with Quantiphi, developed and deployed a new Ask AT&T AI agent to its call center, leading to a 84% decrease in call center analytics costs.

Southern California Edison, working with WWT, is driving Project Orca to enhance data processing and predictions for 100,000+ network assets using agents to reduce downtime, enhance network reliability and enable faster, more efficient ticket resolution.

With the adoption of ServiceNow Dispute Management, built with Visa, banks can use AI agents with the solution to achieve up to a 28% reduction in call center volumes and a 30% decrease in time to resolution.

The Ottawa Hospital, working with Deloitte, deployed a team of 24/7 patient-care agents to provide preoperative support and answer patient questions regarding upcoming procedures for over 1.2 million people in eastern Ontario, Canada.

With the VAST Data Platform, the National Hockey League can unlock over 550,000 hours of historical game footage. This supports sponsorship analysis, helps video producers quickly create broadcast clips and enhances personalized fan content.

State-of-the-Art AI Agents Built With NVIDIA AI Enterprise

AI agents have emerged as versatile tools that can be adapted and adopted across a wide range of industries. These agents connect to organizational knowledge bases to understand the business context they’re deployed in. Their core functionalities — such as question-answering, translation, data processing, predictive analytics and automation — can be tailored to improve productivity and save time and costs, by any organization, in any industry.

NVIDIA AI Enterprise provides the building blocks for enterprise AI agents. It includes NVIDIA NIM microservices for efficient inference of state-of-the-art models — including the new NVIDIA Llama Nemotron reasoning model family — and NVIDIA NeMo tools to streamline data processing, model customization, system evaluation, retrieval-augmented generation and guardrailing.

NVIDIA Blueprints are reference workflows that showcase best practices for developing high-performance agentic systems. With the AI-Q NVIDIA AI Blueprint, developers can build AI agents into larger agentic systems that can reason, then connect these systems to enterprise data to tackle complex problems, harness other tools, collaborate and operate with greater autonomy.

Learn more about AI agent development by watching the NVIDIA GTC keynote and register for sessions from NVIDIA and industry leaders at the show, which runs through March 21.

See notice regarding software product information.

Read More

NVIDIA Aerial Expands With New Tools for Building AI-Native Wireless Networks

NVIDIA Aerial Expands With New Tools for Building AI-Native Wireless Networks

The telecom industry is increasingly embracing AI to deliver seamless connections — even in conditions of poor signal strength — while maximizing sustainability and spectral efficiency, the amount of information that can be transmitted per unit of bandwidth.

Advancements in AI-RAN technology have set the course toward AI-native wireless networks for 6G, built using AI and accelerated computing from the start, to meet the demands of billions of AI-enabled connected devices, sensors, robots, cameras and autonomous vehicles.

To help developers and telecom leaders pioneer these networks, NVIDIA today unveiled new tools in the NVIDIA Aerial Research portfolio.

The expanded portfolio of solutions include the Aerial Omniverse Digital Twin on NVIDIA DGX Cloud, the Aerial Commercial Test Bed on NVIDIA MGX, the NVIDIA Sionna 1.0 open-source library and the Sionna Research Kit on NVIDIA Jetson — helping accelerate AI-RAN and 6G research.

Industry leaders like Amdocs, Ansys, Capgemini, DeepSig, Fujitsu, Keysight, Kyocera, MathWorks, Mediatek, Samsung Research, SoftBank and VIAVI Solutions and more than 150 higher education and research institutions from U.S. and around the world — including Northeastern University, Rice University, The University of Texas at Austin, ETH Zurich, Fraunhofer Institute for Telecommunications, Heinrich-Hertz-Institut, HHI, Singapore University of Technology and Design, and University of Oulu — are harnessing the NVIDIA Aerial Research portfolio to develop, train, simulate and deploy groundbreaking AI-native wireless innovations.

New Tools for Research and Development

The Aerial Research portfolio provides exceptional flexibility and ease of use for developers at every stage of their research — from early experimentation to commercial deployment. Its offerings include:

  • Aerial Omniverse Digital Twin (AODT): A simulation platform to test and fine-tune algorithms in physically precise digital replicas of entire wireless systems, now available on NVIDIA DGX Cloud. Developers can now access AODT everywhere, whether on premises, on laptops, via the public cloud or on an NVIDIA cloud service.
  • Aerial Commercial Test Bed (aka ARC-OTA): A full-stack AI-RAN deployment system that enables developers to deploy new AI models over the air and test them in real time, now available on NVIDIA MGX and available through manufacturers including Supermicro or as a managed offering via Sterling Skywave. ARC-OTA integrates commercial-grade Aerial CUDA-accelerated RAN software with open-source L2+ and 5G core from OpenAirInterface (OAI) and O-RAN-compliant 7.2 split open radio units from WNC and LITEON Technology to enable an end-to-end system for AI-RAN commercial testing.
  • Sionna 1.0: The most widely used GPU-accelerated open-source library for research in communication systems, with more than 135,000 downloads. The latest release of Sionna features a lightning-fast ray tracer for radio propagation, a versatile link-level simulator and new system-level simulation capabilities.
  • Sionna Research Kit: Powered by the NVIDIA Jetson platform, it integrates accelerated computing for AI and machine learning workloads and a software-defined RAN built on OAI. With the kit, researchers can connect 5G equipment and begin prototyping AI-RAN algorithms for next-generation wireless networks in just a few hours.

NVIDIA Aerial Research Ecosystem for AI-RAN and 6G

The NVIDIA Aerial Research portfolio includes the NVIDIA 6G Developer Program, an open community that serves more than 2,000 members, representing leading technology companies, academia, research institutions and telecom operators using NVIDIA technologies to complement their AI-RAN and 6G research.

Testing and simulation will play an essential role in developing AI-native wireless networks. Companies such as Amdocs, Ansys, Keysight, MathWorks and VIAVI are enhancing their simulation solutions with NVIDIA AODT, while operators have created digital twins of their radio access networks to optimize performance with changing traffic scenarios.

Nine out of 10 demonstrations chosen by the AI-RAN Alliance for Mobile World Congress were developed using the NVIDIA Aerial Research portfolio, leading to breakthrough results.

SoftBank and Fujitsu demonstrated an up to 50% throughput gain in poor radio environments using AI-based uplink channel interpolation.

DeepSig developed OmniPHY, an AI-native air interface that eliminates traditional pilot overhead, harnessing neural networks to achieve up to 70% throughput gains in certain scenarios. Using the NVIDIA AI Aerial platform, OmniPHY integrates machine learning into modulation, reception and demodulation to optimize spectral efficiency, reduce power consumption and enhance wireless network performance.

“AI-native signal processing is transforming wireless networks, delivering real-world results,” said Jim Shea, cofounder and CEO of DeepSig. “By integrating deep learning to the air interface and leveraging NVIDIA’s tools, we’re redefining how AI-native wireless networks are designed and built.”

In addition to the Aerial Research portfolio, using the open ecosystem of NVIDIA CUDA-X libraries, built on CUDA, developers can build applications that deliver dramatically higher performance.

Join the NVIDIA 6G Developer Program to access NVIDIA Aerial Research platform tools.

See notice regarding software product information.

Read More

Telecom Leaders Call Up Agentic AI to Improve Network Operations

Telecom Leaders Call Up Agentic AI to Improve Network Operations

Global telecommunications networks can support millions of user connections per day, generating more than 3,800 terabytes of data per minute on average.

That massive, continuous flow of data generated by base stations, routers, switches and data centers — including network traffic information, performance metrics, configuration and topology — is unstructured and complex. Not surprisingly, traditional automation tools have often fallen short on handling massive, real-time workloads involving such data.

To help address this challenge, NVIDIA today announced at the GTC global AI conference that its partners are developing new large telco models (LTMs) and AI agents custom-built for the telco industry using NVIDIA NIM and NeMo microservices within the NVIDIA AI Enterprise software platform. These LTMs and AI agents enable the next generation of AI in network operations.

LTMs — customized, multimodal large language models (LLMs) trained specifically on telco network data — are core elements in the development of network AI agents, which automate complex decision-making workflows, improve operational efficiency, boost employee productivity and enhance network performance.

SoftBank and Tech Mahindra have built new LTMs and AI agents, while Amdocs, BubbleRAN and ServiceNow, are dialing up their network operations and optimization with new AI agents, all using NVIDIA AI Enterprise.

It’s important work at a time when 40% of respondents in a recent NVIDIA-run telecom survey noted they’re deploying AI into their network planning and operations.

LTMs Understand the Language of Networks

Just as LLMs understand and generate human language, and NVIDIA BioNeMo NIM microservices understand the language of biological data for drug discovery, LTMs now enable AI agents to master the language of telecom networks.

The new partner-developed LTMs powered by NVIDIA AI Enterprise are:

  • Specialized in network intelligence — the LTMs can understand real-time network events, predict failures and automate resolutions.
  • Optimized for telco workloads — tapping into NVIDIA NIM microservices, the LTMs are optimized for efficiency, accuracy and low latency.
  • Suited for continuous learning and adaptation — with post-training scalability, the LTMs can use NVIDIA NeMo to learn from new events, alerts and anomalies to enhance future performance.

NVIDIA AI Enterprise provides additional tools and blueprints to build AI agents that simplify network operations and deliver cost savings and operational efficiency, while improving network key performance indicators (KPIs), such as:

  • Reduced downtime — AI agents can predict failures before they happen, delivering network resilience.
  • Improved customer experiences — AI-driven optimizations lead to faster networks, fewer outages and seamless connectivity.
  • Enhanced security — as it continuously scans for threats, AI can help mitigate cyber risks in real time.

Industry Leaders Launch LTMs and AI Agents

Leading companies across telecommunications are using NVIDIA AI Enterprise to advance their latest technologies.

SoftBank has developed a new LTM based on a large-scale LLM base model, trained on its own network data. Initially focused on network configuration, the model — which is available as an NVIDIA NIM microservice — can automatically reconfigure the network to adapt to changes in network traffic, including during mass events at stadiums and other venues. SoftBank is also introducing network agent blueprints to help accelerate AI adoption across telco operations.

Tech Mahindra has developed an LTM with the NVIDIA agentic AI tools to help address critical network operations. Tapping into this LTM, the company’s Adaptive Network Insights Studio provides a 360-degree view of network issues, generating automated reports at various levels of detail to inform and assist IT teams, network engineers and company executives.

In addition, Tech Mahindra’s Proactive Network Anomaly Resolution Hub is powered by the LTM to automatically resolve a significant portion of its network events, lightening engineers’ workloads and enhancing their productivity.

Amdocs’ Network Assurance Agent, powered by amAIz Agents, automates repetitive tasks such as fault prediction. It also conducts impact analysis and prevention methods for network issues, providing step-by-step guidance on resolving any problems that occur. Plus, the company’s Network Deployment Agent simplifies open radio access network (RAN) adoption by automating integration, deployment tasks and interoperability testing, and providing insights to network engineers.

BubbleRAN is developing an autonomous multi-agent RAN intelligence platform on a cloud-native infrastructure, where LTMs can observe the network state, configuration, availability and KPIs to facilitate monitoring and troubleshooting. The platform also automates the process of network reconfiguration and policy enforcement through a high-level set of action tools. The company’s AI agents satisfy user needs by tapping into advanced retrieval-augmented generation pipelines and telco-specific application programming interfaces, answering real-time, 5G deployment-specific questions.

ServiceNow’s AI agents in telecom — built with NVIDIA AI Enterprise on NVIDIA DGX Cloud — drive productivity by generating resolution playbooks and predicting potential network disruptions before they occur. This helps communications service providers reduce resolution time and improve customer satisfaction. The new, ready-to-use AI agents also analyze network incidents, identifying root causes of disruptions so they can be resolved faster and avoided in the future.

Learn more about the latest agentic AI advancements at NVIDIA GTC, running through Friday, March 21, in San Jose, California.

Read More

AI on the Menu: Yum! Brands and NVIDIA Partner to Accelerate Restaurant Industry Innovation

AI on the Menu: Yum! Brands and NVIDIA Partner to Accelerate Restaurant Industry Innovation

The quick-service restaurant industry is a marvel of modern logistics, where speed, teamwork and kitchen operations are key ingredients for every order. Yum! Brands is now introducing AI-powered agents at select Pizza Hut and Taco Bell locations to assist and enhance the team member experience.

Today at the NVIDIA GTC conference, Yum! Brands announced a strategic partnership with NVIDIA with a goal of deploying multiple AI solutions using NVIDIA technology in 500 restaurants this year.

World’s Largest Restaurant Company Advances AI Adoption

Spanning more than 61,000 locations, Yum! operates more restaurants than any other company in the world. Globally, customers are drawn to the food, value, service and digital convenience from iconic brands like KFC, Taco Bell, Pizza Hut and Habit Burger & Grill.

Yum!’s industry-leading digital technology team continues to pioneer the company’s AI-accelerated strategy with the recent announcement of Byte by Yum!, Yum!’s proprietary and digital AI-driven restaurant technology platform.

Generative AI-powered customer-facing experiences like automated ordering can help speed operations — but they’re often difficult to scale because of complexity and costs.

To manage that complexity, developers at Byte by Yum! harnessed NVIDIA NIM microservices and NVIDIA Riva to build new AI-accelerated voice ordering agents in under four months. The voice AI is deployed on Amazon EC2 P4d instances accelerated by NVIDIA A100 GPUs, which enables the agents to understand natural speech, process complex menu orders and suggest add-ons — increasing accuracy and customer satisfaction and helping reduce bottlenecks in high-volume locations.

The new collaboration with NVIDIA will help Yum! advance its ongoing efforts to have its engineering and data science teams in control of their own intelligence — and deliver scalable inference costs, making large-scale deployments possible.

“At Yum, we have a bold vision to deliver leading-edge, AI-powered technology capabilities to our customers and team members globally,” said Joe Park, chief digital and technology officer of Yum! Brands, Inc. and president of Byte by Yum!. “We are thrilled to partner with a pioneering company like NVIDIA to help us accelerate this ambition. This partnership will enable us to harness the rich consumer and operational datasets on our Byte by Yum! integrated platform to build smarter AI engines that will create easier experiences for our customers and team members.”

Rollout of AI Solutions Underway

Yum!’s voice AI agents are already being deployed across its brands, including in call centers to handle phone orders when demand surges during events like game days. An expanded rollout of AI solutions at up to 500 restaurants is expected this year.

Computer Vision and Restaurant Intelligence

Beyond AI-accelerated ordering, Yum! is also testing NVIDIA computer vision software to analyze drive-thru traffic and explore new use cases for AI to perceive, alert and adjust staffing, with the goal of optimizing service speed.

Another initiative focuses on NVIDIA AI-accelerated restaurant operational intelligence. Using NIM microservices, Yum! can deploy applications analyzing performance metrics across thousands of locations to generate customized recommendations for managers, identifying what top-performing stores do differently and applying those insights system-wide.

With the NVIDIA AI Enterprise software platform — available on AWS Marketplace — Byte by Yum! is streamlining AI development and deployment through scalable NVIDIA infrastructure in the cloud.

The bottom line: AI is making restaurant operations and dining experiences easier, faster and more personal for the world’s largest restaurant company.

Read More

Enterprises Ignite Big Savings With NVIDIA-Accelerated Apache Spark

Enterprises Ignite Big Savings With NVIDIA-Accelerated Apache Spark

Tens of thousands of companies worldwide rely on Apache Spark to crunch massive datasets to support critical operations, as well as predict trends, customer behavior, business performance and more. The faster a company can process and understand its data, the more it stands to make and save.

That’s why companies with massive datasets — including the world’s largest retailers and banks — have adopted NVIDIA RAPIDS Accelerator for Apache Spark. The open-source software runs on top of the NVIDIA accelerated computing platform to significantly accelerate the processing of end-to-end data science and analytics pipelines — without any code changes.

To make it even easier for companies to get value out of NVIDIA-accelerated Spark, NVIDIA today unveiled Project Aether — a collection of tools and processes that automatically qualify, test, configure and optimize Spark workloads for GPU acceleration at scale.

Project Aether Completes a Year’s Worth of Work in Less Than a Week 

Customers using Spark in production often manage tens of thousands of complex jobs, or more. Migrating from CPU-only to GPU-powered computing offers numerous and significant benefits, but can be a manual and time-consuming process.

Project Aether automates the myriad steps that companies previously have done manually, including analyzing all of their Spark jobs to identify the best candidates for GPU acceleration, as well as staging and performing test runs of each job. It uses AI to fine-tune the configuration of each job to obtain the maximum performance.

To understand the impact of Project Aether, consider an enterprise that has 100 Spark jobs to complete. With Project Aether, each of these jobs can be configured and optimized for NVIDIA GPU acceleration in as little as four days. The same process done manually by a single data engineer could take up to an entire year.

CBA Drives AI Transformation With NVIDIA-Accelerated Apache Spark

Running Apache Spark on NVIDIA accelerated computing helps enterprises around the world complete jobs faster and with less hardware compared with using CPUs only — saving time, space, power and cooling, as well as on-premises capital and operational costs in the cloud.

Australia’s largest financial institution, the Commonwealth Bank of Australia, is responsible for processing 60% of the continent’s financial transactions. CBA was experiencing challenges from the latency and costs associated with running its Spark workloads. Using CPU-only computing clusters, the bank estimates it faced nearly nine years of processing time for its training backlog — on top of handling already taxing daily data demands.

“With 40 million inferencing transactions a day, it was critical we were able to process these in a timely, reliable manner,” said Andrew McMullan, chief data and analytics officer at CBA.

Running RAPIDS Accelerator for Apache Spark on GPU-powered infrastructure provided CBA with a 640x performance boost, allowing the bank to process a training of 6.3 billion transactions in just five days. Additionally, on its daily volume of 40 million transactions, CBA is now able to conduct inference in 46 minutes and reduce costs by more than 80% compared with using a CPU-based solution.

McMullan says another value of NVIDIA-accelerated Apache Spark is how it offers his team the compute time efficiency needed to cost-effectively build models that can help CBA deliver better customer service, anticipate when customers may need assistance with home loans and more quickly detect fraudulent transactions.

CBA also plans to use NVIDIA-accelerated Apache Spark to better pinpoint where customers commonly end their digital journeys, enabling the bank to remediate when needed to reduce the rate of abandoned applications.

Global Ecosystem

RAPIDS Accelerator for Apache Spark is available through a global network of partners. It runs on Amazon Web Services, Cloudera, Databricks, Dataiku, Google Cloud, Microsoft Azure and Oracle Cloud Infrastructure.

Dell Technologies today also announced the integration of RAPIDS Accelerator for Apache Spark with Dell Data Lakehouse.

To get assistance through NVIDIA Project Aether with a large-scale migration of Apache Spark workloads, apply for access.

To learn more, register for NVIDIA GTC and attend these key sessions featuring Walmart, Capital One, CBA and other industry leaders:

See notice regarding software product information.

Read More

Driving Impact: NVIDIA Expands Automotive Ecosystem to Bring Physical AI to the Streets

Driving Impact: NVIDIA Expands Automotive Ecosystem to Bring Physical AI to the Streets

The autonomous vehicle (AV) revolution is here — and NVIDIA is at its forefront, bringing more than two decades of automotive computing, software and safety expertise to power innovation from the cloud to the car.

At NVIDIA GTC, a global AI conference taking place this week in San Jose, California, dozens of transportation leaders are showcasing their latest advancements with NVIDIA technologies that span passenger cars, trucks, commercial vehicles and more.

Mobility leaders are increasingly turning to NVIDIA’s three core accelerated compute platforms: NVIDIA DGX systems for training the AI-based stack in the data center, NVIDIA Omniverse and NVIDIA Cosmos running on NVIDIA OVX systems for simulation and synthetic data generation, and the NVIDIA DRIVE AGX in-vehicle computer to process real-time sensor data for safe, highly automated and autonomous driving capabilities.

For manufacturers and developers in the multitrillion-dollar auto industry, this unlocks new possibilities for designing, manufacturing and deploying functionally safe, intelligent mobility solutions — offering consumers safer, smarter and more enjoyable experiences.

Transforming Passenger Vehicles 

The U.S.’s largest automaker, General Motors (GM), is collaborating with NVIDIA to develop and build its next-generation vehicles, factories and robots using NVIDIA’s accelerated compute platforms. GM has been investing in NVIDIA GPU platforms for training AI models.

The companies’ collaboration now expands to include optimizing factory planning using Omnivese with Cosmos and deploying next-generation vehicles at scale accelerated by the NVIDIA DRIVE AGX. This will help GM build physical AI systems tailored to its company vision, craft and know-how, and ultimately enable mobility that’s safer, smarter and more accessible than ever.

Volvo Cars, which is using the NVIDIA DRIVE AGX in-vehicle computer in its next-generation electric vehicles, and its subsidiary Zenseact use the NVIDIA DGX platform to analyze and contextualize sensor data, unlock new insights and train future safety models that will enhance overall vehicle performance and safety.

Lenovo has teamed with robotics company Nuro to create a robust end-to-end system for level 4 autonomous vehicles that prioritize safety, reliability and convenience. The system is built on NVIDIA DRIVE AGX in-vehicle compute.

Advancements in Trucking

NVIDIA’s AI-driven technologies are also supercharging trucking, helping address pressing challenges like driver shortages, rising e-commerce demands and high operational costs. NVIDIA DRIVE AGX delivers the computational muscle needed for safe, reliable and efficient autonomous operations — improving road safety and logistics on a massive scale.

Gatik is integrating DRIVE AGX for the onboard AI processing necessary for its freight-only class 6 and 7 trucks, manufactured by Isuzu Motors, which offer driverless middle-mile delivery of a wide range of goods to Fortune 500 customers including Tyson Foods, Kroger and Loblaw.

Uber Freight is also adopting DRIVE AGX as the AI computing backbone of its current and future carrier fleets, sustainably enhancing efficiency and saving costs for shippers.

Torc is developing a scalable, physical AI compute system for autonomous trucks. The system uses NVIDIA DRIVE AGX in-vehicle compute and the NVIDIA DriveOS operating system with Flex’s Jupiter platform and manufacturing capabilities to support Torc’s productization and scaled market entry in 2027.

Growing Demand for DRIVE AGX

NVIDIA DRIVE AGX Orin platform is the AI brain behind today’s intelligent fleets — and the next wave of mobility is already arriving, as production vehicles built on the NVIDIA DRIVE AGX Thor centralized car computer start to hit the roads.

Magna is a key global automotive supplier helping to meet the surging demand for the NVIDIA Blackwell architecture-based DRIVE Thor platform — designed for the most demanding processing workloads, including those involving generative AI, vision language models and large language models (LLMs). Magna will develop driving systems built with DRIVE AGX Thor for integration in automakers’ vehicle roadmaps, delivering active safety and comfort functions along with interior cabin AI experiences.

Simulation and Data: The Backbone of AV Development

Earlier this year, NVIDIA announced the Omniverse Blueprint for AV simulation, a reference workflow for creating rich 3D worlds for autonomous vehicle training, testing and validation. The blueprint is expanding to include NVIDIA Cosmos world foundation models (WFMs) to amplify photoreal data variation.

Unveiled at the CES trade show in January, Cosmos is already being adopted in automotive, including by Plus, which is embedding Cosmos physical AI models into its SuperDrive technology, accelerating the development of level 4 self-driving trucks.

Foretellix is extending its integration of the blueprint, using the Cosmos Transfer WFM to add conditions like weather and lighting to its sensor simulation scenarios to achieve greater situation diversity. Mcity is integrating the blueprint into the digital twin of its AV testing facility to enable physics-based modeling of camera, lidar, radar and ultrasonic sensor data.

CARLA, which offers an open-source AV simulator, has integrated the blueprint to deliver high-fidelity sensor simulation. Global systems integrator Capgemini will be the first to use CARLA’s Omniverse integration for enhanced sensor simulation in its AV development platform.

NVIDIA is using Nexar’s extensive, high-quality, edge-case data to train and fine-tune NVIDIA Cosmos’ simulation capabilities. Nexar is tapping into Cosmos, neural infrastructure models and the NVIDIA DGX Cloud platform to supercharge its AI development, refining AV training, high-definition mapping and predictive modeling.

Enhancing In-Vehicle Experiences With NVIDIA AI Enterprise

Mobility leaders are integrating the NVIDIA AI Enterprise software platform, running on DRIVE AGX, to enhance in-vehicle experiences with generative and agentic AI.

At GTC, Cerence AI is showcasing Cerence xUI, its new LLM-based AI assistant platform that will advance the next generation of agentic in-vehicle user experiences. The Cerence xUI hybrid platform runs in the cloud as well as onboard the vehicle, optimized first on NVIDIA DRIVE AGX Orin.

As the foundation for Cerence xUI, the CaLLM family of language models is based on open-source foundation models and fine-tuned on Cerence AI’s automotive dataset. Tapping into NVIDIA AI Enterprise and bolstering inference performance including through the NVIDIA TensorRT-LLM library and NVIDIA NeMo, Cerence AI has optimized CaLLM to serve as the central agentic orchestrator facilitating enriched driver experiences at the edge and in the cloud.

SoundHound will also be demonstrating its next-generation in-vehicle voice assistant, which uses generative AI at the edge with NVIDIA DRIVE AGX, enhancing the in-car experience by bringing cloud-based LLM intelligence directly to vehicles.

The Complexity of Autonomy and NVIDIA’s Safety-First Solution

Safety is the cornerstone in deploying highly automated and autonomous vehicles to the roads at scale. But building AVs is one of today’s most complex computing challenges. It demands immense computational power, precision and an unwavering commitment to safety.

AVs and highly automated cars promise to extend mobility to those who need it most, reducing accidents and saving lives. To help deliver on this promise, NVIDIA has developed NVIDIA Halos, a full-stack comprehensive safety system that unifies vehicle architecture, AI models, chips, software, tools and services for the safe development of AVs from the cloud to the car.

NVIDIA will host its inaugural AV Safety Day at GTC today, featuring in-depth discussions on automotive safety frameworks and implementation.

In addition, NVIDIA will host Automotive Developer Day on Thursday, March 20, offering sessions on the latest advancements in end-to-end AV development and beyond.

New Tools for AV Developers

NVIDIA also released new NVIDIA NIM microservices for automotive — designed to accelerate development and deployment of end-to-end stacks from cloud to car. The new NIM microservices for in-vehicle applications, which utilize the nuScenes dataset by Motional, include:

  • BEVFormer, a state-of-the-art transformer-based model that fuses multi-frame camera data into a unified bird’s-eye-view representation for 3D perception.
  • SparseDrive, an end-to-end autonomous driving model that performs motion prediction and planning simultaneously, outputting a safe planning trajectory.

For automotive enterprise applications, NVIDIA offers a variety of models, including NV-CLIP, a multimodal transformer model that generates embeddings from images and text; Cosmos Nemotron, a vision language model that queries and summarizes images and videos for multimodal understanding and AI-powered perception; and many more.

Learn more about NVIDIA’s latest automotive news by watching the NVIDIA GTC keynote and register for sessions from NVIDIA and industry leaders at the show, which runs through March 21.

Read More

NVIDIA Unveils AI-Q Blueprint to Connect AI Agents for the Future of Work

NVIDIA Unveils AI-Q Blueprint to Connect AI Agents for the Future of Work

AI agents are the new digital workforce, transforming business operations, automating complex tasks and unlocking new efficiencies. Now, with the ability to collaborate, these agents can work together to solve complex problems and drive even greater impact.

Businesses across industries, including sports and finance, can more quickly harness these benefits with AI-Q — a new NVIDIA Blueprint for developing agentic systems that can use reasoning to unlock knowledge in enterprise data.

Smarter Agentic AI Systems With NVIDIA AI-Q and AgentIQ Toolkit

AI-Q provides an easy-to-follow reference for integrating NVIDIA accelerated computing, partner storage platforms, and software and tools — including the new NVIDIA Llama Nemotron reasoning models. AI-Q offers a powerful foundation for enterprises to build digital workforces that break down agentic silos and are capable of handling complex tasks with high accuracy and speed.

AI-Q integrates fast multimodal extraction and world-class retrieval, using NVIDIA NeMo Retriever, NVIDIA NIM microservices and AI agents.

The blueprint is powered by the new NVIDIA AgentIQ toolkit for seamless, heterogeneous connectivity between agents, tools and data. Released today on GitHub, AgentIQ is an open-source software library for connecting, profiling and optimizing teams of AI agents fueled by enterprise data to create multi-agent, end-to-end systems. It can be easily integrated with existing multi-agent systems — either in parts or as a complete solution — with a simple onboarding process that’s 100% opt-in.

The AgentIQ toolkit also enhances transparency with full system traceability and profiling — enabling organizations to monitor performance, identify inefficiencies and gain fine-grained understanding of how business intelligence is generated. This profiling data can be used with NVIDIA NIM and the NVIDIA Dynamo open-source library to optimize the performance of agentic systems.

The New Enterprise AI Agent Workforce

As AI agents become digital employees, IT teams will support onboarding and training. The AI-Q blueprint and AgentIQ toolkit support digital employees by enabling collaboration between agents and optimizing performance across different agentic frameworks.

Enterprises using these tools will be able to more easily connect AI agent teams across solutions — like Salesforce’s Agentforce, Atlassian Rovo in Confluence and Jira, and the ServiceNow AI platform for business transformation — to break down silos, streamline tasks and cut response times from days to hours.

AgentIQ also integrates with frameworks and tools like CrewAI, LangGraph, Llama Stack, Microsoft Azure AI Agent Service and Letta, letting developers work in their preferred environment.

Azure AI Agent Service is integrated with AgentIQ to enable more efficient AI agents and orchestration of multi-agent frameworks using Semantic Kernel, which is fully supported in AgentIQ.

A wide range of industries are integrating visual perception and interactive capabilities into their agents and copilots.

Financial services leader Visa is using AI agents to streamline cybersecurity, automating phishing email analysis at scale. Using the profiler feature of AI-Q, Visa can optimize agent performance and costs, maximizing AI’s role in efficient threat response.

Get Started With AI-Q and AgentIQ

AI-Q integration into the NVIDIA Metropolis VSS blueprint is enabling multimodal agents, combining visual perception with speech, translation and data analytics for enhanced intelligence.

Developers can use the AgentIQ toolkit open-source library today and sign up for this hackathon to build hands-on skills for advancing agentic systems.

Plus, learn how an NVIDIA solutions architect used the AgentIQ toolkit to improve AI code generation.

Agentic systems built with AI-Q require a powerful AI data platform. NVIDIA partners are delivering these customized platforms that continuously process data to let AI agents quickly access knowledge to reason and respond to complex queries.

See notice regarding software product information.

Read More

NVIDIA Unveils Open Physical AI Dataset to Advance Robotics and Autonomous Vehicle Development

NVIDIA Unveils Open Physical AI Dataset to Advance Robotics and Autonomous Vehicle Development

Teaching autonomous robots and vehicles how to interact with the physical world requires vast amounts of high-quality data. To give researchers and developers a head start, NVIDIA is releasing a massive, open-source dataset for building the next generation of physical AI.

Announced at NVIDIA GTC, a global AI conference taking place this week in San Jose, California, this commercial-grade, pre-validated dataset can help researchers and developers kickstart physical AI projects that can be prohibitively difficult to start from scratch. Developers can either directly use the dataset for model pretraining, testing and validation — or use it during post-training to fine-tune world foundation models, accelerating the path to deployment.

The initial dataset is now available on Hugging Face, offering developers 15 terabytes of data representing more than 320,000 trajectories for robotics training, plus up to 1,000 Universal Scene Description (OpenUSD) assets, including a SimReady collection. Dedicated data to support end-to-end autonomous vehicle (AV) development — which will include 20-second clips of diverse traffic scenarios spanning over 1,000 cities across the U.S. and two dozen European countries — is coming soon.

flythrough of synthetically generated objects
The NVIDIA Physical AI Dataset includes hundreds of SimReady assets for rich scenario building.

This dataset will grow over time to become the world’s largest unified and open dataset for physical AI development. It could be applied to develop AI models to power robots that safely maneuver warehouse environments, humanoid robots that support surgeons during procedures and AVs that can navigate complex traffic scenarios like construction zones.

The NVIDIA Physical AI Dataset is slated to contain a subset of the real-world and synthetic data NVIDIA uses to train, test and validate physical AI for the NVIDIA Cosmos world model development platform, the NVIDIA DRIVE AV software stack, the NVIDIA Isaac AI robot development platform and the NVIDIA Metropolis application framework for smart cities.

Early adopters include the Berkeley DeepDrive Center at the University of California, Berkeley, the Carnegie Mellon Safe AI Lab and the Contextual Robotics Institute at University of California, San Diego.

“We can do a lot of things with this dataset, such as training predictive AI models that help autonomous vehicles better track the movements of vulnerable road users like pedestrians to improve safety,” said Henrik Christensen, director of multiple robotics and autonomous vehicle labs at UCSD. “A dataset that provides a diverse set of environments and longer clips than existing open-source resources will be tremendously helpful to advance robotics and AV research.”

Addressing the Need for Physical AI Data

The NVIDIA Physical AI Dataset can help developers scale AI performance during pretraining, where more data helps build a more robust model — and during post-training, where an AI model is trained on additional data to improve its performance for a specific use case.

Collecting, curating and annotating a dataset that covers diverse scenarios and accurately represents the physics and variation of the real world is time-consuming, presenting a bottleneck for most developers. For academic researchers and small enterprises, running a fleet of vehicles over months to gather data for autonomous vehicle AI is impractical and costly — and, since much of the footage collected is uneventful, typically just 10% of data is used for training.

But this scale of data collection is essential to building safe, accurate, commercial-grade models. NVIDIA Isaac GR00T robotics models take thousands of hours of video clips for post-training — the GR00T N1 model, for example, was trained on an expansive humanoid dataset of real and synthetic data. The NVIDIA DRIVE AV end-to-end AI model for autonomous vehicles requires tens of thousands of hours of driving data to develop.

 

This open dataset, comprising thousands of hours of multicamera video at unprecedented diversity, scale and geography — will particularly benefit the field of safety research by enabling new work on identifying outliers and assessing model generalization performance. The effort contributes to NVIDIA Halos’ full-stack AV safety system.

In addition to harnessing the NVIDIA Physical AI Dataset to help meet their data needs, developers can further boost AI development with tools like NVIDIA NeMo Curator, which process vast datasets efficiently for model training and customization. Using NeMo Curator, 20 million hours of video can be processed in just two weeks on NVIDIA Blackwell GPUs, compared with 3.4 years on unoptimized CPU pipelines.

Robotics developers can also tap the new NVIDIA Isaac GR00T blueprint for synthetic manipulation motion generation, a reference workflow built on NVIDIA Omniverse and NVIDIA Cosmos that uses a small number of human demonstrations to create massive amounts of synthetic motion trajectories for robot manipulation.

University Labs Set to Adopt Dataset for AI Development

The robotics labs at UCSD include teams focused on medical applications, humanoids and in-home assistive technology. Christensen anticipates that the Physical AI Dataset’s robotics data could help develop semantic AI models that understand the context of spaces like homes, hotel rooms and hospitals.

“One of our goals is to achieve a level of understanding where, if a robot was asked to put your groceries away, it would know exactly which items should go in the fridge and what goes in the pantry,” he said.

In the field of autonomous vehicles, Christensen’s lab could apply the dataset to train AI models to understand the intention of various road users and predict the best action to take. His research teams could also use the dataset to support the development of digital twins that simulate edge cases and challenging weather conditions. These simulations could be used to train and test autonomous driving models in situations that are rare in real-world environments.

At Berkeley DeepDrive, a leading research center on AI for autonomous systems, the dataset could support the development of policy models and world foundation models for autonomous vehicles.

“Data diversity is incredibly important to train foundation models,” said Wei Zhan, codirector of Berkeley DeepDrive. “This dataset could support state-of-the-art research for public and private sector teams developing AI models for autonomous vehicles and robotics.”

Researchers at Carnegie Mellon University’s Safe AI Lab plan to use the dataset to advance their work evaluating and certifying the safety of self-driving cars. The team plans to test how a physical AI foundation model trained on this dataset performs in a simulation environment with rare conditions — and compare its performance to an AV model trained on existing datasets.

“This dataset covers different types of roads and geographies, different infrastructure, different weather environments,” said Ding Zhao, associate professor at CMU and head of the Safe AI Lab. “Its diversity could be quite valuable in helping us train a model with causal reasoning capabilities in the physical world that understands edge cases and long-tail problems.”

Access the NVIDIA Physical AI dataset on Hugging Face. Build foundational knowledge with courses such as the Learn OpenUSD learning path and Robotics Fundamentals learning path. And to learn more about the latest advancements in physical AI, watch the GTC keynote by NVIDIA founder and CEO Jensen Huang.

See notice regarding software product information.

Read More