STOCK TITAN

NVIDIA Enters Production With Dynamo, the Broadly Adopted Inference Operating System for AI Factories

Rhea-AI Impact
(Moderate)
Rhea-AI Sentiment
(Neutral)
Tags
AI

NVIDIA (NVDA) announced Dynamo 1.0, an open source, production-grade inference operating system for AI factories, available March 16, 2026. Dynamo integrates with TensorRT-LLM and open frameworks to boost inference on NVIDIA Blackwell GPUs by up to 7x, and is adopted by major cloud providers and enterprises.

The software adds cluster traffic control, smarter memory movement, and GPU-to-GPU data routing to lower token cost and scale agentic AI deployments.

Loading...
Loading translation...

Positive

  • Inference performance up to 7x on NVIDIA Blackwell GPUs
  • Dynamo 1.0 available March 16, 2026
  • Integrated with major cloud providers: AWS, Azure, Google Cloud, OCI
  • Adopted by global enterprises and AI partners including PayPal, Pinterest

Negative

  • None.

Key Figures

Dynamo version: 1.0 Blackwell GPU boost: up to 7x GPU deployment scale: millions of GPUs +1 more
4 metrics
Dynamo version 1.0 Initial production release of NVIDIA Dynamo inference operating system
Blackwell GPU boost up to 7x Inference performance improvement for NVIDIA Blackwell GPUs using Dynamo
GPU deployment scale millions of GPUs Dynamo aims to lower token cost and increase revenue opportunity across large installed base
Pinterest users hundreds of millions Scale of users cited for multimodal AI experiences powered by NVIDIA infrastructure

Market Reality Check

Price: $183.19 Vol: Volume 213,712,481 vs 20-...
normal vol
$183.19 Last Close
Volume Volume 213,712,481 vs 20-day average 186,899,209 (relative volume 1.14). normal
Technical Price 183.187 is trading above 200-day MA at 177.64 and 13.67% below the 52-week high.

Peers on Argus

NVDA gained 2.19% while key peers were mostly flat to down: AVGO -0.34%, TSM -0....

NVDA gained 2.19% while key peers were mostly flat to down: AVGO -0.34%, TSM -0.30%, MU -0.98%, NXPI -1.38%, with only AMD modestly up 0.82%, suggesting a stock-specific reaction to the Dynamo 1.0 news.

Common Catalyst AI-related collaboration headlines (e.g., NXPI AI news) appear, but only one peer has same-day AI news, reinforcing that today’s move is primarily NVIDIA-specific.

Previous AI Reports

5 past events · Latest: Mar 11 (Positive)
Same Type Pattern 5 events
Date Event Sentiment Move Catalyst
Mar 11 AI cloud partnership Positive +0.7% NVIDIA invested $2B in Nebius to scale full-stack AI cloud and AI factories.
Mar 03 AI conference preview Positive -1.3% Announcement of GTC 2026 with large attendee base and AI-focused program.
Feb 17 Hyperscale AI deal Positive +1.6% Multiyear partnership with Meta to codesign AI infrastructure and deploy GPUs.
Feb 03 Industrial AI partnership Positive -2.8% Long-term collaboration with Dassault Systèmes on industrial AI and virtual twins.
Jan 26 AI factory expansion Positive -0.6% Expanded CoreWeave collaboration, including $2B investment to build AI factories.
Pattern Detected

Recent AI-related announcements have produced mixed, often modest, next-day price moves, indicating that even substantial AI ecosystem news does not consistently drive large directional reactions.

Recent Company History

Over the past few months, NVIDIA has consistently expanded its AI ecosystem through large-scale partnerships and infrastructure initiatives. AI-tagged news includes multi-billion investments in CoreWeave and Nebius to build over 5 gigawatts of AI factories, and strategic collaborations with Meta and Dassault Systèmes on hyperscale and industrial AI platforms. GTC 2026 was positioned as a major showcase for the full AI stack. The Dynamo 1.0 "AI factory OS" launch fits this pattern of reinforcing NVIDIA’s role at the center of large, globally distributed AI infrastructure.

Historical Comparison

-0.5% avg move · AI-tagged news for NVDA has shown an average next-day move of -0.5%, with mixed positive and negativ...
AI
-0.5%
Average Historical Move AI

AI-tagged news for NVDA has shown an average next-day move of -0.5%, with mixed positive and negative reactions. Today’s Dynamo 1.0 launch and 2.19% gain sit slightly above this typical response range.

AI-tagged history shows NVIDIA moving from major AI infrastructure and hyperscale partnerships toward software and orchestration layers. The Dynamo 1.0 launch extends this path by positioning NVIDIA at the “operating system” layer for AI factories on top of its Blackwell hardware base.

Market Pulse Summary

This announcement introduces NVIDIA Dynamo 1.0 as an open source “operating system” for AI factories...
Analysis

This announcement introduces NVIDIA Dynamo 1.0 as an open source “operating system” for AI factories, aimed at boosting Blackwell GPU inference performance by up to 7x and improving utilization across millions of GPUs. It builds on a series of AI partnerships with hyperscalers and enterprises, reinforcing NVIDIA’s role across hardware and software. Investors may watch adoption by named partners, scaling of agentic and multimodal workloads, and how Dynamo contributes to future data center revenue growth.

Key Terms

inference, gpu, cuda, open source, +3 more
7 terms
inference technical
"NVIDIA Dynamo 1.0 provides a production-grade, open source foundation for inference at scale."
Inference is the process of drawing a conclusion from available evidence or data, like a detective piecing together clues to form a likely story. For investors it matters because these judgments turn raw reports, test results, or market signals into expectations about future performance, risk, or regulatory outcomes—so how someone infers from the same facts can change investment decisions and valuation.
gpu technical
"Dynamo boosts inference performance of NVIDIA Blackwell GPUs by up to 7x, lowering token cost..."
A GPU (graphics processing unit) is a specialized computer chip designed to handle many calculations at once, originally for rendering images and video but now widely used for tasks like artificial intelligence, data analysis and high-performance computing. Investors watch GPU demand and prices because strong sales often signal growth for chip makers and their customers, affect profit margins and capital spending, and can forecast wider trends in gaming, AI adoption and cloud services.
cuda technical
"NVIDIA also contributes TensorRT-LLM CUDA kernels to the FlashInfer project..."
CUDA is a software platform developed to let programmers use graphics processors (GPUs) for heavy computing tasks beyond rendering images, similar to turning a powerful video card into a team of many specialized workers tackling complex calculations in parallel. It matters to investors because widespread use of CUDA can drive demand for GPUs and related data-center services, accelerate development of AI and scientific applications, and therefore influence revenue and valuation for hardware and cloud companies tied to high-performance computing.
open source technical
"Dynamo 1.0 provides a production-grade, open source foundation for inference at scale."
Open source means the underlying software code is made publicly available so anyone can inspect, copy, change, or reuse it, similar to sharing a recipe so others can tweak and improve it. Investors care because open source can lower a company’s development costs, speed innovation through community contributions, and increase adoption, but it also affects how a business makes money and can introduce licensing or support risks that influence value.
agentic ai technical
"As agentic AI systems move into production across industries, scaling inference..."
Agentic AI refers to computer systems that can make their own decisions and take actions without needing someone to tell them what to do each time. It's like giving a robot a degree of independence to solve problems or achieve goals on its own, which matters because it could change how we work and interact with technology in everyday life.
ai factories technical
"Dynamo 1.0 functions as the distributed “operating system” of AI factories..."
AI factories are organized platforms and processes that turn raw data and computing power into finished AI products and services at scale — think of them as automated assembly lines for machine intelligence. For investors, they matter because they concentrate the tools, data and infrastructure that speed up development, lower unit costs and make it easier to roll out new AI features, which can translate into faster revenue growth or cost savings for companies that operate them.
multimodal technical
"Delivering an intuitive, multimodal AI experience to hundreds of millions of users..."
Multimodal describes an approach, product, or system that uses two or more different types of inputs, methods, or channels — for example combining text, images and audio in a technology product, or blending drugs, devices and therapy in medical care. For investors, multimodal solutions can broaden market reach and competitive differentiation but also add development cost, operational complexity and regulatory hurdles; think of it like a hybrid car that offers more capabilities but requires more parts and oversight.

AI-generated analysis. Not financial advice.

News Summary:

  • NVIDIA Dynamo 1.0 provides a production-grade, open source foundation for inference at scale.
  • Dynamo and NVIDIA TensorRT-LLM optimizations integrate natively into open source frameworks such as LangChain, llm-d, LMCache, SGLang and vLLM to boost inference performance.
  • Dynamo boosts inference performance of NVIDIA Blackwell GPUs by up to 7x, lowering token cost and increasing revenue opportunity for millions of GPUs with free, open source software.
  • NVIDIA inference platform integrated by cloud service providers, Amazon Web Services (AWS), Microsoft Azure, Google Cloud and Oracle Cloud Infrastructure (OCI), along with NVIDIA cloud partners Alibaba Cloud, CoreWeave, Together AI and Nebius — and adopted by AI-native companies Cursor and Perplexity; inference endpoint providers Baseten, Deep Infra and Fireworks; and global enterprises ByteDance, Meituan, PayPal and Pinterest.

SAN JOSE, Calif., March 16, 2026 (GLOBE NEWSWIRE) -- GTC -- NVIDIA today announced NVIDIA Dynamo 1.0, open source software for generative and agentic inference at scale, with widespread global adoption. Together with the NVIDIA Blackwell platform, Dynamo 1.0 enables cloud providers, AI innovators and global enterprises to deliver high-performance AI inference with unmatched scale, efficiency and speed.

As agentic AI systems move into production across industries, scaling inference within a data center has become a complex challenge of resource orchestration, with requests of varying sizes and modalities, as well as performance objectives, arriving in unpredictable bursts.

Just as a computer’s operating system coordinates hardware and applications, Dynamo 1.0 functions as the distributed “operating system” of AI factories, seamlessly orchestrating GPU and memory resources across the cluster to power complex AI workloads. In recent industry benchmarks, Dynamo boosted the inference performance of NVIDIA Blackwell GPUs by up to 7x, lowering token cost and increasing revenue opportunity for millions of GPUs with free, open source software.

“Inference is the engine of intelligence, powering every query, every agent and every application,” said Jensen Huang, founder and CEO of NVIDIA. “With NVIDIA Dynamo, we’ve created the first-ever ‘operating system’ for AI factories. The rapid adoption across our ecosystem shows this next wave of agentic AI is here, and NVIDIA is powering it at global scale.”

Dynamo 1.0 splits inference work across GPUs by adding smarter “traffic control” and the ability to move data between GPUs and lower-cost storage, reducing wasted work and easing memory limits. For agentic AI and long prompts, it can route requests to GPUs that already have the most relevant “short-term memory” from earlier steps, then offload that memory when it is not needed.

NVIDIA Inference Platform Gains Momentum
NVIDIA is accelerating the open source ecosystem by integrating Dynamo and NVIDIA TensorRT™-LLM library optimizations into popular frameworks from providers such as LangChain, llm-d, LMCache, SGLang, vLLM and more. Core Dynamo building blocks like KVBM for smarter memory management, NVIDIA NIXL for fast GPU-to-GPU data movement and NVIDIA Grove for simplified scaling are also available as standalone modules. NVIDIA also contributes TensorRT-LLM CUDA® kernels to the FlashInfer project so they can be natively integrated into open source frameworks.

The NVIDIA inference platform is supported across the AI ecosystem, including:

  • Cloud Service Providers: Amazon Web Services (AWS), Microsoft Azure, Google Cloud, OCI
  • NVIDIA Cloud Partners: Alibaba Cloud, CoreWeave, Crusoe, DigitalOcean, Gcore, GMI Cloud, Lightning AI, Nebius, Nscale, Together AI, Vultr
  • AI-Native Companies: Cursor, Hebbia, Perplexity
  • Inference Endpoint Providers: Baseten, Deep Infra, Fireworks
  • Global Enterprises: AstraZeneca, BlackRock, ByteDance, Coupang, Instacart, Meituan, PayPal, Pinterest, Shopee, SoftBank Corp.

Chen Goldberg, executive vice president of product and engineering at CoreWeave, said: “As AI moves from experimental pilots to continuous, large-scale production, the underlying infrastructure must be as dynamic as the models it supports. Supporting NVIDIA Dynamo allows us to offer a more seamless, resilient environment for deploying complex AI agents. This foundation provides the durability and high-performance orchestration required to move the industry’s most ambitious agentic workloads into global production.”

Danila Shtan, chief technology officer of Nebius, said: “Delivering reliable AI inference at scale isn’t just about powerful GPUs, it’s about the software that turns that performance into real customer outcomes. We value how NVIDIA’s software stack, from Dynamo to TensorRT-LLM, brings deep optimization, predictable performance and faster time to deployment, helping us offer customers a simpler, higher-performance path to production AI.”

Matt Madrigal, chief technology officer of Pinterest, said: “Delivering an intuitive, multimodal AI experience to hundreds of millions of users requires real-time intelligence at global scale. As a significant adopter in open source, we’re committed to building scalable AI technologies. With NVIDIA Dynamo optimizing our deployment, we’re expanding the seamless and personalized experiences we deliver, powered by high-performance AI infrastructure.”

Vipul Ved Prakash, cofounder and CEO of Together AI, said: “AI natives require inference that can reliably and efficiently scale with their application. NVIDIA Dynamo 1.0, combined with cutting-edge inference research from Together AI, helps us deliver a high-performance stack to offer accelerated, cost-effective inference for large-scale production workloads.”

Dynamo 1.0 is available today to developers worldwide. To learn more and get started, read the blog and visit the Dynamo webpage.

About NVIDIA
NVIDIA (NASDAQ: NVDA) is the world leader in AI and accelerated computing.

For further information, contact:
Jordan Byrnes
press@nvidia.com

Certain statements in this press release including, but not limited to, statements as to: Inference being the engine of intelligence, powering every query, every agent and every application; NVIDIA powering the next wave of agentic AI at global scale; the benefits, impact, performance, and availability of NVIDIA’s products, services, and technologies; expectations with respect to NVIDIA’s third party arrangements, including with its collaborators and partners; expectations with respect to technology developments; and other statements that are not historical facts are forward-looking statements within the meaning of Section 27A of the Securities Act of 1933, as amended, and Section 21E of the Securities Exchange Act of 1934, as amended, which are subject to the “safe harbor” created by those sections based on management’s beliefs and assumptions and on information currently available to management and are subject to risks and uncertainties that could cause results to be materially different than expectations. Important factors that could cause actual results to differ materially include: global economic and political conditions; NVIDIA’s reliance on third parties to manufacture, assemble, package and test NVIDIA’s products; the impact of technological development and competition; development of new products and technologies or enhancements to NVIDIA’s existing product and technologies; market acceptance of NVIDIA’s products or NVIDIA’s partners’ products; design, manufacturing or software defects; changes in consumer preferences or demands; changes in industry standards and interfaces; unexpected loss of performance of NVIDIA’s products or technologies when integrated into systems; and changes in applicable laws and regulations, as well as other factors detailed from time to time in the most recent reports NVIDIA files with the Securities and Exchange Commission, or SEC, including, but not limited to, its annual report on Form 10-K and quarterly reports on Form 10-Q. Copies of reports filed with the SEC are posted on the company’s website and are available from NVIDIA without charge. These forward-looking statements are not guarantees of future performance and speak only as of the date hereof, and, except as required by law, NVIDIA disclaims any obligation to update these forward-looking statements to reflect future events or circumstances.

© 2026 NVIDIA Corporation. All rights reserved. NVIDIA, the NVIDIA logo, CUDA, NVIDIA Hopper and TensorRT are trademarks and/or registered trademarks of NVIDIA Corporation in the U.S. and other countries. Other company and product names may be trademarks of the respective companies with which they are associated. Features, pricing, availability and specifications are subject to change without notice.

A photo accompanying this announcement is available at https://www.globenewswire.com/NewsRoom/AttachmentNg/5e492b13-9fd8-42ac-be36-fbd2a4bc1c3f


FAQ

What is NVIDIA Dynamo 1.0 and how does it affect NVDA inference workloads?

Dynamo 1.0 is an open source inference operating system that coordinates GPUs and memory for AI workloads. According to NVIDIA, it orchestrates cluster resources, reduces wasted work and can route short-term memory to speed agentic and long-prompt inference for production deployments.

How much performance improvement does Dynamo 1.0 deliver for NVDA Blackwell GPUs?

Dynamo 1.0 can boost inference performance on Blackwell GPUs by up to 7x. According to NVIDIA, that improvement lowers token costs and increases potential revenue opportunity for many GPUs through software optimization and smarter resource routing.

Which cloud providers and partners support NVIDIA Dynamo for NVDA customers?

Major cloud providers supporting Dynamo include AWS, Microsoft Azure, Google Cloud and OCI. According to NVIDIA, cloud partners and AI-native companies like CoreWeave, Together AI, Cursor and Perplexity have integrated the inference platform.

When is NVIDIA Dynamo 1.0 available and how can developers access it for NVDA systems?

Dynamo 1.0 is available today, March 16, 2026, for developers worldwide. According to NVIDIA, developers can read the blog and visit the Dynamo webpage to download the open source modules and integrate them with TensorRT-LLM and popular frameworks.

Which open source frameworks integrate with NVIDIA Dynamo and TensorRT-LLM for NVDA users?

Dynamo and TensorRT-LLM integrate natively with frameworks like LangChain, llm-d, LMCache, SGLang and vLLM. According to NVIDIA, these integrations enable native performance gains and easier deployment across inference frameworks used in production.

What core modules in Dynamo 1.0 help scale NVDA inference clusters?

Core Dynamo modules include KVBM for memory management, NIXL for fast GPU-to-GPU data movement and Grove for simplified scaling. According to NVIDIA, these building blocks are available as standalone modules to optimize cluster orchestration and reduce memory limits.
Nvidia Corporation

NASDAQ:NVDA

View NVDA Stock Overview

NVDA Rankings

NVDA Latest News

NVDA Latest SEC Filings

NVDA Stock Data

4.38T
23.27B
Semiconductors
Semiconductors & Related Devices
Link
United States
SANTA CLARA