NVIDIA's Open-Source Model Debuts Amid Global AI Agent Frenzy

Stock News
昨天

As AI agents like Anthropic's Claude Cowork and OpenClaw gain worldwide popularity, NVIDIA is positioning itself to capitalize on this emerging trend. The company has introduced Nemotron 3 Super, an open-source large language model specifically designed for scalable, complex AI agent systems. In benchmark testing, Nemotron 3 Super achieved top performance among open-source models, with an 85.6% success rate on OpenClaw tasks, rivaling proprietary models such as Claude Opus 4.6 and GPT-5.4.

This development reinforces NVIDIA's transition from being solely an AI chip supplier to becoming a full-stack platform provider encompassing models, toolchains, cloud inference services, and AI ecosystems. This strategic shift could potentially drive NVIDIA's stock to new highs and uplift the global AI computing supply chain.

NVIDIA stated that the model integrates cutting-edge reasoning capabilities to handle massive AI tasks with high precision, making it suitable for enterprise-grade autonomous AI agents. The new 120-billion parameter open model employs a Mixture of Experts architecture and incorporates three key innovations, delivering up to 3x better reasoning performance, 5x higher throughput, and 2x greater accuracy compared to its predecessor.

AI search leader Perplexity has begun offering Nemotron 3 Super to users for agent-driven systematic searches, while tech companies like CodeRabbit, Factory, and Greptile are integrating the model into their AI agent services to improve accuracy and operational efficiency at lower costs. Research institutions such as Edison Scientific and Lila Sciences plan to utilize the model for advanced tasks like deep literature retrieval, data science, and molecular analysis.

Additionally, companies including Amdocs, Palantir, Cadence, Dassault Systèmes, and Siemens are actively deploying and customizing NVIDIA's model to automate workflows in telecommunications, cybersecurity, semiconductor design, and manufacturing.

Nemotron 3 Super features a "high total parameters, low activation" design optimized for enterprise agent workflows. With 120 billion total parameters but only 12 billion activated during inference, it supports a 1-million-token context window and requires a minimum deployment of 8×H100 80GB GPUs. Its architecture combines LatentMoE, Mamba-2, and minimal Attention layers, making it ideal for multi-agent orchestration and long-context tasks rather than single-turn dialogue performance.

Qualcomm CEO Cristiano Amon recently highlighted at MWC Barcelona that the AI agent wave will transform digital ecosystems, predicting 2026 as the "Year of AI Agents." He emphasized that agents will evolve from reactive tools to proactive systems capable of observation, interpretation, and action.

NVIDIA's broader ambition is to become an "AI infrastructure general contractor" rather than just a chip supplier. Nemotron 3 Super emphasizes high throughput and cost efficiency for agent systems without sacrificing accuracy. Technical reports indicate it achieves 2.2x the throughput of GPT-OSS-120B and 7.5x that of Qwen3.5-122B under 8k input/64k output settings.

The model addresses two key enterprise challenges: context explosion, where multi-agent workflows generate 15x more tokens than standard chats, and "thinking tax," where complex reasoning becomes prohibitively expensive. With its 1-million-token context window, Nemotron 3 Super maintains full workflow states in memory, preventing goal drift. On NVIDIA's Blackwell platform, it runs at NVFP4 precision, reducing memory needs and achieving up to 4x faster inference than Hopper FP8 without accuracy loss.

Trained on synthetic data from advanced reasoning models, Nemotron 3 Super is available via build.nvidia.com, Perplexity, OpenRouter, and Hugging Face. This release signals NVIDIA's full-stack strategy, integrating model architecture, inference optimization, and deployment ecosystems into a unified AI infrastructure.

Dell Technologies is incorporating the model into its Enterprise Hub on Hugging Face for local deployment within Dell AI Factory, while Hewlett Packard Enterprise is adding Nemotron to its agents hub for scalable enterprise AI adoption.

NVIDIA launched the Nemotron 3 series in December and will host its GPU Technology Conference from March 16-19, showcasing advancements in physical AI, AI factories, and agentic AI.

With an 85.6% score on PinchBench and OpenClaw tasks, Nemotron 3 Super is positioned as a scalable "agent orchestration brain" for complex multi-step workflows. Under CEO Jensen Huang's leadership, NVIDIA's AI GPU and CUDA ecosystem form a robust moat, further strengthened by this release. The company is shifting its competitive advantage from GPU performance alone to integrated model architecture, inference stacks, and deployment solutions.

Wall Street analysts have turned increasingly bullish on NVIDIA following Nemotron 3 Super's introduction. Morgan Stanley recently reaffirmed NVIDIA as its top semiconductor pick, maintaining an "Overweight" rating and $260 price target, citing attractive buying opportunities after recent consolidation. The average analyst price target compiled by TipRanks sits at $273, implying 47% upside over the next 12 months.

Morgan Stanley's channel checks indicate a growing global AI compute supply-demand gap, with hyperscalers maintaining aggressive AI workload expansion. Even as clients like Amazon and Meta develop custom AI chips or adopt AMD GPUs, their NVIDIA procurement is expected to grow over 80% by 2026. The upcoming GTC conference is anticipated to showcase NVIDIA's technology roadmap, addressing market share concerns and highlighting new opportunities in physical AI.

As model scale, inference chains, and multimodal AI workloads drive exponential compute demand, tech giants are prioritizing AI infrastructure investments. Global investors continue to anchor the "AI bull narrative" around NVIDIA and AMD's product cycles, sustaining momentum for related sectors like power, cooling systems, and optical networking.

Analysts project that Amazon, Alphabet, Meta, Oracle, and Microsoft will collectively spend approximately $650 billion on AI-related capital expenditures in 2026, with some estimates exceeding $700 billion—representing over 70% year-over-year growth. From 2023 to 2026, these five tech giants are expected to invest around $1.5 trillion in AI infrastructure, compared to roughly $600 billion in total historical investments prior to 2022.

免责声明:投资有风险,本文并非投资建议,以上内容不应被视为任何金融产品的购买或出售要约、建议或邀请,作者或其他用户的任何相关讨论、评论或帖子也不应被视为此类内容。本文仅供一般参考,不考虑您的个人投资目标、财务状况或需求。TTM对信息的准确性和完整性不承担任何责任或保证,投资者应自行研究并在投资前寻求专业建议。

热议股票

  1. 1
     
     
     
     
  2. 2
     
     
     
     
  3. 3
     
     
     
     
  4. 4
     
     
     
     
  5. 5
     
     
     
     
  6. 6
     
     
     
     
  7. 7
     
     
     
     
  8. 8
     
     
     
     
  9. 9
     
     
     
     
  10. 10