Close Menu
FSNN | Free Speech News NetworkFSNN | Free Speech News Network
  • Home
  • News
    • Politics
    • Legal & Courts
    • Tech & Big Tech
    • Campus & Education
    • Media & Culture
    • Global Free Speech
  • Opinions
    • Debates
  • Video/Live
  • Community
  • Freedom Index
  • About
    • Mission
    • Contact
    • Support
Trending

A.B. 1043’s Internet Age Gates Hurt Everyone

14 minutes ago

Tax the Rich All You Want. It Won’t Fix the Deficit.

16 minutes ago

Journalist Marisa Kabas discusses her successful fight for bodycam videos of DOGE raid at US Institute of Peace

33 minutes ago
Facebook X (Twitter) Instagram
Facebook X (Twitter) Discord Telegram
FSNN | Free Speech News NetworkFSNN | Free Speech News Network
Market Data Newsletter
Thursday, March 12
  • Home
  • News
    • Politics
    • Legal & Courts
    • Tech & Big Tech
    • Campus & Education
    • Media & Culture
    • Global Free Speech
  • Opinions
    • Debates
  • Video/Live
  • Community
  • Freedom Index
  • About
    • Mission
    • Contact
    • Support
FSNN | Free Speech News NetworkFSNN | Free Speech News Network
Home»Cryptocurrency & Free Speech Finance»Nvidia Drops Nemotron 3 Super Amid $26 Billion Open-Model AI Bet—America’s Answer to Qwen?
Cryptocurrency & Free Speech Finance

Nvidia Drops Nemotron 3 Super Amid $26 Billion Open-Model AI Bet—America’s Answer to Qwen?

News RoomBy News Room2 hours agoNo Comments6 Mins Read249 Views
Share Facebook Twitter Pinterest Copy Link LinkedIn Tumblr Email VKontakte Telegram
Nvidia Drops Nemotron 3 Super Amid  Billion Open-Model AI Bet—America’s Answer to Qwen?
Share
Facebook Twitter Pinterest Email Copy Link

Listen to the article

0:00
0:00

Key Takeaways

Playback Speed

Select a Voice

In brief

  • Nvidia launched Nemotron 3 Super, a 120B open-weight AI model optimized for autonomous agents and ultra-long context tasks.
  • The hybrid Mamba-Transformer MoE architecture delivers faster reasoning and over 5× throughput while running at 4-bit precision.
  • Nvidia’s $26 billion investment into open-source AI wants to counter China’s rise in the field.

Nvidia just shipped Nemotron 3 Super, a 120-billion-parameter open-weight model built to do one thing well: run autonomous AI agents without bleeding your compute budget dry.

That’s not a small problem. Multi-agent systems generate a lot more tokens than a normal chat—every tool call, reasoning step, and slice of context gets re-sent from scratch. As a result, costs explode, models tend to drift, and the agents slowly forget what they were supposed to be doing in the first place… or at least decrease in accuracy.

Nemotron 3 Super is Nvidia’s answer to all of that. The model runs 12 billion active parameters out of 120 billion total, using a mixture-of-experts (MoE) design that keeps inference cheap while retaining the reasoning depth complex workflows need. It packs a 1-million-token context window, so agents can hold an entire codebase, or nearly 750,000 words in memory before collapsing.

To build its model, Nvidia combined three components that rarely appear together in the same architecture: Mamba-2 state-space layers—a faster, memory-efficient alternative to attention for handling long token streams—along with Transformer attention layers for precise recall, and a new “Latent MoE” design that compresses token embeddings before routing them to experts. That allows the model to activate four times as many specialists at the same compute cost.

Introducing NVIDIA Nemotron 3 Super 🎉

Open 120B-parameter (12B active) hybrid Mamba-Transformer MoE model

Native 1M-token context

Built for compute-efficient, high-accuracy multi-agent applications

Plus, fully open weights, datasets and recipes for easy customization and… pic.twitter.com/kMFI23noFc

— NVIDIA AI Developer (@NVIDIAAIDev) March 11, 2026

The model was also pretrained natively in NVFP4, Nvidia’s 4-bit floating-point format. In practice, that means the system learned to operate accurately within 4-bit arithmetic from the very first gradient update, rather than being trained at high precision and compressed afterward, which often causes models to lose accuracy.

For context, a model’s precision is measured in bits. Full precision, known as FP32, is the gold standard—but it is also extremely expensive to run at scale. Developers often reduce precision to save compute while trying to preserve useful performance.

Think of it like shrinking a 4K image down to 1080p: The picture still looks the same at a glance, just with less detail. Normally, dropping from 32-bit precision all the way to 4-bit would cripple a model’s reasoning ability. Nemotron avoids that problem by learning to operate at low precision from the start, instead of being squeezed into it later.

Compared to its own predecessor, Nemotron 3 Super delivers more than five times the throughput. Against external rivals, it’s 2.2x faster than OpenAI’s GPT-OSS 120B on inference throughput, and 7.5x faster than Alibaba’s Qwen3.5-122B.

We ran our own quick test. The reasoning held up well, including on prompts that were deliberately vague, badly worded, or based on wrong information. The model caught small errors in context without being asked to, handled math and logic problems cleanly, and didn’t fall apart when the question itself was slightly off.

The full training pipeline is public: weights on Hugging Face, 10 trillion curated pretraining tokens seen over 25 trillion total during training, 40 million post-training samples, and reinforcement learning recipes across 21 environment configurations. Perplexity, Palantir, Cadence, and Siemens are already integrating the model in their workflows.

The $26 billion bet

The model may be one piece of a larger strategy. A 2025 financial filing shows Nvidia plans to spend $26 billion over the next five years building open-weight AI models. Executives confirmed it, too.

Bryan Catanzaro, VP of applied deep learning research, told Wired the company recently finished pretraining a 550-billion-parameter model. Nvidia released its first Nemotron model back in November 2023, but that filing makes clear this is no longer a side project.

The investment is strategic considering Nvidia’s chips are still the default infrastructure for training and running frontier models. Models tuned to its hardware give customers a built-in reason to stay on Nvidia despite efforts from competitors to use other hardware. But there’s a more urgent pressure behind the move: America is losing the open-source AI race, and losing it fast.

Chinese open models went from barely 1.2% of global open-model usage in late 2024 to roughly 30% by the end of 2025, according to research by OpenRouter and Andreessen Horowitz. Alibaba’s Qwen overtook Meta’s Llama as the most-used self-hosted open-source model, according to Runpod. American companies including Airbnb adopted it for customer service. Startups worldwide are building on top of it. Beyond market share, that kind of adoption creates infrastructure dependencies that are hard to reverse.

While U.S. giants like OpenAI, Anthropic, and Google keep their best models locked behind APIs, Chinese labs from DeepSeek to Alibaba have been flooding the open ecosystem. Meta was the one major American player competing in open source with Llama, but Zuckerberg recently signaled the company might not make future models fully open.

The gap between “best proprietary model” and “best open model” used to be massive—and in America’s favor. That gap is now very small, and the open side of the ledger is increasingly Chinese.

Incredible graph. In just one year, China completely overtook the U.S. in free AI models.

Not a single U.S. model in the top 5 today when last year the top 3 were all American. pic.twitter.com/34ErpBv8rg

— Arnaud Bertrand (@RnaudBertrand) October 14, 2025

There’s also a hardware threat underneath all of this. A new DeepSeek model is widely expected to drop soon, and it’s rumored to have been trained entirely on chips made by Huawei—a sanctioned Chinese company. If that’s confirmed, then it would give developers around the world, particularly in China, a concrete reason to start testing Huawei’s hardware. China’s Ziphu AI is already doing that.

That’s the scenario Nvidia most needs to prevent: Chinese open models and Chinese chips building an ecosystem that doesn’t need Nvidia at all.

Daily Debrief Newsletter

Start every day with the top news stories right now, plus original features, a podcast, videos and more.



Read the full article here

Fact Checker

Verify the accuracy of this article using AI-powered analysis and real-time sources.

Get Your Fact Check Report

Enter your email to receive detailed fact-checking analysis

5 free reports remaining

Continue with Full Access

You've used your 5 free reports. Sign up for unlimited access!

Already have an account? Sign in here

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Telegram Copy Link
News Room
  • Website
  • Facebook
  • X (Twitter)
  • Instagram
  • LinkedIn

The FSNN News Room is the voice of our in-house journalists, editors, and researchers. We deliver timely, unbiased reporting at the crossroads of finance, cryptocurrency, and global politics, providing clear, fact-driven analysis free from agendas.

Related Articles

Media & Culture

Tax the Rich All You Want. It Won’t Fix the Deficit.

16 minutes ago
Legal & Courts

Journalist Marisa Kabas discusses her successful fight for bodycam videos of DOGE raid at US Institute of Peace

33 minutes ago
Cryptocurrency & Free Speech Finance

BTC mining faces price risk, not power cost shock, as oil tops $100

41 minutes ago
Cryptocurrency & Free Speech Finance

Binance Claims ‘Full and Complete Legal Victory‘ in Alabama Court

42 minutes ago
Cryptocurrency & Free Speech Finance

Optimism Team Lays Off 20 Employees Amid Ethereum Scaling Shifts, Base Migration Plans

45 minutes ago
Media & Culture

Don’t Ban Kids From Using Chatbots

1 hour ago
Add A Comment
Leave A Reply Cancel Reply

Editors Picks

Tax the Rich All You Want. It Won’t Fix the Deficit.

16 minutes ago

Journalist Marisa Kabas discusses her successful fight for bodycam videos of DOGE raid at US Institute of Peace

33 minutes ago

BTC mining faces price risk, not power cost shock, as oil tops $100

41 minutes ago

Binance Claims ‘Full and Complete Legal Victory‘ in Alabama Court

42 minutes ago
Latest Posts

Optimism Team Lays Off 20 Employees Amid Ethereum Scaling Shifts, Base Migration Plans

45 minutes ago

Don’t Ban Kids From Using Chatbots

1 hour ago

The Trump Administration Just Declared All Foreign Exports Unfair

1 hour ago

Subscribe to News

Get the latest news and updates directly to your inbox.

At FSNN – Free Speech News Network, we deliver unfiltered reporting and in-depth analysis on the stories that matter most. From breaking headlines to global perspectives, our mission is to keep you informed, empowered, and connected.

FSNN.net is owned and operated by GlobalBoost Media
, an independent media organization dedicated to advancing transparency, free expression, and factual journalism across the digital landscape.

Facebook X (Twitter) Discord Telegram
Latest News

A.B. 1043’s Internet Age Gates Hurt Everyone

14 minutes ago

Tax the Rich All You Want. It Won’t Fix the Deficit.

16 minutes ago

Journalist Marisa Kabas discusses her successful fight for bodycam videos of DOGE raid at US Institute of Peace

33 minutes ago

Subscribe to Updates

Get the latest news and updates directly to your inbox.

© 2026 GlobalBoost Media. All Rights Reserved.
  • Privacy Policy
  • Terms of Service
  • Our Authors
  • Contact

Type above and press Enter to search. Press Esc to cancel.

🍪

Cookies

We and our selected partners wish to use cookies to collect information about you for functional purposes and statistical marketing. You may not give us your consent for certain purposes by selecting an option and you can withdraw your consent at any time via the cookie icon.

Cookie Preferences

Manage Cookies

Cookies are small text that can be used by websites to make the user experience more efficient. The law states that we may store cookies on your device if they are strictly necessary for the operation of this site. For all other types of cookies, we need your permission. This site uses various types of cookies. Some cookies are placed by third party services that appear on our pages.

Your permission applies to the following domains:

  • https://fsnn.net
Necessary
Necessary cookies help make a website usable by enabling basic functions like page navigation and access to secure areas of the website. The website cannot function properly without these cookies.
Statistic
Statistic cookies help website owners to understand how visitors interact with websites by collecting and reporting information anonymously.
Preferences
Preference cookies enable a website to remember information that changes the way the website behaves or looks, like your preferred language or the region that you are in.
Marketing
Marketing cookies are used to track visitors across websites. The intention is to display ads that are relevant and engaging for the individual user and thereby more valuable for publishers and third party advertisers.