Tuesday, 14 Apr 2026
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > Global Market > Nvidia rolls out new GPUs for AI inferencing, large workloads
Global Market

Nvidia rolls out new GPUs for AI inferencing, large workloads

Last updated: September 14, 2025 10:29 pm
Published September 14, 2025
Share
Rethinking IT for the AI era
SHARE

Rubin has two dies with 25 petaFLOPs per die, NVLink interconnect and 288GB of HBM4 high-speed reminiscence. The Rubin CPX has one die with 30 petaFLOPS of efficiency, no NVLink and 128GB of GDDR7 reminiscence. So Rubin CPX is perfect for particular excessive context wants that don’t want lots of reminiscence. CPX will probably be cheaper than the usual Rubin however Nvidia wouldn’t say how a lot.

To course of video, AI fashions can take as much as a million tokens for an hour of content material, which may take many hours if not days to generate. The extra tokens the system can generate, the bigger scale processing it could possibly do.

Rubin CPX delivers as much as 30 petaflops of compute with NVFP4 precision. It options 128GB of GDDR7 reminiscence somewhat than the same old HBM reminiscence, which is costlier than GDDR7. Nvidia says that the GDDR7 has sufficient efficiency, and that Rubin CPX delivers thrice quicker consideration capabilities in contrast with GB300 NVL72 techniques.

Rubin CPX is obtainable in a number of configurations, together with the Vera Rubin NVL144 CPX, that may be mixed with the Quantum‐X800 InfiniBand scale-out compute cloth or the Spectrum-XTM Ethernet networking platform with Nvidia Spectrum-XGS Ethernet expertise and Nvidia ConnectX-9 SuperNICs.

Nvidia can also be saying a brand new Vera Rubin NVL 144 CPX rack. Narasimhan mentioned the NVL 144 CPX permits AI service suppliers to dramatically enhance their profitability by delivering $5 billion of income for each $100 million invested in infrastructure.

It is available in two configurations: single rack, with 144 Rubin CPX GPUs, 144 Rubin GPUs, and 36 Vera CPUs for 8 exaFLOPs of NVFP4 compute and 100TB of quick reminiscence and 1.7 PB/s of reminiscence bandwidth. Nvidia mentioned it’s 7.5 occasions quicker than the present top-of-the-line GB300 NVL72.

See also  Turning Network Issues into Strategic Business Opportunities

Source link

TAGGED: GPUs, inferencing, large, Nvidia, Rolls, Workloads
Share This Article
Twitter Email Copy Link Print
Previous Article PsiQuantum Secures $1B to Build Fault-Tolerant Million-Qubit Computers PsiQuantum Secures $1B to Build Fault-Tolerant Million-Qubit Computers
Next Article five takeaways from the Munich auto show five takeaways from the Munich auto show
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

HPEJ, JHPE or what? | Network World

The issue is that lower than one-eighth of enterprises stated that they believed HPE may…

August 20, 2025

Baidu undercuts rival AI models with ERNIE 4.5 and ERNIE X1

Baidu has launched its newest basis AI fashions, ERNIE 4.5 and ERNIE X1, and is…

March 17, 2025

OpenAI data residency advances enterprise AI governance

For chief information and data officers, particularly in tightly regulated sectors, information governance has been…

October 26, 2025

AI Computing Is on Pace to Consume More Energy Than India, Arm Says | DCN

(Bloomberg) -- AI’s voracious want for computing energy is threatening to overwhelm power sources, requiring…

April 17, 2024

New Data Center Developments: February 2025

The demand for brand spanking new information facilities isn’t exhibiting any signal of slowing. With…

February 6, 2025

You Might Also Like

linux code binary
Global Market

Linux 7.0 debuts with some big changes for networking

By saad
Dumbbells as models like Google Gemma 4 require stronger enterprise AI governance by CISOs as they scramble to secure edge workloads.
AI

Strengthening enterprise governance for rising edge AI workloads

By saad
How do data centres scale for AI when power is the constraint?
Global Market

How do data centres scale for AI when power is the constraint?

By saad
Lambda doubles down on NVIDIA stack with 10,000+ Blackwell GPUs and CPO networking push
Edge Computing

Lambda doubles down on NVIDIA stack with 10,000+ Blackwell GPUs and CPO networking push

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.