Tuesday, 28 Apr 2026
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > Global Market > Nvidia targets inference as AI’s next battleground with Groq 3 LPX
Global Market

Nvidia targets inference as AI’s next battleground with Groq 3 LPX

Last updated: March 18, 2026 4:56 am
Published March 18, 2026
Share
Nvidia high-performance chip technology
SHARE

Contents
The subsequent section of inferencingCoaching versus inferencing

It’s a giant value play, he identified, and it “has to occur in every single place, on a regular basis, for all customers.”

The subsequent section of inferencing

The brand new Groq 3 language processing models (LPUs) are primarily based on mental property (IP) from Groq, which signed a $20 billion licensing settlement with Nvidia late final 12 months. In response to the chip firm, a fleet of LPUs can operate as a “large single processor.”

Whereas Rubin GPUs will proceed to deal with prefill (immediate processing), Groq’s LPX will now deal with latency-sensitive parts of decode (response). Collectively, they will ship a “new class of inference efficiency,” Nvidia says. 

Every LPX rack options 256 LPUs with 128 GB of on-chip static random-access reminiscence (SRAM), 150 terabyte per second (TB/s) bandwidth, chip-to-chip hyperlinks and high-speed connections to NVL72, Nvidia’s liquid-cooled AI supercomputer. Mixed, these can scale back latency to “close to zero,” Nvidia claims.

The LPX integration with Vera Rubin AI factories will probably be accessible within the second half of this 12 months.

Coaching versus inferencing

Coaching and inference stress infrastructure in very alternative ways, famous Sanchit Vir Gogia, chief analyst at Greyhound Analysis. Whereas coaching rewards “huge parallelism and brute-force scale,” inferencing (particularly for lengthy context and interactive reasoning) is much extra delicate to latency, reminiscence motion, cache conduct, concurrency, and cost per delivered token.

Source link

See also  ceτi AI Secures $80M of NVIDIA GPUs and Hardware to Expand Global AI Infrastructure Network by 25x
TAGGED: AIs, battleground, Groq, Inference, LPX, Nvidia, targets
Share This Article
Twitter Email Copy Link Print
Previous Article The UK’s data centre pipeline continues to soar The UK’s data centre pipeline continues to soar
Next Article Cloud demand shifts toward AI as enterprise usage deepens Cloud demand shifts toward AI as enterprise usage deepens
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Augmenting Data Center Revenue with Grid Services: What to Know | DCN

With the expansion of information facilities throughout the nation, operators are beneath strain to scale…

April 25, 2024

Zilliz Cloud boosts vector database performance

San Francisco-based Zilliz has released a new version of its database-as-a-service (DBaaS) offering, Zilliz Cloud.…

February 3, 2024

Singapore to Free Up More Power for Data Center Expansions

(Bloomberg) --Singapore goals to extend the quantity of power it allocates for knowledge facilities by…

May 30, 2024

Engineers develop a fully 3D-printed electrospray engine that can power tiny satellites

MIT engineers have demonstrated the primary totally 3D-printed, droplet-emitting electrospray engine. The gadget, which might…

February 12, 2025

Miami International Holdings Receives $100M Investment from Warburg Pincus

Miami International Holdings, Inc., a technology-driven chief in constructing and working regulated monetary markets throughout…

August 25, 2024

You Might Also Like

Cisco
Global Market

Infected Cisco firewalls need cold start to clear persistent Firestarter backdoor

By saad
What to expect from DCR Live 2026
Global Market

What to expect from DCR Live 2026

By saad
AI活用は社員が嫌いな仕事から始めよ
Global Market

Nvidia’s ‘AI insurance policy’ balances immediate and future AI approaches

By saad
Why legacy data centre networks are no longer fit for purpose
Global Market

Why legacy data centre networks are no longer fit for purpose

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.