Wednesday, 8 Apr 2026
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > Global Market > Nvidia targets inference as AI’s next battleground with Groq 3 LPX
Global Market

Nvidia targets inference as AI’s next battleground with Groq 3 LPX

Last updated: March 18, 2026 4:56 am
Published March 18, 2026
Share
Nvidia high-performance chip technology
SHARE

Contents
The subsequent section of inferencingCoaching versus inferencing

It’s a giant value play, he identified, and it “has to occur in every single place, on a regular basis, for all customers.”

The subsequent section of inferencing

The brand new Groq 3 language processing models (LPUs) are primarily based on mental property (IP) from Groq, which signed a $20 billion licensing settlement with Nvidia late final 12 months. In response to the chip firm, a fleet of LPUs can operate as a “large single processor.”

Whereas Rubin GPUs will proceed to deal with prefill (immediate processing), Groq’s LPX will now deal with latency-sensitive parts of decode (response). Collectively, they will ship a “new class of inference efficiency,” Nvidia says. 

Every LPX rack options 256 LPUs with 128 GB of on-chip static random-access reminiscence (SRAM), 150 terabyte per second (TB/s) bandwidth, chip-to-chip hyperlinks and high-speed connections to NVL72, Nvidia’s liquid-cooled AI supercomputer. Mixed, these can scale back latency to “close to zero,” Nvidia claims.

The LPX integration with Vera Rubin AI factories will probably be accessible within the second half of this 12 months.

Coaching versus inferencing

Coaching and inference stress infrastructure in very alternative ways, famous Sanchit Vir Gogia, chief analyst at Greyhound Analysis. Whereas coaching rewards “huge parallelism and brute-force scale,” inferencing (particularly for lengthy context and interactive reasoning) is much extra delicate to latency, reminiscence motion, cache conduct, concurrency, and cost per delivered token.

Source link

See also  Namla pairs with NVIDIA to unlock streamlined deployment of edge AI
TAGGED: AIs, battleground, Groq, Inference, LPX, Nvidia, targets
Share This Article
Twitter Email Copy Link Print
Previous Article The UK’s data centre pipeline continues to soar The UK’s data centre pipeline continues to soar
Next Article Cloud demand shifts toward AI as enterprise usage deepens Cloud demand shifts toward AI as enterprise usage deepens
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Vertiv launches one-day installation package for AI data center systems

Knowledge middle infrastructure vendor Vertiv has launched Vertiv OneCore, a totally modular information middle constructing…

August 26, 2025

A sector shaping the global landscape

Pieter Schaap, Director at Soben, considers a few of the greatest developments taking place within…

March 19, 2024

RemotePass Raises $5.5M in Series A Funding

RemotePass, an Abu Dhabi, UAE-based supplier of a HR and fintech platform, raised $5.5M in…

March 5, 2024

Keysight’s FITS-8CH enhances validation for AI data centres

Keysight Applied sciences has launched the FITS-8CH platform, an answer designed to assist the reliability…

March 24, 2026

Aim Security, Interview With CEO Matan Getz

Matan Getz, CEO, Aim Security Aim Security, a Tel Aviv, Israel-based cybersecurity startup offering enterprises…

January 31, 2024

You Might Also Like

Stellanor expands to 11 UK data centres with Imagination Technologies deal
Global Market

Stellanor expands to 11 UK data centres with Imagination Technologies deal

By saad
Intel
Global Market

Intel bets on Terafab to help it reassert itself in the AI chip race

By saad
Security, laser shield
Global Market

Cisco joins Anthropic’s multivendor effort to secure AI software

By saad
NVIDIA and T-Mobile push AI-RAN to turn 5G networks into distributed edge compute platforms
Edge Computing

NVIDIA and T-Mobile push AI-RAN to turn 5G networks into distributed edge compute platforms

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.