Friday, 1 May 2026
Subscribe
logo
  • AI Compute
  • Infrastructure
  • Power & Cooling
  • Security
  • Colocation
  • Cloud Computing
  • More
    • Sustainability
    • Industry News
    • About Data Center News
    • Terms & Conditions
Font ResizerAa
Data Center NewsData Center News
Search
  • AI Compute
  • Infrastructure
  • Power & Cooling
  • Security
  • Colocation
  • Cloud Computing
  • More
    • Sustainability
    • Industry News
    • About Data Center News
    • Terms & Conditions
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > AI & Compute > Cerebras becomes the world’s fastest host for DeepSeek R1, outpacing Nvidia GPUs by 57x
AI & Compute

Cerebras becomes the world’s fastest host for DeepSeek R1, outpacing Nvidia GPUs by 57x

Last updated: February 2, 2025 11:22 am
Published February 2, 2025
Share
Cerebras becomes the world’s fastest host for DeepSeek R1, outpacing Nvidia GPUs by 57x
SHARE

Be a part of our each day and weekly newsletters for the newest updates and unique content material on industry-leading AI protection. Be taught Extra


Cerebras Systems introduced as we speak it’s going to host DeepSeek’s breakthrough R1 artificial intelligence model on U.S. servers, promising speeds as much as 57 instances sooner than GPU-based options whereas holding delicate knowledge inside American borders. The transfer comes amid rising considerations about China’s fast AI development and knowledge privateness.

The AI chip startup will deploy a 70-billion-parameter model of DeepSeek-R1 operating on its proprietary wafer-scale {hardware}, delivering 1,600 tokens per second — a dramatic enchancment over conventional GPU implementations which have struggled with newer “reasoning” AI fashions.

Response instances of main AI platforms, measured in seconds. Cerebras achieves the quickest response at simply over one second, whereas Novita’s system takes almost 38 seconds to generate its first output — a crucial metric for real-world purposes. (Supply: Synthetic Evaluation)

Why DeepSeek’s reasoning fashions are reshaping enterprise AI

“These reasoning fashions have an effect on the economic system,” stated James Wang, a senior govt at Cerebras, in an unique interview with VentureBeat. “Any data employee principally has to do some type of multi-step cognitive duties. And these reasoning fashions would be the instruments that enter their workflow.”

The announcement follows a tumultuous week by which DeepSeek’s emergence triggered Nvidia’s largest-ever market value loss, nearly $600 billion, elevating questions in regards to the chip large’s AI supremacy. Cerebras’ resolution straight addresses two key considerations which have emerged: the computational calls for of superior AI fashions, and knowledge sovereignty.

“If you happen to use DeepSeek’s API, which could be very fashionable proper now, that knowledge will get despatched straight to China,” Wang defined. “That’s one extreme caveat that [makes] many U.S. corporations and enterprises…not keen to think about [it].”

See also  Zuckerberg's $15B talent war explained
Cerebras demonstrates dramatic efficiency benefits in output velocity, processing 1,508 tokens per second — almost six instances sooner than its closest competitor, Groq, and roughly 100 instances sooner than conventional GPU-based options like Novita. (Supply: Synthetic Evaluation)

How Cerebras’ wafer-scale know-how beats conventional GPUs at AI velocity

Cerebras achieves its velocity benefit by way of a novel chip structure that retains complete AI fashions on a single wafer-sized processor, eliminating the reminiscence bottlenecks that plague GPU-based methods. The corporate claims its implementation of DeepSeek-R1 matches or exceeds the efficiency of OpenAI’s proprietary fashions, whereas operating fully on U.S. soil.

The event represents a major shift within the AI panorama. DeepSeek, based by former hedge fund govt Liang Wenfeng, shocked the {industry} by reaching subtle AI reasoning capabilities reportedly at simply 1% of the price of U.S. rivals. Cerebras’ internet hosting resolution now provides American corporations a solution to leverage these advances whereas sustaining knowledge management.

“It’s really a pleasant story that the U.S. analysis labs gave this present to the world. The Chinese language took it and improved it, however it has limitations as a result of it runs in China, has some censorship issues, and now we’re taking it again and operating it on U.S. knowledge facilities, with out censorship, with out knowledge retention,” Wang stated.

Efficiency benchmarks exhibiting DeepSeek-R1 operating on Cerebras outperforming each GPT-4o and OpenAI’s o1-mini throughout query answering, mathematical reasoning, and coding duties. The outcomes counsel Chinese language AI improvement could also be approaching or surpassing U.S. capabilities in some areas. (Credit score: Cerebras)

U.S. tech management faces new questions as AI innovation goes world

The service will likely be obtainable by way of a developer preview beginning as we speak. Whereas it will likely be initially free, Cerebras plans to implement API access controls because of sturdy early demand.

The transfer comes as U.S. lawmakers grapple with the implications of DeepSeek’s rise, which has uncovered potential limitations in American trade restrictions designed to keep up technological benefits over China. The flexibility of Chinese language corporations to realize breakthrough AI capabilities regardless of chip export controls has prompted calls for brand new regulatory approaches.

See also  How CrowdStrike's 78-minute outage reshaped enterprise cybersecurity

Business analysts counsel this improvement may speed up the shift away from GPU-dependent AI infrastructure. “Nvidia is now not the chief in inference efficiency,” Wang famous, pointing to benchmarks exhibiting superior efficiency from numerous specialised AI chips. “These different AI chip corporations are actually sooner than GPUs for operating these newest fashions.”

The affect extends past technical metrics. As AI fashions more and more incorporate subtle reasoning capabilities, their computational calls for have skyrocketed. Cerebras argues its structure is healthier suited to these rising workloads, probably reshaping the aggressive panorama in enterprise AI deployment.


Source link
TAGGED: 57x, Cerebras, DeepSeek, Fastest, GPUs, Host, Nvidia, outpacing, Worlds
Share This Article
Twitter Email Copy Link Print
Previous Article Survive until 25? We're off to a rough start | The DeanBeat Survive until 25? We’re off to a rough start | The DeanBeat
Next Article These Yale and Berkeley dropouts just raised $2 million to build an AI assistant that could rival OpenAI These Yale and Berkeley dropouts just raised $2 million to build an AI assistant that could rival OpenAI
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

‘Studio Ghibli’ AI image trend overwhelms OpenAI’s new GPT-4o feature, delaying free tier

Be part of our every day and weekly newsletters for the newest updates and unique…

March 27, 2025

CyrusOne breaks ground on first Italian data centre

CyrusOne has formally damaged floor on its first knowledge heart in Italy. The MIL1 facility,…

April 15, 2026

AMD Strikes Blow in AI Chip War With OpenAI Deal

Superior Micro Gadgets (AMD) on Monday struck a large take care of OpenAI for six…

October 6, 2025

Scaling smarter: How enterprise IT teams can right-size their compute for AI

This text is a part of VentureBeat’s particular situation, “The Actual Price of AI: Efficiency,…

July 5, 2025

Microsoft’s new rStar-Math technique upgrades small models to outperform OpenAI’s o1-preview at math problems

Be a part of our each day and weekly newsletters for the most recent updates…

January 10, 2025

You Might Also Like

STL launches Neuralis data centre connectivity suite in the U.S.
AI & Compute

STL launches Neuralis data centre connectivity suite in the U.S.

By saad
What is optical interconnect and why Lightelligence's $10B debut says it matters for AI
AI & Compute

What is optical interconnect and why Lightelligence’s $10B debut says it matters for AI

By saad
IBM launches AI platform Bob to regulate SDLC costs
AI & Compute

IBM launches AI platform Bob to regulate SDLC costs

By saad
The evolution of encoders: From simple models to multimodal AI
AI & Compute

The evolution of encoders: From simple models to multimodal AI

By saad

About Us

Data Center News is your dedicated source for data center infrastructure, AI compute, cloud, and industry news.

Top Categories

  • AI & Compute
  • Cloud Computing
  • Power & Cooling
  • Colocation
  • Security
  • Infrastructure
  • Sustainability
  • Industry News

Useful Links

  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

Find Us on Socials

© 2026 Data Center News. All Rights Reserved.

© 2026 Data Center News. All Rights Reserved.
Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.