Saturday, 11 Apr 2026
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > Cloud Computing > Cerebras Introduces ‘World’s Fastest AI Chip’ and New AI Server | DCN
Cloud Computing

Cerebras Introduces ‘World’s Fastest AI Chip’ and New AI Server | DCN

Last updated: March 13, 2024 7:28 pm
Published March 13, 2024
Share
The Cerebras WSE-3 chip features four million transistors and 900,000 cores
SHARE

AI {hardware} startup Cerebras Techniques has launched a brand new, third-generation AI processor that it claims to be the quickest on the earth. The WSE-3 chip doubles the efficiency of its predecessor, which was the earlier file holder, the corporate mentioned at present (March 13).

“As soon as once more, we’ve delivered the most important and quickest AI chip on the planet with the identical dinner plate-size kind issue,” mentioned Andy Hock, Cerebras’ vice chairman of product administration.

Associated: Key Developments and Applied sciences Impacting Information Facilities in 2024 and Past

The Sunnyvale, California-based startup entered the {hardware} market in 2019 when it launched a super-sized AI chip, known as the Wafer Scale Engine (WSE), which measured eight inches by eight inches. It was 56 instances bigger than the biggest GPU and featured 1.2 trillion transistors and 400,000 computing cores, making it the quickest and largest AI chip out there on the time.

Then in 2021, Cerebras launched the WSE-2, a 7-nanometer chip that doubled the efficiency of the unique with 2.6 trillion transistors and 850,000 cores.

900,000 Cores

Associated: Naked-Steel vs. Devoted Servers: Completely different Names for the Similar Factor?

The corporate at present almost doubled efficiency once more with the WSE-3 chip, which options 4 million transistors and 900,000 cores, delivering 125 petaflops of efficiency. The brand new 5-nanometer processor powers Cerebras’ new CS-3 AI server, which is designed to coach the biggest AI fashions.

“The CS-3 is an enormous step ahead for us,” Hock informed DCN. “It’s two instances extra efficiency than our CS-2 [server]. So, it’s two instances sooner coaching for giant AI fashions with the identical energy draw, and it’s out there on the similar value [as the CS-2] to our clients.”  

See also  7 Things to Consider When Choosing a Server Antivirus Program

Since its launch, Cerebras has positioned itself as a substitute for Nvidia GPU-powered AI programs. The startup’s pitch: as a substitute of utilizing 1000’s of GPUs, they will run their AI coaching on Cerebras {hardware} utilizing considerably fewer chips.

“One [Cerebras] server can do the identical work as 10 racks of GPUs,” mentioned Karl Freund, founder and principal analyst of Cambrian AI Analysis.

Cerebras

The WSE-3 processor powers Cerebras’ new CS-3 AI server, which is designed to coach the biggest AI fashions

Cerebras Makes Inroads Into AI Market

Nvidia dominates the AI market with its GPUs capturing about 85% of the AI chip market, whereas the remaining gamers reminiscent of AMD, Intel, Google, AWS, Microsoft, Cerebras and others have captured about 15%, the analyst mentioned.

Whereas the competitors has not but confirmed that they will steal an enormous chunk of market share from Nvidia, Cerebras has discovered success because it launched its first product 5 years in the past, mentioned Freund, who calls Cerebras probably the most profitable AI startup at present.

“From the start, Cerebras took a really totally different method,” he mentioned. “Everyone else is making an attempt to outdo Nvidia, which is actually laborious to do. Cerebras mentioned, ‘We’re going to construct a complete wafer-scale AI engine,’ which nobody has ever performed. The profit is extremely excessive efficiency.”

Cloud Entry

Cerebras doesn’t earn a living promoting its processors. It makes cash promoting servers that run on these chips, which, in accordance with an organization spokesperson, value thousands and thousands of {dollars} every. Cerebras makes its CS-3 programs out there to clients over the cloud, nevertheless it additionally sells to massive enterprises, authorities companies, and worldwide cloud suppliers.

See also  Four Strategies to Meet Changing Data Center Demands | DCN

For instance, Cerebras just lately added healthcare supplier Mayo Clinic to its rising roster of shoppers, which incorporates Argonne Nationwide Laboratory and pharmaceutical large GlaxoSmithKline.  

Cerebras in July 2023 additionally introduced it inked a $100 million deal to construct the primary of 9 interconnected, cloud-based AI supercomputers for G42, a expertise holding group based mostly within the United Arab Emirates.

Since then, the 2 firms have built two supercomputers totaling eight exaflops of AI compute. Accessible over the cloud, the supercomputers are optimized for coaching massive language fashions and generative AI fashions and are being utilized by organizations throughout totally different industries for local weather, well being and vitality analysis and different initiatives.

Cerebras and G42 are at the moment constructing a 3rd supercomputer, the Condor Galaxy 3 in Dallas, which will probably be powered by 64 CS-3 programs and can produce eight exaflops of AI compute. By the top of 2024, the businesses plan to finish the 9 supercomputers, which is able to whole 55.6 exaflops of compute.

“The truth that Cerebras has now produced a third-generation Wafer Scale Engine is a testomony to its buyer traction. They generated the form of income they wanted to pay for all that engineering,” Freund mentioned.

In Numbers: WSE-3 Chip and CS-3 AI System

Cerebras’ WSE-3 options 52 instances extra cores than Nvidia’s H100 Tensor Core. When in comparison with an Nvidia DGX H100 system, the Cerebras CS-3 system – powered by the WSE-3 chip – performs coaching eight instances sooner, options 1,900 instances extra reminiscence and might prepare AI fashions as much as 24 trillion parameters, which is 600 instances bigger than a DGX H100’s capabilities, Cerebras executives mentioned.  

See also  Virginia Faces New Headwinds in Data Center Growth

A Llama 70 billion parameter mannequin that takes 30 days to coach on GPUs may be educated in at some point utilizing a CS-3 cluster, Hock mentioned.

Cerebras Companions with Qualcomm on AI Inferencing

As a result of Cerebras’ {hardware} focuses on AI coaching, it beforehand didn’t have a solution for purchasers’ AI inferencing wants. Now it does because of a brand new partnership with Qualcomm.

The 2 firms at present mentioned they’ve collaborated, in order that the fashions educated on Cerebras’ {hardware} are optimized to run inferencing on Qualcomm’s Cloud A100 Extremely accelerator.

“They optimized the output of the large CS-3 machines to run rather well on these very low-cost, low-power Qualcomm AI inferencing engines,” Freund mentioned.

Source link

Contents
900,000 CoresCerebras Makes Inroads Into AI MarketCloud EntryIn Numbers: WSE-3 Chip and CS-3 AI SystemCerebras Companions with Qualcomm on AI Inferencing
TAGGED: Cerebras, Chip, DCN, Fastest, introduces, Server, Worlds
Share This Article
Twitter Email Copy Link Print
Previous Article Roku hackers breach 15,000 accounts and are selling them online Roku hackers breach 15,000 accounts and are selling them online
Next Article smart cities Pioneering the era of smart cities
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Powering the AI Revolution By Quiver Quantitative

© Reuters. Blackstone's $25 Billion Data Center Empire: Powering the AI Revolution Quiver Quantitative -…

January 30, 2024

Rerun Raises $17M in Seed Funding

Rerun, a Stockholm, Sweden-based firm constructing a multimodal knowledge stack for Bodily AI, raised $17M…

March 22, 2025

Forecasting water shortages could be made harder by AI boom

The speedy progress of AI is leaving the Setting Company unable to precisely forecast future…

June 19, 2025

AI wars heat up: OpenAI’s SearchGPT takes on Google’s search dominance

Be part of our every day and weekly newsletters for the newest updates and unique…

July 28, 2024

Cooking.City Bringing Back Value Redistribution to Solana Fair Launches

Hong Kong, Hong Kong, July third, 2025, Chainwire Cooking.Metropolis, a lately launched fair-launch platform, emphasizes…

July 3, 2025

You Might Also Like

Uber expands use of AWS chips for AI workloads
Cloud Computing

Uber expands use of AWS chips for AI workloads

By saad
Tanium introduces autonomy driven by AI and security innovation
Cloud Computing

Tanium introduces autonomy driven by AI and security innovation

By saad
Intel
Global Market

Intel bets on Terafab to help it reassert itself in the AI chip race

By saad
Why Google is building more data centres in Asia
Cloud Computing

Why Google is building more data centres in Asia

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.