Friday, 1 May 2026
Subscribe
logo
  • AI Compute
  • Infrastructure
  • Power & Cooling
  • Security
  • Colocation
  • Cloud Computing
  • More
    • Sustainability
    • Industry News
    • About Data Center News
    • Terms & Conditions
Font ResizerAa
Data Center NewsData Center News
Search
  • AI Compute
  • Infrastructure
  • Power & Cooling
  • Security
  • Colocation
  • Cloud Computing
  • More
    • Sustainability
    • Industry News
    • About Data Center News
    • Terms & Conditions
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > AI & Compute > Ant Group uses domestic chips to train AI models and cut costs
AI & Compute

Ant Group uses domestic chips to train AI models and cut costs

Last updated: April 3, 2025 8:04 pm
Published April 3, 2025
Share
Ant Group uses domestic chips to train AI models and cut costs
SHARE

Ant Group is counting on Chinese language-made semiconductors to coach synthetic intelligence fashions to scale back prices and reduce dependence on restricted US know-how, in response to folks acquainted with the matter.

The Alibaba-owned firm has used chips from home suppliers, together with these tied to its mum or dad, Alibaba, and Huawei Applied sciences to coach massive language fashions utilizing the Combination of Consultants (MoE) methodology. The outcomes had been reportedly corresponding to these produced with Nvidia’s H800 chips, sources declare. Whereas Ant continues to make use of Nvidia chips for a few of its AI improvement, one sources stated the corporate is popping more and more to options from AMD and Chinese language chip-makers for its newest fashions.

The event indicators Ant’s deeper involvement within the rising AI race between Chinese language and US tech companies, significantly as corporations search for cost-effective methods to coach fashions. The experimentation with home {hardware} displays a broader effort amongst Chinese language companies to work round export restrictions that block entry to high-end chips like Nvidia’s H800, which, though not probably the most superior, continues to be one of many extra highly effective GPUs accessible to Chinese language organisations.

Ant has revealed a analysis paper describing its work, stating that its fashions, in some checks, carried out higher than these developed by Meta. Bloomberg News, which initially reported the matter, has not verified the corporate’s outcomes independently. If the fashions carry out as claimed, Ant’s efforts could symbolize a step ahead in China’s try to decrease the price of working AI functions and scale back the reliance on overseas {hardware}.

See also  The Hidden Costs of AI: Securing Inference in an Age of Attacks

MoE fashions divide duties into smaller knowledge units dealt with by separate elements, and have gained consideration amongst AI researchers and knowledge scientists. The method has been utilized by Google and the Hangzhou-based startup, DeepSeek. The MoE idea is just like having a staff of specialists, every dealing with a part of a process to make the method of manufacturing fashions extra environment friendly. Ant has declined to touch upon its work with respect to its {hardware} sources.

Coaching MoE fashions relies on high-performance GPUs which will be too costly for smaller corporations to amass or use. Ant’s analysis centered on lowering that price barrier. The paper’s title is suffixed with a transparent goal: Scaling Fashions “with out premium GPUs.” [our quotation marks]

The path taken by Ant and the usage of MoE to scale back coaching prices distinction with Nvidia’s strategy. CEO Officer Jensen Huang has stated that demand for computing energy will proceed to develop, even with the introduction of extra environment friendly fashions like DeepSeek’s R1. His view is that corporations will search extra highly effective chips to drive income development, reasonably than aiming to chop prices with cheaper options. Nvidia’s technique stays centered on constructing GPUs with extra cores, transistors, and reminiscence.

Based on the Ant Group paper, coaching one trillion tokens – the essential models of knowledge AI fashions use to be taught – price about 6.35 million yuan (roughly $880,000) utilizing typical high-performance {hardware}. The corporate’s optimised coaching methodology diminished that price to round 5.1 million yuan by utilizing lower-specification chips.

See also  Agentic AI is all about the context — engineering, that is

Ant stated it plans to use its fashions produced on this approach – Ling-Plus and Ling-Lite – to industrial AI use circumstances like healthcare and finance. Earlier this yr, the corporate acquired Haodf.com, a Chinese language on-line medical platform, to additional Ant’s ambition to deploy AI-based options in healthcare. It additionally operates different AI companies, together with a digital assistant app referred to as Zhixiaobao and a monetary advisory platform generally known as Maxiaocai.

“For those who discover one level of assault to beat the world’s finest kung fu grasp, you may nonetheless say you beat them, which is why real-world software is vital,” stated Robin Yu, chief know-how officer of Beijing-based AI agency, Shengshang Tech.

Ant has made its fashions open supply. Ling-Lite has 16.8 billion parameters – settings that assist decide how a mannequin capabilities – whereas Ling-Plus has 290 billion. For comparability, estimates recommend closed-source GPT-4.5 has round 1.8 trillion parameters, in response to MIT Know-how Assessment.

Regardless of progress, Ant’s paper famous that coaching fashions stays difficult. Small changes to {hardware} or mannequin construction throughout mannequin coaching generally resulted in unstable efficiency, together with spikes in error charges.

(Photograph by Unsplash)

See additionally: DeepSeek V3-0324 tops non-reasoning AI fashions in open-source first

Wish to be taught extra about AI and large knowledge from trade leaders? Try AI & Big Data Expo going down in Amsterdam, California, and London. The great occasion is co-located with different main occasions together with Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Discover different upcoming enterprise know-how occasions and webinars powered by TechForge here.

See also  Cryptocurrency markets a testbed for AI forecasting models

Source link

TAGGED: Ant, Chips, Costs, Cut, Domestic, Group, models, train
Share This Article
Twitter Email Copy Link Print
Previous Article Pulsant to acquire two data centres from SCC Pulsant to acquire two data centres from SCC
Next Article U.S. Energy Chief Plans to Use Federal Land to Build Data Centers U.S. Energy Chief Plans to Use Federal Land to Build Data Centers
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Mistral AI launches Devstral, powerful new open source SWE agent model that runs on laptops

Be a part of our every day and weekly newsletters for the newest updates and…

May 25, 2025

Deploying agentic finance AI for immediate business ROI

Agentic finance AI improves enterprise effectivity and ROI solely when deployed with strict governance and…

February 25, 2026

On a more intelligent approach to connectivity delivery

GNX is on a mission to ship connectivity ‘the best way it must be’, selling…

September 17, 2025

Nvidia Sees $5.5B Hit From New Trump China Curbs on Chips

(Bloomberg) -- President Donald Trump’s administration has barred Nvidia Company from promoting its H20 chip…

April 17, 2025

Capacity Europe 2025: Where global connectivity & digital infra converge

Regular 0 false false false EN-US X-NONE X-NONE In an period the place AI is…

August 14, 2025

You Might Also Like

STL launches Neuralis data centre connectivity suite in the U.S.
AI & Compute

STL launches Neuralis data centre connectivity suite in the U.S.

By saad
What is optical interconnect and why Lightelligence's $10B debut says it matters for AI
AI & Compute

What is optical interconnect and why Lightelligence’s $10B debut says it matters for AI

By saad
IBM launches AI platform Bob to regulate SDLC costs
AI & Compute

IBM launches AI platform Bob to regulate SDLC costs

By saad
The evolution of encoders: From simple models to multimodal AI
AI & Compute

The evolution of encoders: From simple models to multimodal AI

By saad

About Us

Data Center News is your dedicated source for data center infrastructure, AI compute, cloud, and industry news.

Top Categories

  • AI & Compute
  • Cloud Computing
  • Power & Cooling
  • Colocation
  • Security
  • Infrastructure
  • Sustainability
  • Industry News

Useful Links

  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

Find Us on Socials

© 2026 Data Center News. All Rights Reserved.

© 2026 Data Center News. All Rights Reserved.
Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.