Saturday, 11 Apr 2026
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > Cloud Computing > Energy-Efficient NPU Technology Cuts AI Power Use by 44%
Cloud Computing

Energy-Efficient NPU Technology Cuts AI Power Use by 44%

Last updated: July 11, 2025 7:01 am
Published July 11, 2025
Share
Energy-Efficient NPU Technology Cuts AI Power Use by 44%
SHARE

Researchers on the Korea Superior Institute of Science and Know-how (KAIST) have developed energy-efficient NPU know-how that demonstrates substantial efficiency enhancements in laboratory testing. 

Their specialised AI chip ran AI fashions 60% quicker whereas utilizing 44% much less electrical energy than the graphics playing cards at the moment powering most AI programs, primarily based on outcomes from managed experiments. 

To place it merely, the analysis, led by Professor Jongse Park from KAIST’s College of Computing in collaboration with HyperAccel Inc., addresses one of the vital urgent challenges in fashionable AI infrastructure: the large vitality and {hardware} necessities of large-scale generative AI fashions. 

Present programs resembling OpenAI’s ChatGPT-4 and Google’s Gemini 2.5 demand not solely excessive reminiscence bandwidth but additionally substantial reminiscence capability, driving corporations like Microsoft and Google to buy a whole bunch of 1000’s of NVIDIA GPUs.

The reminiscence bottleneck problem

The core innovation lies within the crew’s strategy to fixing reminiscence bottleneck points that plague present AI infrastructure. Their energy-efficient NPU know-how focuses on “light-weight” the inference course of whereas minimising accuracy loss—a essential steadiness that has confirmed difficult for earlier options.

PhD scholar Minsu Kim and Dr Seongmin Hong from HyperAccel Inc., serving as co-first authors, introduced their findings on the 2025 Worldwide Symposium on Pc Structure (ISCA 2025) in Tokyo. The analysis paper, titled “Oaken: Fast and Efficient LLM Serving with Online-Offline Hybrid KV Cache Quantization,” particulars their complete strategy to the issue.

The know-how centres on KV cache quantisation, which the researchers determine as accounting for most reminiscence utilization in generative AI programs. By optimising this part, the crew allows the identical stage of AI infrastructure efficiency utilizing fewer NPU gadgets in comparison with conventional GPU-based programs.

See also  Flexible nanogenerator with enhanced power density could one day rival the power of solar panels

Technical innovation and structure

The KAIST crew’s energy-efficient NPU know-how employs a three-pronged quantisation algorithm: threshold-based online-offline hybrid quantisation, group-shift quantisation, and fused dense-and-sparse encoding. This strategy permits the system to combine with present reminiscence interfaces with out requiring modifications to operational logic in present NPU architectures.

The {hardware} structure incorporates page-level reminiscence administration methods for environment friendly utilisation of restricted reminiscence bandwidth and capability. Moreover, the crew launched new encoding methods particularly optimised for quantised KV cache, addressing the distinctive necessities of their strategy.

“This analysis, via joint work with HyperAccel Inc., discovered an answer in generative AI inference light-weighting algorithms and succeeded in creating a core NPU know-how that may resolve the reminiscence downside,” Professor Park defined. 

“By way of this know-how, we applied an NPU with over 60% improved efficiency in comparison with the newest GPUs by combining quantisation methods that scale back reminiscence necessities whereas sustaining inference accuracy.”

Sustainability implications

The environmental influence of AI infrastructure has develop into a rising concern as generative AI adoption accelerates. The energy-efficient NPU know-how developed by KAIST presents a possible path towards extra sustainable AI operations. 

With 44% decrease energy consumption in comparison with present GPU options, widespread adoption may considerably scale back the carbon footprint of AI cloud providers. Nonetheless, the know-how’s real-world influence will rely upon a number of elements, together with manufacturing scalability, cost-effectiveness, and trade adoption charges. 

The researchers acknowledge that their resolution represents a big step ahead, however widespread implementation would require continued growth and trade collaboration.

Trade context and future outlook

The timing of this energy-efficient NPU know-how breakthrough is especially related as AI corporations face rising strain to steadiness efficiency with sustainability. The present GPU-dominated market has created provide chain constraints and elevated prices, making various options more and more enticing.

See also  Transparent Power: Serverfarm’s Renewable Energy Commitment

Professor Park famous that the know-how “has demonstrated the opportunity of implementing high-performance, low-power infrastructure specialised for generative AI, and is predicted to play a key position not solely in AI cloud knowledge centres but additionally within the AI transformation (AX) setting represented by dynamic, executable AI resembling agentic AI.”

The analysis represents a big step towards extra sustainable AI infrastructure, however its final influence will probably be decided by how successfully it may be scaled and deployed in industrial environments. Because the AI trade continues to grapple with vitality consumption issues, improvements like KAIST’s energy-efficient NPU know-how provide hope for a extra sustainable future in synthetic intelligence computing.

(Photograph by Korea Superior Institute of Science and Know-how)

See additionally: The 6 practices that guarantee extra sustainable knowledge centre operations

Need to be taught extra about cybersecurity and the cloud from trade leaders? Try Cyber Security & Cloud Expo going down in Amsterdam, California, and London.

Discover different upcoming enterprise know-how occasions and webinars powered by TechForge here.

Source link

TAGGED: Cuts, EnergyEfficient, NPU, Power, technology
Share This Article
Twitter Email Copy Link Print
Previous Article Torch Torch Acquires Praxis Labs
Next Article Ultrathin clay membrane layers offer low-cost alternative for extracting lithium from water Ultrathin clay membrane layers offer low-cost alternative for extracting lithium from water
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Autogon AI Receives Funding from Fast Forward Venture Studio

Autogon AI, a Houston, TX-based no-code AI orchestration platform, obtained funding from Quick Ahead Enterprise Studio.…

February 29, 2024

The Role of Data Governance in Key Regulatory Areas

Geoffrey Kee (KY), VP, Information Governance | AML Analytics | Improvements…

May 23, 2024

Shell launches Direct Liquid Cooling Fluid

This propylene glycol-based fluid enhances Shell’s current immersion cooling fluid portfolio, providing wonderful warmth switch…

June 4, 2025

Applied Digital Completes Financing for Jamestown HPC Data Center Campus

Utilized Digital CompanyDALLAS, March 05, 2024 (GLOBE NEWSWIRE) -- Applied Digital Corporation (Nasdaq: APLD) ("Utilized…

March 5, 2024

Soma Global Receives Majority Investment From Greater Sum Ventures

Soma Global, a Tampa, FL-based company which specializes in cloud-native public safety software solutions, received…

February 11, 2024

You Might Also Like

Heat emission from the chimneys of a large data and server complex.
Global Market

OpenAI puts part of Stargate project on hold over runaway power costs

By saad
Uber expands use of AWS chips for AI workloads
Cloud Computing

Uber expands use of AWS chips for AI workloads

By saad
Tanium introduces autonomy driven by AI and security innovation
Cloud Computing

Tanium introduces autonomy driven by AI and security innovation

By saad
Nscale moves into power with AIPCorp deal, building 8GW U.S. AI campus to bypass energy bottlenecks
Edge Computing

Nscale moves into power with AIPCorp deal, building 8GW U.S. AI campus to bypass energy bottlenecks

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.