Monday, 12 Jan 2026
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > AI > Liquid AI is revolutionizing LLMs to work on edge devices like smartphones with new ‘Hyena Edge’ model
AI

Liquid AI is revolutionizing LLMs to work on edge devices like smartphones with new ‘Hyena Edge’ model

Last updated: April 25, 2025 11:32 pm
Published April 25, 2025
Share
Liquid AI is revolutionizing LLMs to work on edge devices like smartphones with new 'Hyena Edge' model
SHARE

Be part of our day by day and weekly newsletters for the newest updates and unique content material on industry-leading AI protection. Study Extra


Liquid AI, the Boston-based basis mannequin startup spun out of the Massachusetts Institute of Know-how (MIT), is looking for to maneuver the tech {industry} past its reliance on the Transformer structure underpinning hottest massive language fashions (LLMs) corresponding to OpenAI’s GPT sequence and Google’s Gemini household.

Yesterday, the corporate introduced “Hyena Edge,” a brand new convolution-based, multi-hybrid mannequin designed for smartphones and different edge units prematurely of the International Conference on Learning Representations (ICLR) 2025.

The convention, one of many premier occasions for machine studying analysis, is happening this 12 months in Vienna, Austria.

New convolution-based mannequin guarantees quicker, extra memory-efficient AI on the edge

Hyena Edge is engineered to outperform robust Transformer baselines on each computational effectivity and language mannequin high quality.

In real-world checks on a Samsung Galaxy S24 Extremely smartphone, the mannequin delivered decrease latency, smaller reminiscence footprint, and higher benchmark outcomes in comparison with a parameter-matched Transformer++ mannequin.

A brand new structure for a brand new period of edge AI

Not like most small fashions designed for cellular deployment — together with SmolLM2, the Phi fashions, and Llama 3.2 1B — Hyena Edge steps away from conventional attention-heavy designs. As an alternative, it strategically replaces two-thirds of grouped-query consideration (GQA) operators with gated convolutions from the Hyena-Y household.

The brand new structure is the results of Liquid AI’s Synthesis of Tailor-made Architectures (STAR) framework, which makes use of evolutionary algorithms to mechanically design mannequin backbones and was introduced again in December 2024.

See also  Huawei's new Ascend chips to power world's most powerful cluster

STAR explores a variety of operator compositions, rooted within the mathematical idea of linear input-varying methods, to optimize for a number of hardware-specific goals like latency, reminiscence utilization, and high quality.

Benchmarked instantly on client {hardware}

To validate Hyena Edge’s real-world readiness, Liquid AI ran checks instantly on the Samsung Galaxy S24 Extremely smartphone.

Outcomes present that Hyena Edge achieved as much as 30% quicker prefill and decode latencies in comparison with its Transformer++ counterpart, with velocity benefits growing at longer sequence lengths.

Prefill latencies at quick sequence lengths additionally outpaced the Transformer baseline — a crucial efficiency metric for responsive on-device purposes.

When it comes to reminiscence, Hyena Edge persistently used much less RAM throughout inference throughout all examined sequence lengths, positioning it as a powerful candidate for environments with tight useful resource constraints.

Outperforming Transformers on language benchmarks

Hyena Edge was educated on 100 billion tokens and evaluated throughout customary benchmarks for small language fashions, together with Wikitext, Lambada, PiQA, HellaSwag, Winogrande, ARC-easy, and ARC-challenge.

On each benchmark, Hyena Edge both matched or exceeded the efficiency of the GQA-Transformer++ mannequin, with noticeable enhancements in perplexity scores on Wikitext and Lambada, and better accuracy charges on PiQA, HellaSwag, and Winogrande.

These outcomes counsel that the mannequin’s effectivity beneficial properties don’t come at the price of predictive high quality — a typical tradeoff for a lot of edge-optimized architectures.

Hyena Edge Evolution: A take a look at efficiency and operator traits

For these looking for a deeper dive into Hyena Edge’s growth course of, a latest video walkthrough supplies a compelling visible abstract of the mannequin’s evolution.

See also  From static classifiers to reasoning engines: OpenAI’s new model rethinks content moderation

The video highlights how key efficiency metrics — together with prefill latency, decode latency, and reminiscence consumption — improved over successive generations of structure refinement.

It additionally provides a uncommon behind-the-scenes take a look at how the interior composition of Hyena Edge shifted throughout growth. Viewers can see dynamic adjustments within the distribution of operator varieties, corresponding to Self-Consideration (SA) mechanisms, varied Hyena variants, and SwiGLU layers.

These shifts provide perception into the architectural design ideas that helped the mannequin attain its present degree of effectivity and accuracy.

By visualizing the trade-offs and operator dynamics over time, the video supplies precious context for understanding the architectural breakthroughs underlying Hyena Edge’s efficiency.

Open-source plans and a broader imaginative and prescient

Liquid AI stated it plans to open-source a sequence of Liquid basis fashions, together with Hyena Edge, over the approaching months. The corporate’s objective is to construct succesful and environment friendly general-purpose AI methods that may scale from cloud datacenters down to non-public edge units.

The debut of Hyena Edge additionally highlights the rising potential for various architectures to problem Transformers in sensible settings. With cellular units more and more anticipated to run subtle AI workloads natively, fashions like Hyena Edge might set a brand new baseline for what edge-optimized AI can obtain.

Hyena Edge’s success — each in uncooked efficiency metrics and in showcasing automated structure design — positions Liquid AI as one of many rising gamers to look at within the evolving AI mannequin panorama.


Source link
TAGGED: devices, edge, Hyena, liquid, LLMs, Model, Revolutionizing, smartphones, Work
Share This Article
Twitter Email Copy Link Print
Previous Article nanotechnology How miniaturisation is transforming technology
Next Article Aqualung Carbon Capture Closes Phase 1 2025 Financing Round Aqualung Carbon Capture Closes Phase 1 2025 Financing Round
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

BSO unveils DataOne | Data Centre Solutions

Situated in France and financed with the assist of the Ardian debt fund and main…

December 4, 2024

Tech hiring slows, more IT jobs lost

CompTIA estimates that throughout your complete economic system, tech occupations declined by 14,000 and the…

August 6, 2024

US to Tighten Rules Aimed at Keeping Advanced Chips Out of China | DCN

(Bloomberg) -- The Biden administration is refining rules aimed at keeping advanced chips and manufacturing…

February 12, 2024

HashKey Global Launches 2nd HashKey Launchpool: Earn ATH Tokens by Locking ATH & USDT

Burmuda, Bermuda, June eighth, 2024, Chainwire HashKey Global will launch the second part of the…

June 8, 2024

Microsoft gains major AI client as TikTok spends $20 million

The place many have struggled to show their cloud companies right into a worthwhile endeavour,…

August 2, 2024

You Might Also Like

Engineer
Global Market

AI, edge, and security: Shaping the need for modern infrastructure management

By saad
Autonomy without accountability: The real AI risk
AI

Autonomy without accountability: The real AI risk

By saad
The future of personal injury law: AI and legal tech in Philadelphia
AI

The future of personal injury law: AI and legal tech in Philadelphia

By saad
How AI code reviews slash incident risk
AI

How AI code reviews slash incident risk

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.