Sunday, 14 Dec 2025
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > Cloud Computing > Building the future of AI systems at Meta
Cloud Computing

Building the future of AI systems at Meta

Last updated: December 4, 2024 7:31 am
Published December 4, 2024
Share
Building the future of AI systems at Meta
SHARE

Meta’s Ye (Charlotte) Qi took the stage at QCon San Francisco 2024, to debate the challenges of working LLMs at scale.

As reported by InfoQ, her presentation targeted on what it takes to handle huge fashions in real-world methods, highlighting the obstacles posed by their dimension, complicated {hardware} necessities, and demanding manufacturing environments.

She in contrast the present AI increase to an “AI Gold Rush,” the place everyone seems to be chasing innovation however encountering important roadblocks. In response to Qi, deploying LLMs successfully isn’t nearly becoming them onto current {hardware}. It’s about extracting each little bit of efficiency whereas conserving prices beneath management. This, she emphasised, requires shut collaboration between infrastructure and mannequin improvement groups.

Making LLMs match the {hardware}

One of many first challenges with LLMs is their huge urge for food for sources — many fashions are just too giant for a single GPU to deal with. To sort out this, Meta employs methods like splitting the mannequin throughout a number of GPUs utilizing tensor and pipeline parallelism. Qi harassed that understanding {hardware} limitations is essential as a result of mismatches between mannequin design and obtainable sources can considerably hinder efficiency.

Her recommendation? Be strategic. “Don’t simply seize your coaching runtime or your favorite framework,” she stated. “Discover a runtime specialised for inference serving and perceive your AI downside deeply to select the appropriate optimisations.”

Pace and responsiveness are non-negotiable for functions counting on real-time outputs. Qi spotlighted methods like steady batching to maintain the system working easily, and quantisation, which reduces mannequin precision to make higher use of {hardware}. These tweaks, she famous, can double and even quadruple efficiency.

See also  Vespertec, Sardina Systems, and Borealis Data Center join forces

When prototypes meet the true world

Taking an LLM from the lab to manufacturing is the place issues get actually tough. Actual-world situations deliver unpredictable workloads and stringent necessities for pace and reliability. Scaling isn’t nearly including extra GPUs — it entails rigorously balancing value, reliability, and efficiency.

Meta addresses these points with methods like disaggregated deployments, caching methods that prioritise often used information, and request scheduling to make sure effectivity. Qi acknowledged that constant hashing — a technique of routing-related requests to the identical server — has been notably useful for enhancing cache efficiency.

Automation is extraordinarily necessary within the administration of such difficult methods. Meta depends closely on instruments that monitor efficiency, optimise useful resource use, and streamline scaling selections, and Qi claims Meta’s customized deployment options permit the corporate’s providers to answer altering calls for whereas conserving prices in examine.

The large image

Scaling AI methods is greater than a technical problem for Qi; it’s a mindset. She stated corporations ought to take a step again and take a look at the larger image to determine what actually issues. An goal perspective helps companies concentrate on efforts that present long-term worth, consistently refining methods.

Her message was clear: succeeding with LLMs requires greater than technical experience on the mannequin and infrastructure ranges – though on the coal-face, these components are of paramount significance. It’s additionally about technique, teamwork, and specializing in real-world impression.

(Photograph by Unsplash)

See additionally: Samsung chief engages Meta, Amazon and Qualcomm in strategic tech talks

Wish to be taught extra about cybersecurity and the cloud from business leaders? Take a look at Cyber Security & Cloud Expo happening in Amsterdam, California, and London. Discover different upcoming enterprise expertise occasions and webinars powered by TechForge here.

Tags: AI, cloud, GPU

See also  Tokyo Residents Seek to Block Building of Massive Data Center

Source link

Contents
Making LLMs match the {hardware}When prototypes meet the true worldThe large image
TAGGED: Building, Future, Meta, Systems
Share This Article
Twitter Email Copy Link Print
Previous Article Platform allows AI to learn from constant, nuanced human feedback rather than large datasets Platform allows AI to learn from constant, nuanced human feedback rather than large datasets
Next Article Reflexivity Raises $30M in Series B Funding MakersHub Raises $7M in Additional Seed Funding
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Lata Varghese – Rackspace Technology –

Lata Varghese has been named Senior Vice President of Enterprise Transformation by Rackspace Know-how. Lata…

November 25, 2025

EU Unveils Plan to Ease Water Shortages and Curb Data Center Use

(Bloomberg) -- The European Union unveiled plans to spice up resilience to droughts and enhance…

June 10, 2025

The Brazil Data Center Market to Reach USD 4.43 Billion by 2028 – Arizton

Get Insights on 51 Existing Data Centers and 22 Upcoming Facilities across Brazil According to…

January 25, 2024

Novel system turns quantum bottlenecks into breakthroughs

A flowchart of the HyperQ workflow. Credit score: Jason Nieh and Ronghui Gu Quantum computer…

July 9, 2025

The race to atomic-layer devices

Ellie Gabel discusses how computing on the thinnest scale is driving breakthroughs in quantum know-how,…

August 26, 2025

You Might Also Like

Newsweek: Building AI-resilience for the next era of information
AI

Newsweek: Building AI-resilience for the next era of information

By saad
atNorth's Iceland data centre epitomises circular economy
Cloud Computing

atNorth’s Iceland data centre epitomises circular economy

By saad
Experimental AI concludes as autonomous systems rise
AI

Experimental AI concludes as autonomous systems rise

By saad
How cloud infrastructure shapes the modern Diablo experience 
Cloud Computing

How cloud infrastructure shapes the modern Diablo experience 

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.