Sunday, 3 May 2026
Subscribe
logo
  • AI Compute
  • Infrastructure
  • Power & Cooling
  • Security
  • Colocation
  • Cloud Computing
  • More
    • Sustainability
    • Industry News
    • About Data Center News
    • Terms & Conditions
Font ResizerAa
Data Center NewsData Center News
Search
  • AI Compute
  • Infrastructure
  • Power & Cooling
  • Security
  • Colocation
  • Cloud Computing
  • More
    • Sustainability
    • Industry News
    • About Data Center News
    • Terms & Conditions
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > AI & Compute > Rapt AI and AMD work to make GPU utilization more efficient
AI & Compute

Rapt AI and AMD work to make GPU utilization more efficient

Last updated: March 30, 2025 1:33 pm
Published March 30, 2025
Share
Rapt AI and AMD work to make GPU utilization more efficient
SHARE

Rapt AI, a supplier of AI-powered AI-workload automation for GPUs and AI accelerators, has teamed with AMD to reinforce AI infrastructure.

The long-term strategic collaboration goals to enhance AI inference and coaching workload administration and efficiency on AMD Intuition GPUs, providing prospects a scalable and cost-effective resolution for deploying AI purposes.

As AI adoption accelerates, organizations are grappling with useful resource allocation, efficiency bottlenecks, and complicated GPU administration.

By integrating Rapt’s clever workload automation platform with AMD Intuition MI300X, MI325X and upcoming MI350 sequence GPUs, this collaboration delivers a scalable, high-performance, and cost-effective resolution that allows prospects to maximise AI inference and coaching effectivity throughout on-premises and multi-cloud infrastructures.

A extra environment friendly resolution

AMD Intuition MI325X GPU.

Charlie Leeming, CEO of Rapt AI, stated in a press briefing, “The AI fashions we’re seeing immediately are so giant and most significantly are so dynamic and unpredictable. The older instruments for optimizing don’t actually match in any respect. We noticed these dynamics. Enterprises are throwing plenty of cash. Hiring a brand new set of expertise in AI. It’s one among these disruptive applied sciences. We now have a state of affairs the place CFOs and CIOs are asking the place is the return. In some instances, there may be tens of thousands and thousands, tons of of thousands and thousands or billions of {dollars} spend on GPU-related infrastructure.”

Leeming stated Anil Ravindranath, CTO of Rapt AI, noticed the answer. And that concerned deploying screens to allow observations of the infrastructure.

“We really feel we have now the correct resolution on the proper time. We got here out of stealth final fall. We’re in a rising variety of Fortune 100 corporations. Two are working the code amongst cloud service suppliers,” Leeming stated.

See also  E.U. Will Work on Setting Water Use Caps for Thirsty Data Centers

And he stated, “We do have strategic companions however our conversations with AMD went extraordinarily nicely. They’re constructing large GPUs, AI accelerators. We’re recognized for placing the utmost quantity of workload on GPUs. Inference is taking off. It’s in manufacturing stage now. AI workloads are exploding. Their information scientists are working as quick as they will. They’re panicking, they want instruments, they want effectivity, they want automation. It’s screaming for the correct resolution. Inefficiencies — 30% GPU underutilization. Prospects do need flexibility. Giant prospects are asking when you assist AMD.”

Enhancements that may take 9 hours might be performed in three minutes, he stated. Ravindranath stated in a press briefing the Rapt AI platform allows as much as 10 occasions mannequin run capability on the similar AI compute spending degree, as much as 90% value financial savings, and 0 people in a loop and no code modifications. For productiveness, this implies no extra ready for compute and time spent tuning infrastructure.

Lemming stated different methods have been round for some time and haven’t lower it. Run AI, a rival, overlaps in a aggressive manner considerably. He stated his firm observes in minutes as an alternative of hours after which optimizes the infrastructure. Ravindranath stated Run AI is extra like a scheduler however Rapt AI positions itself for unpredictable outcomes and offers with it.

“We run the mannequin and determine it out, and that’s an enormous profit for inference workloads. It ought to simply routinely run,” Ravindranath stated.

The advantages: decrease prices, higher GPU utilization

AMD Intuition MI300X GPU.

The businesses stated that AMD Intuition GPUs, with their industry-leading reminiscence capability, mixed with
Rapt’s clever useful resource optimization, helps guarantee most GPU utilization for AI workloads, serving to decrease complete value of possession (TCO).

See also  BigQuery is 5x bigger than Snowflake and Databricks: What Google is doing to make it even better

Rapt’s platform streamlines GPU administration, eliminating the necessity for information scientists to spend worthwhile time on trial-and-error infrastructure configurations. By routinely optimizing useful resource allocation for his or her particular workloads, it empowers them to concentrate on innovation slightly than infrastructure. It seamlessly helps numerous GPU environments (AMD and others, whether or not within the cloud, on premises or each) by way of a single occasion, serving to guarantee most infrastructure flexibility.

The mixed resolution intelligently optimizes job density and useful resource allocation on AMD Intuition GPUs, leading to higher inference efficiency and scalability for manufacturing AI deployments. Rapt’s auto-scaling capabilities additional assist guarantee environment friendly useful resource use based mostly on demand, lowering latency and maximizing value effectivity.

Rapt’s platform works out-of-the-box with AMD Intuition GPUs, serving to guarantee fast efficiency advantages. Ongoing collaboration between Rapt and AMD will drive additional optimizations in thrilling areas similar to GPU scheduling, reminiscence utilization and extra, serving to guarantee prospects are geared up with a future prepared AI infrastructure.

“At AMD, we’re dedicated to delivering high-performance, scalable AI options that empower organizations to unlock the total potential of their AI workloads.” stated Negin Oliver, company vice chairman of enterprise growth for information middle GPU enterprise at AMD, in an announcement. “Our collaboration with Rapt AI combines the cutting-edge capabilities of AMD Intuition GPUs with Rapt’s clever workload automation, enabling prospects to realize higher effectivity, flexibility, and price financial savings throughout their AI infrastructure.”

Source link

Contents
A extra environment friendly resolutionThe advantages: decrease prices, higher GPU utilization
TAGGED: AMD, efficient, GPU, Rapt, utilization, Work
Share This Article
Twitter Email Copy Link Print
Previous Article Corning launches GlassWorks AI solutions Corning launches GlassWorks AI solutions
Next Article Vertiv Apprenticeship Programme boosts jobs and skills in Derry and Donegal Vertiv Apprenticeship Programme boosts jobs and skills in Derry and Donegal
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Teaching the model: Designing LLM feedback loops that get smarter over time

Need smarter insights in your inbox? Join our weekly newsletters to get solely what issues…

August 16, 2025

Google DeepMind researchers introduce new benchmark to improve LLM factuality, reduce hallucinations

Be a part of our day by day and weekly newsletters for the newest updates…

January 11, 2025

A standard, open framework for building AI agents is coming from Cisco, LangChain and Galileo

Be a part of our each day and weekly newsletters for the most recent updates…

March 9, 2025

Imagination unveils E-Series GPUs for graphics and AI at the edge

Creativeness Applied sciences is unveiling its E-Sequence graphics processing items (GPUs) for graphics and AI…

May 8, 2025

Turnkey AI inference solution for data centres

As international AI inference calls for soar, conventional datacenters grapple with prolonged deployment timelines of…

July 12, 2025

You Might Also Like

STL launches Neuralis data centre connectivity suite in the U.S.
AI & Compute

STL launches Neuralis data centre connectivity suite in the U.S.

By saad
What is optical interconnect and why Lightelligence's $10B debut says it matters for AI
AI & Compute

What is optical interconnect and why Lightelligence’s $10B debut says it matters for AI

By saad
IBM launches AI platform Bob to regulate SDLC costs
AI & Compute

IBM launches AI platform Bob to regulate SDLC costs

By saad
The evolution of encoders: From simple models to multimodal AI
AI & Compute

The evolution of encoders: From simple models to multimodal AI

By saad

About Us

Data Center News is your dedicated source for data center infrastructure, AI compute, cloud, and industry news.

Top Categories

  • AI & Compute
  • Cloud Computing
  • Power & Cooling
  • Colocation
  • Security
  • Infrastructure
  • Sustainability
  • Industry News

Useful Links

  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

Find Us on Socials

© 2026 Data Center News. All Rights Reserved.

© 2026 Data Center News. All Rights Reserved.
Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.