Sunday, 8 Feb 2026
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > Power & Cooling > Red Hat Unveils AI Inference Server in Latest Product Expansion
Power & Cooling

Red Hat Unveils AI Inference Server in Latest Product Expansion

Last updated: May 20, 2025 8:37 pm
Published May 20, 2025
Share
Red Hat Unveils AI Inference Server in Latest Product Expansion
SHARE

Crimson Hat has launched the Crimson Hat AI Inference Server, which permits enterprises to run generative AI functions quicker and extra effectively, the corporate introduced at this time (Could 20).

Launched at this week’s Crimson Hat Summit in Boston, the brand new AI inference server software program builds upon the open supply vLLM project and incorporates expertise from Crimson Hat’s latest acquisition of startup Neural Magic.

It options instruments that compress skilled AI fashions so that they run extra effectively. It additionally makes extra environment friendly use of processor reminiscence, enabling quicker inferencing throughout hybrid cloud environments, the corporate stated.

In accordance with business analysts, the corporate’s actions spotlight how AI acceleration encompasses not simply quick processors but in addition optimized software program.

“AI places loads of stress on computing programs, and with the appearance of AI brokers, it can put much more stress sooner or later,” Rick Villars, IDC’s group vice chairman of worldwide analysis, advised DCN. “Crimson Hat is saying they need to allow you to optimize your investments.

“As you go from mannequin constructing to embedding it to what you are promoting processes or buyer experiences, they are going to do every little thing they will on the software program degree to be sure you get most efficiency.”

Optimized AI Fashions

Associated:Knowledge Middle Outages Decline for Fourth Straight Yr, However Points Persist

The Crimson Hat AI Inference Server accelerates inferencing, which means it supplies quicker generative AI mannequin responses and handles extra customers concurrently with out requiring further {hardware}, the corporate stated.

See also  Mark Zuckerberg unveils his vision for superintelligence

The software program does so by optimizing using GPUs by way of methods akin to higher reminiscence administration and steady batching. A Crimson Hat spokesperson stated the expertise can optimize AMD and Nvidia GPUs, Intel’s Gaudi AI accelerators and Google TPUs.

The AI inference server may also be used to optimize AI fashions, akin to DeepSeek, Google’s Gemma, Meta’s open supply Llama, Mistral, Microsoft’s Phi and different massive language fashions.

Crimson Hat makes validated and optimized AI fashions accessible on Hugging Face, the corporate stated.

“Pre-optimized fashions working on vLLM typically ship two to 4 occasions extra token manufacturing – so a a lot larger degree of effectivity,” stated Brian Stevens, Crimson Hat’s senior vice chairman and AI chief expertise officer, throughout a media briefing.

Earlier than vLLM launched two years in the past, inference server choices have been restricted, however Nvidia supplied one in its software program stack, Stevens stated. Now, vLLM has gained traction due to its ease of use, the power to run fashions from Hugging Face, its OpenAI-compatible interface and its help for a number of AI accelerators.

Associated:Rack-Scale Revolution: AI Drives New Period of Knowledge Middle Structure

The AI Inference Server, which is Crimson Hat’s implementation of vLLM, could be deployed as a standalone containerized providing. It may also be deployed as an built-in element of Crimson Hat’s AI software program portfolio. That features Crimson Hat Enterprise Linux AI, a model of the open supply OS tailor-made for AI, and Crimson Hat OpenShift AI, a platform for constructing and deploying AI functions in containerized Kubernetes environments on-premises and within the cloud.

See also  OpenAI faces diminishing returns with latest AI model

Virtualization Market Development

Throughout this week’s Crimson Hat Summit, executives stated they’ve seen greater than 150% progress in Crimson Hat OpenShift Virtualization deployments since 2024.

To draw extra virtualization prospects, Crimson Hat stated Google Cloud, Microsoft Azure, and Oracle Cloud Infrastructure are making Crimson Hat OpenShift Virtualization accessible as expertise or public previews.

The corporate additionally introduced the overall availability of its virtualization software program on Amazon Internet Providers (AWS) and IBM Cloud.

“Clients, when they’re selecting their next-generation virtualization platform, need to go wherever their infrastructure selection leads them, and we needed to actually hone out and construct out these relationships with our cloud suppliers,” stated Mike Barrett, vice chairman and normal supervisor of Crimson Hat’s Hybrid Cloud Platforms, in a media briefing.

Associated:AMD Introduces New Server CPU for SMBs

Jim Mercer, IDC’s program vice chairman of software program growth, DevOps and DevSecOps, stated Crimson Hat has put loads of effort in bettering its virtualization software program. And whereas the corporate shouldn’t be saying it’s matching rival Broadcom characteristic for characteristic, Crimson Hat is implying that it gives a lot of the main virtualization options that prospects need.

“Quite a lot of prospects who’ve Crimson Hat OpenShift even have VMware vSphere, So Crimson Hat already has a foothold,” Mercer stated. “Crimson Hat is making an attempt to reap the benefits of the truth that, ‘You recognize us as associate. We’re going that can assist you with the migration, and we’re going to make the migration as simple as doable for you.’”

Crimson Hat Summit 2025: Extra Key Bulletins

On the Crimson Hat Summit, the corporate additionally introduced:

  • Crimson Hat Enterprise Linux 10. The brand new OS, accessible at this time, contains new safety features that defend towards assaults from future quantum computer systems. The ‘picture mode’ characteristic permits the OS to be deployed as a bootable container picture. By containerizing the OS and functions, enterprises can streamline administration utilizing the identical, constant instruments and workflow, the corporate stated.

  • New llm-d open supply group to scale inferencing. Crimson Hat introduced the launch of llm-d group, whose objective is to leverage vLLM and scale inferencing by way of a distributed strategy. Founding contributors are CoreWeave, Google, IBM Analysis, and Nvidia. Different members embrace AMD, Cisco, Intel, Lambda, and Mistral AI.

  • Lightspeed generative AI assistants. To deal with the abilities hole, Crimson Hat launched Lightspeed in Enterprise Linux 10, permitting IT directors to make use of pure language to get help for every little thing from troubleshooting frequent issues to managing advanced environments. In June, Crimson Hat will launch OpenShift Lightspeed, a generative AI assistant for managing and troubleshooting the OpenShift setting.

  • Crimson Hat Superior Developer Suite. Crimson Hat introduced the Superior Developer Suite, which mixes platform engineering instruments and safety capabilities.

  • Extra cloud information. Crimson Hat OpenShift is now accessible on Oracle Cloud Infrastructure, whereas Crimson Hat AI Inference Server is offered on Google Cloud.



Source link

Contents
Optimized AI FashionsVirtualization Market DevelopmentCrimson Hat Summit 2025: Extra Key Bulletins
TAGGED: Expansion, Hat, Inference, Latest, product, Red, Server, unveils
Share This Article
Twitter Email Copy Link Print
Previous Article Data Center Growth Demands Better Physical Security Standards Data Center Growth Demands Better Physical Security Standards
Next Article AI and quantum threats top security agendas AI and quantum threats top security agendas
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Microsoft’s agentic AI OmniParser rockets up open source charts

Be a part of our every day and weekly newsletters for the most recent updates…

November 4, 2024

Astera Labs Debuts PCIe Cables for Scaled Cloud and AI Deployments

Astera Labs, a renowned player in the semiconductor connectivity sphere has unveiled its Aries PCIe…

January 31, 2024

Industry groups drive Ethernet upgrades for AI, HPC

UEC model 1.0 Work on the UEC specs is following what the group calls a…

March 31, 2024

OpenAI, Nvidia, and Hugging Face unveil small AI models: GPT-4o Mini, Mistral-Nemo, and SmolLM lead industry shift

Be part of our every day and weekly newsletters for the newest updates and unique…

July 21, 2024

Colt Technology Services sells eight European data centres

Colt Expertise Companies is promoting six of its knowledge centres in main cities throughout Europe…

April 22, 2025

You Might Also Like

atNorth expands Nordic footprint with new Stockholm data centre
Power & Cooling

atNorth expands Nordic footprint with new Stockholm data centre

By saad
Microsoft unveils method to detect sleeper agent backdoors
AI

Microsoft unveils method to detect sleeper agent backdoors

By saad
Mark Powell named MD at Weatherite Air Conditioning
Power & Cooling

Mark Powell named MD at Weatherite Air Conditioning

By saad
Nokia and Blaize sign edge AI inference MOU targeting APAC networks
Edge Computing

Nokia and Blaize sign edge AI inference MOU targeting APAC networks

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.