Monday, 19 Jan 2026
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > Global Market > Red Hat Execs on Inference Scaling and Open-Source AI
Global Market

Red Hat Execs on Inference Scaling and Open-Source AI

Last updated: July 20, 2025 8:45 am
Published July 20, 2025
Share
Red Hat Execs on Inference Scaling and Open-Source AI
SHARE

This week on ‘No Math AI,’ Akash and Isha go to the Pink Hat Summit to fulfill with Pink Hat CEO Matt Hicks and CTO Chris Wright. They speak concerning the sensible necessities of introducing inference time scaling, also referred to as take a look at time scaling/compute, to company customers across the globe.

Matt Hicks examines how an AI platform is important for abstracting complexity and bearing bills as AI evolves from static fashions to dynamic, agentic purposes. These purposes make in depth use of inference time scaling strategies, together with particle filtering and reasoning, which produce numerous tokens so as to enhance accuracy. Hicks highlights the need of platforms that cut back the unit price of those capabilities, make it easy for companies to implement such methods, and encourage confidence by providing price transparency so as to recover from the “concern response” associated to unexpected prices when doing extra inferencing.

Chris Wright describes the open-source AI roadmap for implementing these novel, inference-based applied sciences in manufacturing in a dependable method. Shifting from single-instance inference to a distributed infrastructure that may help a number of customers directly and successfully handle the large token manufacturing wanted by these scaled inference processes is likely one of the difficulties he addresses. Wright presents the brand new Pink Hat undertaking LLM-d, which goals to offer a distributed inference platform commonplace. By way of integration with Kubernetes, LLM-d seeks to enhance {hardware} utilization, handle distributed KV caches, and intelligently route requests relying on {hardware} wants. By way of cooperative open-source initiatives, the target is to offer repeatable blueprints for a shared structure to handle these inference-time-scaling workloads.

See also  Linux in your car: Red Hat’s milestone collaboration with exida

An important impediment, as Hicks and Wright level out, is effectively increasing the underlying inference structure from single-server cases to a steady, distributed, and clear platform. For company AI to advance and for inference time scaling to be broadly used, this barrier should be addressed by means of group initiatives.

Source link

TAGGED: execs, Hat, Inference, opensource, Red, Scaling
Share This Article
Twitter Email Copy Link Print
Previous Article Motif Raises $46M in Seed and Series A Funding Blockskye Raises $15.8M in Funding
Next Article Receive Raises $4M in Seed Funding Receive Raises $4M in Seed Funding
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Bybit Named Exclusive Payment Partner for Tomorrowland Brasil 2025-26, Launches Cardholder Presale

Dubai, UAE, March twenty second, 2025, Chainwire   Bybit, the world’s second-largest cryptocurrency change by…

March 22, 2025

Kao Data secures new customer contract

Kao Information has secured a brand new buyer contract with 20i, one of many UK's…

March 31, 2025

Nobl Q Acquires Propel

Nobl Q, a Dallas, TX-based international AI- and cloud-based options supplier, acquired Propel, a software…

August 12, 2024

Cloudflare discloses breach related to stolen Okta data

Last fall, Cloudflare announced it mitigated an attempted cyberattack stemming from the infamous Okta breach.…

February 4, 2024

What to expect when starting your AI journey

Pavan Bedadala, Senior Director of Product Administration at Commvault, discusses the way it’s vital for…

March 20, 2024

You Might Also Like

AI Factory 16:9
Global Market

What exactly is an AI factory?

By saad
Huawei MWC 2025.
Global Market

Chinese AI firm trains state-of-the-art model entirely on Huawei chips

By saad
Future-proof your IT career
Global Market

Cisco’s 2026 agenda prioritizes AI-ready infrastructure, connectivity

By saad
Nvidia high-performance chip technology
Global Market

Nvidia H200 chips in China: US says yes, China says no

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.