Monday, 9 Feb 2026
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > AI > ​​IBM wants to be the enterprise LLM king with its new open-source Granite 3.1 models
AI

​​IBM wants to be the enterprise LLM king with its new open-source Granite 3.1 models

Last updated: December 26, 2024 11:06 am
Published December 26, 2024
Share
​​IBM wants to be the enterprise LLM king with its new open-source Granite 3.1 models
SHARE

Be a part of our every day and weekly newsletters for the newest updates and unique content material on industry-leading AI protection. Study Extra


IBM is staking its declare on the prime of the open-source AI leaderboard with its new Granite 3.1 sequence out as we speak.

The Granite 3.1 giant language fashions (LLMs) supply enterprise customers prolonged context size of 128K tokens, new embedding fashions, built-in hallucination detection and improved efficiency. In accordance with IBM, the brand new Granite 8B Instruct mannequin tops open-source rivals of the identical measurement together with Meta Llama 3.1, Qwen 2.5 and Google Gemma 2. IBM ranked its fashions throughout a sequence of educational benchmarks included within the OpenLLM Leaderboard. 

The brand new fashions are a part of the accelerated launch cadence of IBM’s Granite open-source fashions. Granite 3.0 was simply launched in October. On the time, IBM claimed that it has a $2 billion e book of enterprise associated to generative AI. With the Granite 3.1 replace, IBM is specializing in packing extra functionality into smaller fashions. The essential concept is that smaller fashions are simpler for enterprises to run and are extra cost-efficient to function.

“We’ve additionally simply boosted all of the numbers — all of the efficiency of just about all the things throughout the board has improved,” David Cox, VP for AI fashions at IBM Analysis, advised VentureBeat. “We use Granite for a lot of completely different use instances, we use it internally at IBM for our merchandise, we use it for consulting, we make it accessible to our prospects and we launch it as open supply, so now we have to be type of good at all the things.”

See also  What PubMatic's AgenticOS signals for enterprise marketing

Why efficiency and smaller fashions matter for enterprise AI

There are any variety of methods an enterprise can consider the efficiency of an LLM with benchmarks.

The route that IBM is taking is to run fashions by way of a gamut of educational and real-world checks. Cox emphasised that IBM examined and educated its fashions to be optimized for enterprise use instances. Efficiency isn’t nearly some summary measure of pace, both; somewhat, it’s a considerably extra nuanced measure of effectivity.

One side of effectivity that IBM is aiming to push ahead helps customers spend much less time to get desired outcomes.

“You must spend much less time fidgeting with prompts,” stated Cox. “So, the stronger a mannequin is in an space, the much less time it’s important to spend engineering prompts.”

Effectivity can also be about mannequin measurement. The bigger a mannequin, the extra compute and GPU sources it sometimes requires, which additionally means extra value.

“When individuals are doing minimal viable prototype type of work, they typically soar to very giant fashions, so that you would possibly go to a 70 billion parameter mannequin or a 405 billion parameter mannequin to construct your prototype,” stated Cox. “However the actuality is that lots of these will not be economical, so the opposite factor we’ve been attempting to do is drive as a lot capability as potential into the smallest package deal potential.”

Context issues for enterprise agentic AI

Apart from the promise of improved efficiency and effectivity, IBM has dramatically expanded Granite’s context size.

With the preliminary Granite 3.0 launch, the context size was restricted to 4k. In Granite 3.1, IBM has prolonged  that to 128k, permitting for the processing of for much longer paperwork. The prolonged context is a major improve for enterprise AI customers, each for retrieval-augmented technology (RAG) and for agentic AI.

See also  Open-source DeepSeek-R1 uses pure reinforcement learning to match OpenAI o1 — at 95% less cost

Agentic AI methods and AI brokers typically have to course of and motive over longer sequences of knowledge, equivalent to bigger paperwork, log traces or prolonged conversations. The elevated 128k context size permits these agentic AI methods to have entry to extra contextual info, enabling them to raised perceive and reply to advanced queries or duties.

IBM can also be releasing a sequence of embedding fashions to assist speed up the method of changing knowledge into vectors. The Granite-Embedding-30M-English mannequin can obtain efficiency of 0.16 seconds per question, which IBM claims is quicker than rival choices together with Snowflake’s Arctic.

How IBM has improved Granite 3.1 to serve enterprise AI wants

So how did IBM handle to enhance its efficiency for Granite 3.1? It wasn’t anybody particular factor, however somewhat a sequence of course of and technical improvements, Cox defined.

IBM has developed more and more superior multi-stage coaching pipelines, he stated. This has allowed the corporate to extract extra efficiency from fashions. Additionally, a essential a part of any LLM coaching is knowledge. Fairly than simply specializing in growing the amount of coaching knowledge, IBM has put a robust emphasis on enhancing the standard of information used to coach the Granite fashions.

“It’s not a amount recreation,” stated Cox. “It’s not like we’re going to exit and get 10 occasions extra knowledge and that’s magically going to make fashions higher.”

Decreasing hallucination straight within the mannequin

A standard strategy to lowering the danger of hallucinations and errant outputs in LLMs is to make use of guardrails. These are sometimes deployed as exterior options alongside an LLM.

See also  Anthropic just made it harder for AI to go rogue with its updated safety policy

With Granite 3.1, IBM is integrating hallucination safety straight into the mannequin. The Granite Guardian 3.1 8B and 2B fashions now embody a function-calling hallucination detection functionality.

“The mannequin can natively do its personal guardrailing, which can provide completely different alternatives to builders to catch issues,” stated Cox. 

He defined that performing hallucination detection within the mannequin itself optimizes the general course of. Inner detection means fewer inference calls, making the mannequin extra environment friendly and correct.

How enterprises can use Granite 3.1 as we speak, and what’s subsequent

The brand new Granite fashions are all now freely accessible as open supply to enterprise customers. The fashions are additionally accessible by way of IBM’s Watsonx enterprise AI service and will probably be built-in into IBM’s industrial merchandise.

The corporate plans on retaining an aggressive tempo for updating the Granite fashions. Trying ahead, the plan for Granite 3.2 is so as to add multimodal performance that may debut in early 2025. 

“You’re gonna see us over the subsequent few level releases, including extra of those varieties of various options which are differentiated, main as much as the stuff that we’ll announce on the IBM Suppose convention subsequent 12 months,” stated Cox.


Source link
TAGGED: enterprise, Granite, IBM, king, LLM, models, opensource
Share This Article
Twitter Email Copy Link Print
Previous Article AI Accelerated Servers Fuel Growth in Data Center Spending Demand for Google’s TPU Chips ‘Challenges NVIDIA’s Dominance’
Next Article Quantum Bridge, Juniper, Eurofiber Enable Quantum-Safe Connectivity Quantum Bridge, Juniper, Eurofiber Enable Quantum-Safe Connectivity
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

HPE Aruba targets retailers with IoT, edge computing launches

The 750 Collection APs help as much as 18.7 Gbps tri‑band combination knowledge charges and…

January 14, 2025

Microsoft just launched powerful AI ‘agents’ that could completely transform your workday — and challenge Google’s workplace dominance

Be part of our each day and weekly newsletters for the most recent updates and…

April 27, 2025

Elon Musk revives OpenAI legal battle with fresh allegations

Elon Musk’s revived lawsuit towards OpenAI consists of recent allegations towards the corporate and two…

August 5, 2024

Airalo Becomes Unicorn With an Investment Round of $220M

Airalo, a Lewes, DE-based eSIM supplier, acquired a $220M funding led by new investor, CVC.…

July 10, 2025

Google to invest $2B in northeast Indiana data center

Google on Friday stated it's creating a $2 billion knowledge middle in Fort Wayne, Ind.,…

April 27, 2024

You Might Also Like

SuperCool review: Evaluating the reality of autonomous creation
AI

SuperCool review: Evaluating the reality of autonomous creation

By saad
Top 7 best AI penetration testing companies in 2026
AI

Top 7 best AI penetration testing companies in 2026

By saad
Intuit, Uber, and State Farm trial AI agents inside enterprise workflows
AI

Intuit, Uber, and State Farm trial enterprise AI agents

By saad
How separating logic and search boosts AI agent scalability
AI

How separating logic and search boosts AI agent scalability

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.