Friday, 8 May 2026
Subscribe
logo
  • AI Compute
  • Infrastructure
  • Power & Cooling
  • Security
  • Colocation
  • Cloud Computing
  • More
    • Sustainability
    • Industry News
    • About Data Center News
    • Terms & Conditions
Font ResizerAa
Data Center NewsData Center News
Search
  • AI Compute
  • Infrastructure
  • Power & Cooling
  • Security
  • Colocation
  • Cloud Computing
  • More
    • Sustainability
    • Industry News
    • About Data Center News
    • Terms & Conditions
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > AI & Compute > Researchers find you don’t need a ton of data to train LLMs for reasoning tasks
AI & Compute

Researchers find you don’t need a ton of data to train LLMs for reasoning tasks

Last updated: February 15, 2025 7:03 am
Published February 15, 2025
Share
Researchers find you don’t need a ton of data to train LLMs for reasoning tasks
SHARE

Be a part of our each day and weekly newsletters for the most recent updates and unique content material on industry-leading AI protection. Study Extra


Giant language fashions (LLMs) can be taught complicated reasoning duties with out counting on massive datasets, in line with a new study by researchers at Shanghai Jiao Tong College. Their findings present that with only a small batch of well-curated examples, you may prepare an LLM for duties that have been thought to require tens of hundreds of coaching cases. 

This effectivity is as a result of inherent information that trendy LLMs acquire throughout the pre-training part. With new coaching strategies changing into extra data- and compute-efficient, enterprises would possibly be capable of create custom-made fashions with out requiring entry to the assets of huge AI labs.

Much less is extra (LIMO)

Of their examine, the researchers problem the idea that you just want massive quantities of knowledge to coach LLMs for reasoning duties. They introduce the idea of “much less is extra” (LIMO). Their work builds on prime of previous research that confirmed LLMs could possibly be aligned with human preferences with just a few examples.

Much less is Extra (LIMO) for reasoning (supply: arXiv)

Of their experiments, they demonstrated that they may create a LIMO dataset for complicated mathematical reasoning duties with just a few hundred coaching examples. An LLM fine-tuned on the dataset was capable of create complicated chain-of-thought (CoT) reasoning chains that enabled it to perform the duties at a really excessive success charge.

For instance, a Qwen2.5-32B-Instruct mannequin fine-tuned on 817 coaching examples chosen based mostly on LIMO reached 57.1% accuracy on the extremely difficult AIME benchmark and 94.8% on MATH, outperforming fashions that have been educated on 100 occasions extra examples. It additionally scored increased on the benchmarks than reasoning fashions similar to QwQ-32B-Preview (a model of the Qwen mannequin that has been educated for reasoning) and OpenAI o1-preview, each of which have been educated with bigger knowledge and compute assets.

See also  Net-Zero AI Data Center Project Gets a $5B Boost in Saudi Arabia

Furthermore, LIMO-trained fashions generalize to examples drastically totally different from their coaching knowledge. For instance, on the OlympiadBench scientific benchmark, the LIMO mannequin outperformed QwQ-32B-Preview, and on the difficult GPQA benchmark, it achieved 66.7% accuracy, near OpenAI-o1-preview’s main rating of 73.3%.

What does it imply for enterprise AI?

Customizing LLMs is a beautiful use case for enterprise purposes. Due to strategies similar to retrieval-augmented era (RAG) and in-context studying, LLMs could be custom-made to make use of bespoke knowledge or carry out new duties with out the necessity for costly fine-tuning. 

Nevertheless, reasoning duties typically require coaching and fine-tuning LLMs. The widely-held perception has been that such duties require massive volumes of coaching examples with extremely detailed reasoning chains and options. Creating such datasets is sluggish and impractical for a lot of purposes and firms.

Extra just lately, researchers have proven that pure reinforcement studying approaches can allow fashions to coach themselves for reasoning duties by producing many options and selecting those that work greatest. Whereas this strategy requires much less guide effort, it nonetheless calls for costly compute assets which can be past the attain of many enterprises.

However, crafting just a few hundred examples is an endeavor that many corporations can deal with, bringing specialised reasoning fashions throughout the attain of a wider vary of organizations.

“This discovery has profound implications for synthetic intelligence analysis: It means that even competition-level complicated reasoning skills could be successfully elicited by means of minimal however curated coaching samples,” the researchers write.

Why LIMO works

Of their experiments, the researchers determine two key the explanation why LLMs can be taught complicated reasoning duties with fewer examples.

See also  LangChain shows AI agents aren't human-level yet because they're overwhelmed by tools

First, state-of-the-art basis fashions have been educated on a really great amount of mathematical content material and code throughout pre-training. Because of this these LLMs already possess wealthy reasoning information of their parameters that may be activated by means of carefully-crafted examples.

Second, new post-training strategies have proven that permitting fashions to generate prolonged reasoning chains considerably improves their reasoning capability. In essence, giving the fashions extra time to “assume” permits them to unpack and apply their pre-trained information extra successfully.

“We hypothesize that profitable reasoning emerges from the synergy of those two elements: wealthy pre-trained information and enough computational assets at inference time,” the researchers write. “These developments collectively counsel a hanging risk: If fashions possess wealthy reasoning information and are given sufficient computational house, then activating their reasoning capabilities might require solely a small variety of high-quality coaching samples that encourage prolonged deliberation, moderately than huge fine-tuning datasets.”

Selecting extra complicated issues to incorporate within the coaching dataset can have a major impact on the educated mannequin’s accuracy in reasoning duties (supply: arXiv)

In response to the researchers’ findings, creating helpful LIMO datasets hinges on choosing the proper issues and options. Information curators ought to prioritize difficult issues that require complicated reasoning chains, numerous thought processes and information integration. The issues must also deviate from the mannequin’s coaching distribution to encourage new reasoning approaches and pressure it towards generalization.

Accordingly, options needs to be clearly and well-organized, with the reasoning steps tailored to the complexity of the issue. Excessive-quality options must also present strategic instructional help by progressively constructing understanding by means of rigorously structured explanations. 

See also  Qlik introduces Data Flow for faster Data-to-Decisions in Qlik Cloud Analytics

“By specializing in a minimal but meticulously curated set of reasoning chains, we embody the core precept of LIMO: Excessive-quality demonstrations, moderately than sheer knowledge quantity, are key to unlocking complicated reasoning capabilities,” the researchers write.

The researchers have released the code and data used to coach the LIMO fashions of their experiments. Sooner or later, they plan to broaden the idea to different domains and purposes.


Source link
TAGGED: data, Dont, find, LLMs, reasoning, researchers, tasks, ton, train
Share This Article
Twitter Email Copy Link Print
Previous Article AirTrunk to build second Johor data centre to meet cloud and AI demand AirTrunk to build second Johor data centre for cloud and AI growth
Next Article Perplexity just made AI research crazy cheap—what that means for the industry Perplexity just made AI research crazy cheap—what that means for the industry
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Managing AI-era cloud storage costs with Datadog

The rising use of AI is driving up cloud storage prices as enterprises battle to…

November 11, 2025

How Moonshot AI beat GPT-5 & Claude at a fraction of the cost

A Chinese language AI startup, Moonshot, has disrupted expectations in synthetic intelligence growth after its…

November 11, 2025

Vertiv supplies Digital Realty’s first Italian data centre

Vertiv, a worldwide chief in vital digital infrastructure, lately introduced its partnership with Digital Realty…

November 11, 2025

Apple scraps data protection tool for UK customers

Apple has pulled the plug on its highest degree knowledge safety device within the UK,…

February 21, 2025

Vertiv expands power portfolio | Data Centre Solutions

Vertiv has launched the Vertiv™ PowerSwitch 7000, a next-generation static switch swap that strengthens Vertiv's…

May 25, 2025

You Might Also Like

STL launches Neuralis data centre connectivity suite in the U.S.
AI & Compute

STL launches Neuralis data centre connectivity suite in the U.S.

By saad
What is optical interconnect and why Lightelligence's $10B debut says it matters for AI
AI & Compute

What is optical interconnect and why Lightelligence’s $10B debut says it matters for AI

By saad
IBM launches AI platform Bob to regulate SDLC costs
AI & Compute

IBM launches AI platform Bob to regulate SDLC costs

By saad
STL launches Neuralis data centre connectivity suite in the U.S.
Power & Cooling

STL launches Neuralis data centre connectivity suite in the U.S.

By saad

About Us

Data Center News is your dedicated source for data center infrastructure, AI compute, cloud, and industry news.

Top Categories

  • AI & Compute
  • Cloud Computing
  • Power & Cooling
  • Colocation
  • Security
  • Infrastructure
  • Sustainability
  • Industry News

Useful Links

  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

Find Us on Socials

© 2026 Data Center News. All Rights Reserved.

© 2026 Data Center News. All Rights Reserved.
Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.