Saturday, 13 Dec 2025
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > AI > Weibo's new open source AI model VibeThinker-1.5B outperforms DeepSeek-R1 on $7,800 post-training budget
AI

Weibo's new open source AI model VibeThinker-1.5B outperforms DeepSeek-R1 on $7,800 post-training budget

Last updated: November 12, 2025 8:03 pm
Published November 12, 2025
Share
Weibo's new open source AI model VibeThinker-1.5B outperforms DeepSeek-R1 on $7,800 post-training budget
SHARE

One other day in late 2025, one other spectacular end result from a Chinese language firm in open supply synthetic intelligence.

Chinese language social networking firm Weibo’s AI division recently released its open source VibeThinker-1.5B—a 1.5 billion parameter massive language mannequin (LLM) that could be a fine-tuned variant of rival Chinese language tech agency Alibaba’s Qwen2.5-Math-1.5B.

It is obtainable now at no cost obtain and utilization by researchers and enterprise builders—even for industrial functions—below a permissive MIT License on Hugging Face, GitHub and ModelScope, with a technical report on open entry science publishing website arxiv.org.

And but, regardless of its compact dimension, VibeThinker-1.5B achieves benchmark-topping reasoning efficiency on math and code duties, rivaling or surpassing fashions tons of of occasions its dimension, even outperforming Chinese language rival DeepSeek’s famed R1 that went viral initially of this 12 months—a 671-billion parameter mannequin—on formal reasoning benchmark.

It additional eclipses Mistral AI’s Magistral Medium and holds its personal in opposition to Anthropic’s Claude Opus 4 and OpenAI’s gpt-oss-20B Medium, all whereas requiring a fraction of the infrastructure and funding.

It additionally does so having been post-trained on a funds of merely $7800 USD for compute assets (3900 GPU hours on Nvidia H800s) — far lower than the tens, and even tons of, of 1000’s of {dollars} sometimes required to fine-tune fashions of comparable or bigger scale.

Recall this isn’t the whole price of the mannequin’s growth, nonetheless: LLMs are skilled in phases. First comes pre-training, when the mannequin learns fundamental language construction and common data by predicting the following phrase throughout huge quantities of textual content from the web, books, and articles. This offers it fluency however not a lot sense of the way to comply with directions or maintain a dialog

Put up-training comes subsequent, utilizing a lot smaller, higher-quality datasets—sometimes collections of instance questions, prompts, and expert-written solutions—to show the mannequin the way to reply helpfully, motive via issues, and align with human expectations. Nonetheless, Weibo’s post-training price effectiveness on VibeThinker-1.5B is noteworthy and needs to be recommended.

The open-source launch upends assumptions about parameter scale, compute depth, and the minimal viable dimension for high-performance LLMs.

See also  Stellium Datacenters, Submer, and ExxonMobil join forces on Open Compute Project

A Completely different Coaching Strategy: Spectrum-to-Sign

VibeThinker-1.5B owes its efficiency to not scale, however to the coaching framework behind it: the Spectrum-to-Sign Precept (SSP).

As an alternative of optimizing a mannequin purely for single-answer correctness (Move@1), the SSP framework decouples supervised fine-tuning (SFT) and reinforcement studying (RL) into two distinct phases with completely different targets:

  • SFT (“Spectrum Section”): The mannequin is skilled to maximise range throughout potential right solutions, enhancing its Move@Okay rating. This builds a variety of believable resolution paths.

  • RL (“Sign Section”): A second-stage reinforcement studying system (referred to as MaxEnt-Guided Coverage Optimization, or MGPO) is used to determine and amplify essentially the most right paths from this numerous resolution pool. MGPO prioritizes issues the place the mannequin is most unsure, utilizing entropy-based weighting to focus studying.

The authors argue this separation permits small fashions to discover reasoning area extra successfully—attaining sign amplification with out counting on large parameter counts.

VibeThinker-1.5B makes a compelling case that the trade’s reliance on parameter scaling as the one route to higher reasoning efficiency could also be outdated.

By adopting a diversity-first coaching pipeline, WeiboAI has proven that smaller, extra accessible fashions can match and even outperform billion-dollar programs in logic-heavy duties.

The low useful resource footprint is among the many most important facets of VibeThinker-1.5B. At below $8,000, the post-training price is 30–60x decrease than fashions like DeepSeek R1 and MiniMax-M1, which price between $294K and $535K to coach.

Efficiency Throughout Domains

Regardless of its small dimension, VibeThinker-1.5B delivers cross-domain reasoning that outpaces many bigger open-source and industrial fashions:

Mannequin

AIME25

LiveCodeBench v6

GPQA-Diamond

VibeThinker-1.5B

74.4

51.1

46.7

GPT-OSS-20B-Medium

72.1

54.9

66.0

Claude Opus 4

69.2

56.6

79.6

MiniMax M1 (456B)

74.6

62.3

69.2

DeepSeek R1 (671B)

70.0

65.9

71.5

Kimi K2 (1.09T)

49.5

53.7

75.1

VibeThinker was benchmarked in opposition to each reasoning-centric fashions (Magistral, Claude, OpenAI o3-mini) and non-reasoning LLMs (GPT-4.1, Kimi K2, DeepSeek V3). Throughout structured reasoning benchmarks, the mannequin constantly outperformed non-reasoning fashions, no matter dimension:

  • On AIME24 (math), it beat Kimi K2 (1.09T) by over 10 factors (80.3 vs. 69.6).

  • On LiveCodeBench v6, it surpassed Claude Opus 4 (51.1 vs. 47.4).

  • On GPQA, it scored beneath GPT-4.1 and Claude, however nonetheless doubled its base mannequin (from 16.4 to 46.7).

See also  Covid RNA Analyzed With Tee on OVM: Open (Incl. RSS3) Supercharges DeSci

This helps the authors’ declare that dimension isn’t the one path to reasoning functionality—with correct coaching design, smaller fashions can attain and even exceed the efficiency of far bigger programs in focused duties.

Notably, it achieves parity with fashions tons of of occasions bigger on math and code, although it lags behind normally data reasoning (GPQA), the place bigger fashions keep an edge.

This means a possible specialization trade-off: whereas VibeThinker excels at structured logical duties, it has much less capability for wide-ranging encyclopedic recall, a identified limitation of smaller architectures.

Steering for Enterprise Adoption

The discharge consists of advisable inference settings (temperature = 0.6, top_p = 0.95, max tokens = 40960).

The mannequin is sufficiently small to be deployed on edge gadgets, together with cell phones and vehicle-embedded programs, whereas inference prices are estimated to be 20–70x cheaper than with massive fashions.

This positions VibeThinker-1.5B not simply as a analysis achievement, however as a possible basis for cost-efficient, domestically deployable reasoning programs.

Weibo’s Technique and Market Place

Weibo, launched by Sina Company in 2009, stays a cornerstone of China’s social media ecosystem. Usually described as China’s model of X (previously Twitter), the platform blends microblogging, multimedia content material, and trending-topic options with a regulatory surroundings formed by tight authorities oversight.

Regardless of counting 600 million month-to-month energetic customers (greater than twice that of X), investors are not optimistic about its advertising revenue growth potential within the close to time period, and Weibo is navigating intensifying competitors from video-first platforms like Douyin, that are drawing youthful customers and growing time-spent elsewhere.

In response, Weibo has leaned into creator-economy monetization, live-streaming, and vertical video—including instruments for influencer engagement, e-commerce integration, and richer analytics for manufacturers.

The platform’s position as a digital public sq. additionally makes it a spotlight of regulatory scrutiny. Chinese language authorities proceed to use strain on points starting from content material governance to information safety. In September 2025, Weibo was among the platforms cited in official warnings, highlighting its ongoing publicity to coverage dangers.

See also  Peer raises $10.5M for metaverse engine, launches 3D personal planets

Weibo’s push into AI R&D—exemplified by the discharge of VibeThinker-1.5B—alerts a shift in ambition. Past being a media platform, Weibo is positioning itself as a participant within the subsequent section of Chinese language AI growth, utilizing its capital reserves, consumer conduct information, and in-house analysis capability to pursue adjoining technical domains.

What It Means for Enterprise Technical Determination Makers

For engineering leaders and enterprise AI groups, VibeThinker’s launch has sensible implications for the whole lot from orchestration pipelines to price modeling.

A 1.5B-parameter mannequin that outperforms 100x bigger fashions on math and programming duties doesn’t simply save compute—it shifts the architectural stability. It permits LLM inference on constrained infrastructure, reduces latency on the edge, and lowers the barrier to entry for purposes that in any other case would have required API entry to closed, frontier-scale fashions.

That issues for enterprise ML leads making an attempt to deploy reasoning-capable brokers inside present programs, or for platform house owners tasked with integrating LLMs into automated workflows.

It additionally speaks to these operating reinforcement studying from human suggestions (RLHF) pipelines or managing inference optimization throughout hybrid cloud environments.

The mannequin’s post-training methodology—significantly its entropy-targeted reinforcement studying strategy—affords a roadmap for groups trying to refine smaller checkpoints as an alternative of counting on large-scale pretraining.

VibeThinker’s benchmark transparency and information decontamination steps additionally handle one other rising precedence in enterprise AI: auditability. Whereas its efficiency on general-knowledge checks nonetheless trails massive frontier fashions, its task-specific reliability makes it a beautiful candidate for managed environments the place correctness issues greater than protection.

In brief, VibeThinker-1.5B isn’t only a analysis milestone—it’s a powerful candidate for sensible enterprise use, deployment and learnings. It suggests {that a} new class of compact, reasoning-optimized fashions is viable for enterprise use circumstances that have been beforehand the area of far bigger programs. For organizations making an attempt to stability price, latency, interpretability, and management, it’s a great new choice to the lengthy, rising listing of Chinese language open supply choices.

Source link

TAGGED: Budget, DeepSeekR1, Model, Open, outperforms, posttraining, source, VibeThinker1.5B, Weibo039s
Share This Article
Twitter Email Copy Link Print
Previous Article Anthropic to Pour $50B into US Data Centers Anthropic to Pour $50B into US Data Centers
Next Article A Game-Changer for Data Centers? A Game-Changer for Data Centers?
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Gcore integrates StackPath’s API protection offering to improve its security solution

Edge AI, cloud, community, and safety options supplier Gcore has acquired StackPath’s internet utility and…

March 9, 2024

Anthropic signs billion-dollar deal with Google Cloud

US-based AI firm Anthropic has signed a major deal with Google Cloud that's stated to…

October 25, 2025

H5 Data Centers partners with DE-CIX

This strategic partnership empowers SAT-IX clients at H5 San Antonio to seamlessly join and peer…

February 20, 2025

OVHcloud Reinforces AI Inference with SambaNova Partnership

OVHcloud, a worldwide cloud participant and the main European cloud supplier, has made a strategic…

November 25, 2025

Gartner: $5.26 Trillion IT Spending in 2024, Fueled by Data Center Demand

Gartner's newest forecast anticipates world IT spending to succeed in $5.26 trillion in 2024, marking…

July 18, 2024

You Might Also Like

Google’s new framework helps AI agents spend their compute and tool budget more wisely
AI

Google’s new framework helps AI agents spend their compute and tool budget more wisely

By saad
BBVA embeds AI into banking workflows using ChatGPT Enterprise
AI

BBVA embeds AI into banking workflows using ChatGPT Enterprise

By saad
Ai2's new Olmo 3.1 extends reinforcement learning training for stronger reasoning benchmarks
AI

Ai2's new Olmo 3.1 extends reinforcement learning training for stronger reasoning benchmarks

By saad
Experimental AI concludes as autonomous systems rise
AI

Experimental AI concludes as autonomous systems rise

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.