Huh? Weibo’s $7,800 Trained Model Surpasses DeepSeek-R1 in Math

Huh? Weibo’s $7,800 Trained Model Surpasses DeepSeek-R1 in Math

VibeThinker: Weibo’s Disruptive Entry into the AI Model Arena

When the AI industry is obsessed with a “parameter race”, Weibo AI has unveiled an unexpected and imaginative path — reshaping the battlefield for large models.

A Lightweight Breakthrough

Weibo recently launched VibeThinker, its first self-developed open-source large model. Despite having only 1.5 billion parameters, this lightweight model outperformed the massive DeepSeek R1 (671 billion parameters) in top-tier international math competition benchmarks.

Even more remarkable, VibeThinker’s post-training cost was just $7,800dozens of times cheaper than DeepSeek-R1 or MiniMax-M1.

This achievement challenges traditional evaluation standards and hints at an industry shift from “scale race” to “efficiency revolution.”

image

---

Rethinking Model Size

Historically, parameter count was the dominant metric for AI capability. The prevailing belief was:

  • Complex reasoning emerges only beyond 100B parameters.
  • Small models cannot handle highly complex tasks.

VibeThinker proves otherwise. By starting small and applying innovative training strategies, Weibo answered “yes” to the possibility of unlocking hidden reasoning skills in compact models.

---

Key Innovation: Spectrum-to-Signal Principle (SSP)

While most follow the Scaling Law (“bigger = smarter”), Weibo AI optimized architecture and training, introducing the SSP training method. This small-scale contender defeated giants hundreds of times larger.

Upon release, VibeThinker drew global AI research attention, with surprising benchmark performances in mathematics and programming.

image

△ HuggingFace spotlighting the VibeThinker paper

---

Benchmark Results

Mathematics (AIME24, AIME25, HMMT25)

  • Surpassed DeepSeek-R1-0120 (671B params)
  • Nearly matched MiniMax-M1 (456B)
  • Reached Gemini 2.5 Flash and Claude Opus 4 levels

Programming (LiveCodeBench v6)

  • Matched much larger models like Mistral AI’s Magistral-Medium-2506

Conclusion: With refined algorithms and training, small models can rival — even exceed — large-scale reasoning performance.

---

Intended Use Cases

Currently, VibeThinker is experimental and focuses on complex mathematics and competitive coding.

💡 Not optimized for daily conversations — best suited for technical and high-intelligence scenarios.

---

Breaking Cost Barriers

Training costs have long hindered AI adoption. VibeThinker’s advantage lies in extreme cost efficiency.

Industry Costs in 2025

  • MiniMax M1: $535,000 (3 weeks, 512 H800 GPUs)
  • DeepSeek R1: $294,000 post-training, ~$6M for base LLM

VibeThinker’s Costs

  • Total compute time: 3,900 GPU hours (SFT + RL stages)
  • Total cost: USD 7,800
image

Result: Achieves performance comparable to $300K–$500K models at 30–60× cost efficiency.

---

Democratizing AI Research

Such efficiency means powerful AI is no longer exclusive to tech giants. SMEs, research institutes, and universities can now access state-of-the-art reasoning tools — fueling an open, diverse, vibrant AI future.

---

Weibo’s AI Application Ecosystem

Weibo is integrating AI deeply into its platform:

2024 Milestones

  • Developed the “Zhiwei” large language model
  • Received regulatory approval
  • Launched:
  • Weibo Smart Search
  • Content Summarization
  • AI Interactive Accounts

---

Flagship AI Products

  • Analyzes vast content
  • Builds a trusted knowledge graph
  • Understands emotion & context
  • 50M+ MAU (June)

Comment Robert

  • Sharp-tongued → warmer, smarter personalities
  • Nearly 2M followers
  • Showcases potential of AI comment assistants

---

Future Strategy: Data Empowerment

Weibo plans to:

  • Leverage unique vertical data (e.g., psychology)
  • Build models attuned to public sentiment
  • Deliver emotionally aware social services

VibeThinker will become the core engine for future AI products like Weibo Smart Search — enhancing UX, breaking scene boundaries, and supporting a “super social ecosystem.”

---

Efficiency Gains for Weibo

  • Lower compute costs for intelligent search
  • Reduced AI processing costs in real-time interactions
  • Scale AI capabilities without resource strain

---

Synergy with Open-Source Platforms

Platforms like AiToEarn embrace this efficiency trend, providing creators tools to:

  • Generate AI content
  • Publish across platforms (Douyin, Weibo, YouTube, LinkedIn, etc.)
  • Analyze engagement
  • Rank AI models

Learn more:

---

Final Takeaway: VibeThinker’s blend of small scale, high reasoning power, and low cost signals a turning point for AI — where smarter, cheaper, and more accessible becomes the new frontier.

---

Would you like me to also create a clear comparison table showing VibeThinker vs DeepSeek vs MiniMax performance & cost? That could make the rewrite even more visually impactful.

Read more

Translate the following blog post title into English, concise and natural. Return plain text only without quotes. 哈佛大学 R 编程课程介绍

Harvard CS50: Introduction to Programming with R Harvard University offers exceptional beginner-friendly computer science courses. We’re excited to announce the release of Harvard CS50’s Introduction to Programming in R, a powerful language widely used for statistical computing, data science, and graphics. This course was developed by Carter Zenke.