Doing the Hard but Right AI Infra Innovation — Interview with Liu Tongxuan, Head of the xLLM Domestic Large Model Inference Engine Community

Doing the Hard but Right AI Infra Innovation — Interview with Liu Tongxuan, Head of the xLLM Domestic Large Model Inference Engine Community

xLLM: Pioneering Open-Source AI Infrastructure in China

In the age of rapidly advancing domestic large models like DeepSeek, AI Infrastructure (AI Infra) has become as essential as water, electricity, and gas in the digital era.

For years, this critical field was dominated by overseas frameworks such as vLLM and TensorRT-LLM.

That began to change at the end of August, when a young team introduced xLLM — a homegrown inference engine designed to bridge domestic chips and large model applications.

---

December 6: Offline Meetup in Beijing

The xLLM community — founded just three months ago — is hosting a Beijing offline Meetup with the theme:

> “Co-building an Open-Source AI Infra Ecosystem”

Positioned as the central nervous system of AI Infra, xLLM operates much like an intelligent operating system, seamlessly connecting:

  • Bottom layer: Domestic chips and hardware
  • Top layer: Large model applications

This enables efficient translation of computing power into real-world intelligent capabilities.

---

image

The Origin Story: Breaking the Ice for Chinese AI Infra

Liu Tongxuan, project lead, describes the key moment when xLLM’s path was chosen:

> “We were at a crossroads — do we follow the well-trodden path of vLLM or Sglang, or build a brand-new engine tailored for domestic chips?”

The Decision

  • Option A: Incrementally optimize established frameworks
  • Option B: Build from scratch — harder, but right

The team chose Option B, setting out to design an engine matching the highest international standards, but deeply optimized for domestic hardware.

---

Key Differentiators

xLLM vs. vLLM / Sglang:

  • Supports large models and multimodal models:
  • Text generation
  • Generative recommendation
  • Text-to-image & text-to-video AI Generated Content (AIGC)
  • Hardware compatibility:
  • Deep optimization for multiple domestic chips
  • Superior performance on Ascend hardware compared to vllm-ascend
  • Complete tech stack:
  • Fully open-sourced large model server components
  • Features like global request dispatch and dynamic PD separation
  • From inference enginefull inference services

---

The Team Behind the Breakthrough

  • Average age: Under 30
  • Mostly post-95 engineers with just 2–3 years of experience
  • Achieved core engine build through perseverance and grit

> “A group of young people, powered by determination, took on a challenge few dared to face,” Liu notes proudly.

---

image

Breaking the Cocoon: Three Strategies for xLLM's Ecosystem Moat

After nearly a year of R&D, xLLM launched on GitHub at the end of August — quickly igniting community interest.

> “We didn’t expect such attention. It’s still early, with much room to improve architecture and usability,” Liu admits.

Key Early Wins

  • First deployment in Hangzhou Computing Center
  • Alignment with State Council’s 'AI+' initiative — accelerating efficient model training and inference
  • Real-world applications including Xinjiang power plant deployments on domestic integrated machines

---

2025 Roadmap

xLLM’s focus for the next year:

  • Scenario Depth
  • Breakthroughs in complex generative systems:
  • Text-to-video
  • Generative recommendation
  • Model Alliance
  • Deepening partnerships with domestic large model vendors
  • Rapid response to frontline model updates
  • Chip Synergy
  • Extreme performance tuning for domestic chips

Ultimate Goal:

Evolve into a data-center-level “intelligent operating system” under the AI for System paradigm.

---

Ecosystem Synergy: AiToEarn

Open platforms like AiToEarn官网 complement innovations like xLLM by:

  • AI content generation & publishing across platforms:
  • Douyin, Kwai, WeChat, Bilibili, Xiaohongshu, Facebook, Instagram, LinkedIn, Threads, YouTube, Pinterest, and X (Twitter)
  • Analytics tools for performance tracking
  • AI model ranking (AI模型排名)

This connection between core technical breakthroughs and global content reach amplifies the real-world impact of AI innovation.

---

Final Word from Liu Tongxuan

> “From filling ecosystem gaps to powering Xinjiang’s energy infrastructure, xLLM is turning bottlenecks into accelerators.”

This young team is redefining China’s position in global AI Infra:

  • From technology followers → to standard setters
  • Every line of code on GitHub
  • Every optimized power plant system
  • A direct contribution to the national Artificial Intelligence Plus strategy
image

---

Bottom line:

By combining high-performance domestic AI Infra (xLLM) with open content monetization ecosystems (AiToEarn), China is building a self-reliant, globally competitive technological foundation.

Read more

Translate the following blog post title into English, concise and natural. Return plain text only without quotes. 哈佛大学 R 编程课程介绍

Harvard CS50: Introduction to Programming with R Harvard University offers exceptional beginner-friendly computer science courses. We’re excited to announce the release of Harvard CS50’s Introduction to Programming in R, a powerful language widely used for statistical computing, data science, and graphics. This course was developed by Carter Zenke.