AI news

Huawei launches "near-trillion-scale MoE inference" with two killer optimization technologies released as open source

AI news

Huawei launches "near-trillion-scale MoE inference" with two killer optimization technologies released as open source

Machine Heart Report: Ultra-Large MoE Inference Breakthroughs --- 2025 Landscape: Inference Efficiency Takes Center Stage As 2025 concludes, large AI models have evolved from niche tools into foundational infrastructure powering enterprise systems. In this shift, inference efficiency has become the critical factor for scalable deployment. For ultra-large-scale MoE (Mixture-of-Experts) models,