Honghao Wang

Breaking the Terabyte-Scale Model “Memory Wall”: Collaborative Compression Framework Fits 1.3TB MoE Model into a 128GB Laptop

AI news

Breaking the Terabyte-Scale Model “Memory Wall”: Collaborative Compression Framework Fits 1.3TB MoE Model into a 128GB Laptop

Collaborative Compression: Breaking the Memory Wall for Trillion-Parameter MoE Models This article introduces the Collaborative Compression framework, which — for the first time — successfully deployed a trillion-parameter Mixture-of-Experts (MoE) large model on a consumer-grade PC with 128 GB RAM, achieving over 5 tokens/second in local inference. Developed by the Moxin

Translate the following blog post title into English, concise and natural. Return plain text only without quotes.

AI 时代的技术团队如何进行批判性思考

AI Playbooks

Translate the following blog post title into English, concise and natural. Return plain text only without quotes. AI 时代的技术团队如何进行批判性思考

In a time where AI can generate code, design ideas, and occasionally plausible answers on demand, the need for human critical thinking is greater than ever. Even the smartest automation can’t replace the ability to ask the right questions, challenge assumptions, and think independently at this time. This essay