The New China AI Trifecta

The video highlights the emergence of three leading Chinese AI labs—Moonshot AI, ZAI (Zhipu AI), and MiniMax—that are rapidly advancing open-source large language models with innovative techniques and a focus on practical, real-world applications. These labs are challenging established Western AI giants by excelling in new benchmarks, emphasizing usability, and fostering open-source collaboration.

The video discusses the rapid advancements in open-source large language models (LLMs) coming out of China, highlighting a new “AI trifecta” of leading research labs: Moonshot AI, ZAI (Zhipu AI), and MiniMax. Over the past several months, these labs have released a series of state-of-the-art models, each employing different techniques and excelling in various areas. Notably, three of these labs have reached the top of the field for the first time, challenging established players like Google and OpenAI with their innovative approaches and impressive results.

A major trend in recent model development is the shift from focusing solely on benchmarks for reasoning and raw knowledge to prioritizing practical, application-driven capabilities. The industry has moved away from older benchmarks like MMLU and ARC Challenge to more specialized and advanced ones such as SweetBench and LifeCodeBench. This shift has allowed smaller labs to compete directly with tech giants, as models are now evaluated more on real-world usability and user experience rather than just academic metrics.

Moonshot AI is highlighted as a particularly innovative lab, known for its risk-taking research and rapid progress since its founding in 2023. Their Kim K2 series introduced novel techniques like replacing the Adam optimizer with Muon and developing hybrid attention mechanisms for improved long-context performance. Moonshot AI also stands out for its quantization-aware training, ensuring that their models perform efficiently in real-world settings, and for its open research culture, exemplified by detailed technical blogs from its researchers.

ZAI (Zhipu AI), the oldest of the three, has evolved from a university research group into a major player with a focus on practical, agentic AI. Their GLM series, particularly GLM 4.7, has achieved state-of-the-art results while being significantly smaller and more efficient than competitors. ZAI’s models are designed for real-world coding and tool use, with innovations like switching from JSON to XML for code representation. Their open-source approach has enabled high-speed deployment on specialized hardware, and their competitive pricing positions them as a strong alternative to expensive proprietary models.

MiniMax, the third lab in the trifecta, initially focused on AI roleplay and generative media before pivoting to LLMs in 2025. Their early models experimented with linear attention for long context windows, but after discovering limitations in reasoning, they quickly shifted to more effective attention mechanisms. The result was MiniMax M2, which now tops several open-source benchmarks for instruction following and agentic tasks. Despite a lower valuation compared to its peers, MiniMax’s agility and technical achievements make it a formidable contender. Collectively, these three labs represent a new wave of Chinese AI development, emphasizing practical applications and open-source collaboration, and are poised to challenge the dominance of Western AI giants in both research and real-world deployment.