What happened
Chinese companies, exemplified by DeepSeek AI, are developing efficient AI models, such as DeepSeek-V3, which utilise Mixture of Experts (MoE) architectures. DeepSeek-V3, with 671 billion parameters, activates only 37 billion per token during inference, enabling comparable performance to larger models like GPT-4 at a fraction of the training cost and computing power. This contrasts with the US focus on large, complex models and AGI, as China prioritises practical, state-driven AI deployment.
Why it matters
The proliferation of highly efficient AI models, developed with significantly reduced computational and financial overheads, introduces a new competitive constraint for organisations relying on traditional large-model development paradigms. This increases due diligence requirements for R&D and procurement teams to assess the true cost-performance ratio of AI solutions, and for strategic planning to adapt to a rapidly evolving, resource-optimised AI landscape. It also creates a visibility gap regarding the optimal investment in AI infrastructure and model development, as established benchmarks for scale and cost are being redefined.
Subscribe for Weekly Updates
Stay ahead with our weekly AI and tech briefings, delivered every Tuesday.




