Rakuten AI: Efficiency prioritised

Rakuten AI: Efficiency prioritised

23 December 2025

Rakuten Group is expanding its AI division, now comprising 1,000 employees, and is focusing on cost-efficient AI model development. Led by AI chief Ting Cai, the company utilises thousands of Nvidia chips to support its AI initiatives. Rakuten's AI strategy prioritises minimising costs while augmenting the company's various businesses.

Rakuten AI 3.0, the company's latest large language model (LLM), uses a Mixture of Experts (MoE) architecture with approximately 700 billion parameters, activating only around 40 billion parameters per token. This design lowers computing requirements during inference compared to dense models. Rakuten claims its new model is 90% cheaper to run than comparable LLMs, though this has not been independently verified. The company reported that AI contributed ¥10.5 billion to its operating income in 2024 and aims to double this figure in 2025.

Model training occurred on an in-house multi-node GPU cluster within a secure environment, keeping all data internal. Rakuten AI 3.0 achieved a Japanese MT-Bench score of 8.88, outperforming GPT-4o (8.67) and Rakuten's previous model. The model will be released as an open-weight model in Spring 2026.

AI generated content may differ from the original.

Published on 23 December 2025
aimachinelearningllmrakuten
  • AI Minds Meet, Discuss Future

    AI Minds Meet, Discuss Future

    Read more about AI Minds Meet, Discuss Future
  • Luma AI: Video Generation Evolved

    Luma AI: Video Generation Evolved

    Read more about Luma AI: Video Generation Evolved
  • LeCun's AI Startup Targets €3B

    LeCun's AI Startup Targets €3B

    Read more about LeCun's AI Startup Targets €3B
  • Amazon Bets Big on AI

    Amazon Bets Big on AI

    Read more about Amazon Bets Big on AI