In a major move to reclaim its lead in the Chinese AI landscape, Moonshot AI officially released Kimi K2.5 on January 27, 2026.
The new flagship model marks a pivot from simple chat to high-complexity “agentic” workflows. By introducing native multimodality and a first-of-its-kind “Agent Cluster” capability, Moonshot is positioning Kimi as a direct rival to Western models like Claude 4.5 and GPT-5.2.

1. The “Agent Cluster” Breakthrough
The standout feature of Kimi K2.5 is its transition from a single AI assistant to an Agent Cluster paradigm.
- Autonomous Teams: When faced with a massive task (e.g., auditing 1,000 legal documents or coding a full-stack app), K2.5 can autonomously instantiate hundreds of “avatars” or domain-specific agents.
- Parallel Execution: These agents work in parallel, collaborating and cross-checking each other. Moonshot claims this improves operational efficiency by 10x to 100x for large-scale search and data synthesis scenarios.
- Tool Mastery: The model has been specifically trained to navigate advanced office suites (Excel, PPT, Word) with professional-level precision.
2. Native Multimodality: “Visual Causal Flow”
Kimi K2.5 moves away from being a text-first model with a vision “plug-in.” It is a native multimodal model trained on approximately 15 trillion mixed tokens.
- Screen-to-Code: A major highlight for developers is K2.5’s ability to watch a screen recording or look at a UI screenshot and reproduce the underlying logic and professional code instantly.
- MoonViT Encoder: Utilizing the proprietary 400M parameter MoonViT encoder, the model can understand complex video workflows and design drafts with a 91% accuracy rate in visual-reasoning benchmarks.
3. Technical Specifications & Benchmarks
Built on the Kimi-K2-Base, the K2.5 model continues the use of a massive Mixture-of-Experts (MoE) architecture.
| Feature | Specification (Jan 2026) |
| Total Parameters | 1.0 Trillion |
| Activated Parameters | 32 Billion (per forward pass) |
| Context Window | 256,000 Tokens |
| Architecture | Mixture-of-Experts (MoE) with MLA Attention |
| Inference Modes | “Instant” (Low Latency) & “Thinking” (Deep Reasoning) |
| Licensing | Open Source (Modified MIT License) |
Benchmark Comparison
In Moonshot’s self-reported data, Kimi K2.5 (Thinking Mode) shows it can now trade blows with the world’s most advanced proprietary systems:
- HLE-Full (w/ Tools): 50.2 (Beating GPT-5.2’s 45.5).
- SWE-bench Verified: Top-tier performance for open-source models, rivaling Claude 4.5 in bug-fixing tasks.
4. Valuation and Strategic Position
The launch comes as Moonshot AI’s valuation has surged to $4.8 billion, following a fresh $500 million funding round led by Alibaba and Tencent in late January.
- Kimi Code Release: Alongside the main model, Moonshot launched Kimi Code, a dedicated programming assistant integrated with VSCode and Cursor.
- Ecosystem Compatibility: The Kimi API remains 100% compatible with OpenAI’s SDK, making it seamless for international developers to swap in K2.5 for regional Chinese or cost-effective agentic workloads.
Conclusion: The Agentic Future
Kimi K2.5 represents a clear signal that the “Chatbot Era” is evolving into the “Agent Era.” By open-sourcing a 1-trillion parameter model with native swarm capabilities, Moonshot AI is providing the infrastructure for a new generation of autonomous enterprise software. The model is available starting today on the Kimi App, website, and through the Moonshot Open Platform API.


