Moonshot AI has discreetly implemented a major upgrade to its flagship model, introducing a new version named Kimi K2.5, marking a critical advancement for the company in multimodal capabilities and AI agent cluster collaboration. This upgrade was not accompanied by a public launch event but was instead rolled out directly via product updates, aiming to consolidate its leading position in China's fiercely competitive artificial intelligence market through substantive enhancements in technical prowess.
According to a company statement released on Tuesday, the latest K2.5 model adopts a native multimodal architecture, enabling it to process text, images, and video simultaneously from a single prompt. The model's most significant technical breakthrough lies in its "AI agent cluster" capability, which allows it to autonomously orchestrate up to 100 sub-agents working in parallel. This architecture reduces the execution time for complex tasks by up to 4.5 times compared to a single-agent configuration, significantly boosting processing efficiency.
This move comes as competition among China's top large language model developers intensifies. On the eve of competitor DeepSeek's anticipated major update, Moonshot AI is showcasing its technical reserves through this upgrade. Informed sources revealed that Moonshot AI recently completed a funding round, raising $500 million from investors including Alibaba and IDG Capital, achieving a post-money valuation of $4.3 billion, and is currently seeking a new funding round with a valuation of up to $5 billion.
Currently, Kimi K2.5 is available to users via the Kimi.com web version, App, and API platform. This upgrade not only strengthens its traditional advantage in long-context text processing but also, by introducing visual understanding and automated code generation tools, attempts to secure a more favorable position in enterprise-level applications and developer ecosystems.
Kimi K2.5 represents a leap from single-modality text processing to an all-in-one visual assistant in its fundamental architecture. The model was pre-trained using approximately 15 trillion mixed visual and text tokens, endowing it with deep visual comprehension capabilities beyond simple OCR text recognition.
According to official descriptions, users can directly upload complex circuit diagrams, handwritten mathematical formulas, or financial statements, and K2.5 can understand the underlying logic and perform analytical deductions. In the programming domain, the model demonstrates robust visual coding capabilities, enabling it to generate complete front-end interface code directly from image or video inputs, supporting visual debugging, thereby lowering the barrier for users to express development intent visually.
The core highlight of this update lies in the "AI agent cluster" paradigm. K2.5 incorporates Parallel Agent Reinforcement Learning (PARL) technology, allowing it to act as a coordinator that autonomously manages and orchestrates a cluster of up to 100 sub-agents.
When handling large-scale searches or complex workflows, the model can decompose tasks into parallelizable subtasks without pre-defining roles or workflows. For instance, when screening video creators in a specific field, K2.5 can simultaneously activate 100 sub-agents to conduct parallel searches and data aggregation. Internal test data indicates that this parallel processing mechanism not only reduces end-to-end runtime by 80% but also supports parallel workflows with up to 1,500 coordination steps, significantly surpassing the limitations of traditional single-agent systems.
Regarding performance metrics, Moonshot AI stated that K2.5 outperforms open-source counterparts on multiple benchmarks. The model showed outstanding performance on agent benchmarks such as HLE, BrowseComp, and SWE-Verified. Particularly in programming and logical reasoning, K2.5 has narrowed the gap with top-tier proprietary models, and its newly introduced automated coding tools are positioned to compete with Claude Code from Anthropic PBC.
For practical office scenarios, K2.5 demonstrates the ability to handle high-density, large-scale knowledge work. Internal "AI Office benchmark" tests showed that compared to the previous K2 Thinking version, the new model's performance improved by 59.3% on end-to-end tasks like processing documents, spreadsheets, and building financial models.
Moonshot AI was founded by former Tsinghua University professor Yang Zhilin, who has experience working on AI projects at both Meta Platforms Inc. and Google. Although the company is advancing its commercialization through subscription plans and enterprise services, it still faces fierce competition for market share from rivals like Zhipu and MiniMax Group Inc., the latter two having collectively raised over $1 billion through recent initial public offerings (IPOs) in Hong Kong.
Following the breakthrough success of the DeepSeek R1 model in early 2025, the "hundred-model battle" in China's LLM market has entered an elimination phase, with many smaller players falling behind due to difficulties in keeping pace with rapid technological upgrades and substantial funding requirements. By releasing K2.5 ahead of competitors and aligning it with a new high-valuation funding plan, Moonshot AI aims to demonstrate its sustained leadership in both technological iteration and capital attraction.