
Kimi K2 0905 is the latest update to Moonshot AI’s large-scale Mixture-of-Experts (MoE) language model, which is well-suited for complex agent-like tasks. With its advanced coding and reasoning capabilities, and extended context length, it delivers outstanding performance in the field of artificial intelligence.
- Agent-like intelligence: It doesn’t just answer questions, it also performs actions. This includes advanced tool usage, reasoning, and code synthesis. It automatically understands how to use given tools to complete a task without having to write complex workflows.
- Long-context inference: Supports long-context inference of up to 256k tokens, which has been extended from the previous 128k.
- Coding: It has improved agent-like coding, with higher accuracy and better generalization across frameworks. It also offers advanced front-end coding with more aesthetic and functional outputs for web, 3D and related tasks. It performs well on coding benchmarks such as LiveCodeBench and SWE-bench.
- Reasoning and Knowledge: Achieves state-dependent performance in boundary knowledge, mathematics and coding among non-thinking models. It performs well on reasoning benchmarks such as ZebraLogic and GPQA.
- Tool Usage: Performs well on tool usage benchmarks such as Tau2 and AceBench. To strengthen tool invocation capabilities, the model can independently decide when and how to invoke its tools.
Links
Twitter: https://x.com/Kimi_Moonshot/status/1963802687230947698Kimi-K2: https://moonshotai.github.io/Kimi-K2/Hugging Face: https://huggingface.co/moonshotai/Kimi-K2-Instruct-0905Tech report: https://github.com/MoonshotAI/Kimi-K2/blob/main/tech_report.pdfUser Manual: https://platform.moonshot.ai/docs/introduction#text-generation-modelKimi Chat: https://www.kimi.com/Openrouter MoonshotAI: Kimi K2 0905: https://openrouter.ai/moonshotai/kimi-k2-0905Groq: https://groq.com/blog/introducing-kimi-k2-0905-on-groqcloud