Kimi K2 0905, a new state of the art open weight models for agentic reasoning, tool use, and coding, is now available! Try Now

Kimi K2 Instruct

Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities.

Try Model

Fireworks Features

Fine-tuning

Kimi K2 Instruct can be customized with your data to improve responses. Fireworks uses LoRA to efficiently train and deploy your personalized model

Learn More

Serverless

Immediately run model on pre-configured GPUs and pay-per-token

Learn More

On-demand Deployment

On-demand deployments give you dedicated GPUs for Kimi K2 Instruct using Fireworks' reliable, high-performance system with no rate limits.

Learn More

Kimi K2 Instruct is a 1 trillion parameter Mixture-of-Experts (MoE) language model developed by Moonshot AI. It activates 32 billion parameters per forward pass and is instruction-tuned for general-purpose chat, agentic reasoning, and tool use. It is available on Fireworks AI for high-performance inference.

Info & Pricing

Model Type

LLM

Context Length

131072

Serverless

Available

Fine-Tuning

Available

Pricing Per 1M Tokens Input/Output

$0.6 / $2.5