Latest Qwen3 coder model, 30B with 3B active parameter model
Fine-tuningDocs | Qwen3 Coder 30B A3B Instruct can be customized with your data to improve responses. Fireworks uses LoRA to efficiently train and deploy your personalized model |
On-demand DeploymentDocs | On-demand deployments give you dedicated GPUs for Qwen3 Coder 30B A3B Instruct using Fireworks' reliable, high-performance system with no rate limits. |
Qwen3-Coder 30B A3B Instruct is a Mixture-of-Experts (MoE) instruction-tuned coding model developed by Qwen (Alibaba Group). It features 30.5 billion total parameters with 3.3 billion active per forward pass, and is trained for advanced code reasoning, agentic systems, and browser-integrated coding tasks.
The model is designed for:
The model natively supports a context window of 262,144 tokens (262.1K).
The full 262.1K token window is usable in on-demand deployments on Fireworks, which provide dedicated GPU access.
The recommended output length is up to 65,536 tokens, constrained by the 262.1K total context limit.
<think> tokensYes. Fireworks supports LoRA-based fine-tuning for this model using its RFT (Reserved Fine-Tuning) infrastructure.
Fireworks charges based on total input + output tokens, respecting the 262.1K context limit.
The model is released under the Apache 2.0 license, which permits unrestricted commercial use.