Updated FP8 version of Qwen3-30B-A3B thinking mode, with better tool use, coding, instruction following, logical reasoning and text comprehension capabilities
Fine-tuningDocs | Qwen3 30B A3B Thinking 2507 can be customized with your data to improve responses. Fireworks uses LoRA to efficiently train and deploy your personalized model |
On-demand DeploymentDocs | On-demand deployments allow you to use Qwen3 30B A3B Thinking 2507 on dedicated GPUs with Fireworks' high-performance serving stack with high reliability and no rate limits. |