Excited to launch a multi-year partnership bringing Fireworks to Microsoft Azure Foundry! Learn more

FINE TUNING

Train Your Own AI

Train, deploy, evaluate, and continuously improve custom models — all on one platform. No vendor sprawl. No infrastructure headaches.

Flywheel

Train

Fine-tune with SFT, DPO, or RFT via Training Agent, Managed Training, or the Training API. Each loop starts here — with fresh data from the previous cycle.

Deploy

One-click deployment to an auto-scaling inference endpoint. Your model is live in seconds — no separate serving stack, no format conversion.

Monitor

Track latency, accuracy, and cost in production. Identify where your model struggles — the gaps become your next training dataset.

Collect

Gather new data from real production traffic: user feedback, failure cases, edge cases, tool-call logs. Feed it back into Train to close the loop.

TRAINING API

Customer success

Notion logo dark

"By partnering with Fireworks to fine-tune models, we reduced latency from about 2 seconds to 350 milliseconds, significantly improving performance and enabling us to launch AI features at scale. That improvement is a game changer for delivering reliable, enterprise-scale AI"

Sarah Sachs
Sarah Sachs | AI Lead at Notion
Sourcegraph

"Fireworks has been a fantastic partner in building AI dev tools at Sourcegraph. Their fast, reliable model inference lets us focus on fine-tuning, AI-powered code search, and deep code context, making Cody the best AI coding assistant. They are responsive and ship at an amazing pace."

Beyang Liu Testimonial
Beyang Liu | CTO at Sourcegraph
rLLM

"The rLLM team is dedicated to pushing the boundaries of autonomous AI, which means our time is best spent on innovation rather than managing backend clusters. The Fireworks Training SDK lets us focus on our research instead of wrestling with infrastructure. The platform is fast, well-optimized, and just works."

rLLM
Kyle Montgomery & Sijun Tan | Core Contributors, rLLM at rLLM
Notion logo dark

"By partnering with Fireworks to fine-tune models, we reduced latency from about 2 seconds to 350 milliseconds, significantly improving performance and enabling us to launch AI features at scale. That improvement is a game changer for delivering reliable, enterprise-scale AI"

Sarah Sachs
Sarah Sachs | AI Lead at Notion
Sourcegraph

"Fireworks has been a fantastic partner in building AI dev tools at Sourcegraph. Their fast, reliable model inference lets us focus on fine-tuning, AI-powered code search, and deep code context, making Cody the best AI coding assistant. They are responsive and ship at an amazing pace."

Beyang Liu Testimonial
Beyang Liu | CTO at Sourcegraph
rLLM

"The rLLM team is dedicated to pushing the boundaries of autonomous AI, which means our time is best spent on innovation rather than managing backend clusters. The Fireworks Training SDK lets us focus on our research instead of wrestling with infrastructure. The platform is fast, well-optimized, and just works."

rLLM
Kyle Montgomery & Sijun Tan | Core Contributors, rLLM at rLLM
Notion logo dark

"By partnering with Fireworks to fine-tune models, we reduced latency from about 2 seconds to 350 milliseconds, significantly improving performance and enabling us to launch AI features at scale. That improvement is a game changer for delivering reliable, enterprise-scale AI"

Sarah Sachs
Sarah Sachs | AI Lead at Notion
THREE PRODUCTS, ONE PLATFORM

Start where you are. Go as far as you need.

Each product feeds into the same flywheel — train, deploy, improve, repeat.

Training Agent

Describe what you want. The agent does the rest.

For Product Managers & App Builders

From AI natives to enterprises, Fireworks powers everything from rapid prototyping to mission-critical workloads

Managed Training

Your data, your model, our infrastructure.

For ML Engineers

Use defined methods like SFT, DPO, and RFT. We provide reliable, scalable infrastructure for standard tuning with one-click deployment to production.

Training API

Own your model's intelligence.

For Advanced ML Platform Teams & Researchers

Full parameter tuning, 128k+ context window, custom loss functions, raw primitives. Push the frontier with complete algorithmic control on managed infrastructure.

Training AgentManaged TrainingTraining API
YOU BRINGData + description of what you wantFormatted data + method choiceYour training loop + loss functions
WE HANDLEData prep, model selection, evals, training, deploymentGPU provisioning, distributed training, checkpointing, scalingGPU execution, model parallelism, weight syncing, checkpointing, preemption recovery
ABSTRACTIONFully automatedMethod-controlledRecipes → SDK → raw primitives
LORA / FULL PARAMLoRALoRA and Full ParameterLoRA and Full Parameter
PRICINGPer job, confirmed upfrontPer token / GPU-hrPer GPU-hr
TRAINING METHODS

Three modalities, all feeding the same loop

Each method is purpose-built for a different stage of model maturity.

Supervised Fine-Tuning

Teach the model specific formats, tone, and domain knowledge from your labeled examples.

  • Classification & extraction
  • Tone & style alignment
  • Domain-specific Q&A

Reinforcement Fine-Tuning

Teach the model how to think, use tools, and solve multi-step problems with reward functions. The most powerful method for agentic AI.

  • Multi-turn agent workflows
  • Complex tool calling
  • Chain-of-thought reasoning

Preference Optimization

Align the model with human preferences and penalize undesirable behaviors without reward modeling.

  • Safety & guardrails
  • Hallucination reduction
  • Brand voice alignment
REINFORCEMENT FINE-TUNING

Improve model quality with reinforcement fine-tuning

Train frontier open-source models for multi-turn agents. User-friendly for app developers, powerful enough to surpass closed-source models in performance, quality, and tool interaction.

RFT chart

Train expert open models with just a few examples

RFT lets open models match frontier quality up to 10× faster, with just an evaluator and a few examples.

Fine tuning

Superior multi-turn agent performance

Customize your model for complex multi-turn agents, dramatically improving performance with internal tools and function-calling.

Fine tuning

No vendor lock-in, easy model swapping

Easily test with and train new models as the frontier evolves. Keep your same evaluator and pipeline, just swap the base model.

Fine tuning

Enterprise-grade security and reliability

Choose flexible deployment methods, including full management by Fireworks or self-managed rollouts within your own environments.

Fine tuning
SUPERVISED FINE-TUNING

Fine-tune with your own data

Customize model behavior by fine-tuning with your own data. Fireworks makes supervised fine-tuning fast, reliable, and cost-effective with an optimized training stack. Train large, state-of-the-art models using advanced methods like quantization-aware training to achieve ideal results.

  • Optimized training stack — Purpose-built for speed and reliability at scale.
  • Quantization-aware training — Train large models efficiently without sacrificing quality.
  • LoRA & full parameter — Choose the right tradeoff between speed and model fidelity.
DPO vs GRPO
PREFERENCE OPTIMIZATION

Align models to human preferences

Direct Preference Optimization (DPO) and ORPO let you shape model behavior using preference pairs — no reward model required. Ideal for safety, brand voice, and reducing hallucinations without complex RL setups.

  • No reward model needed — Simpler pipeline than RLHF — just chosen/rejected pairs.
  • Safety & guardrails — Penalize harmful or off-brand outputs directly.
  • Hallucination reduction — Teach the model what not to say, not just what to say.

WHY FIREWORKS

The only platform that combines full-spectrum training and built-in inference

Serve personalized models at scale

100s

Models per GPU

1-click

Deployment

Zero

Extra infra cost
Multi-lora
ALTERNATIVEEXAMPLESTHE LIMITATIONFIREWORKS ADVANTAGE
Closed ModelsOpenAI, AnthropicNo weight ownership. High cost. Zero portability. No retraining loop.[check] Open-source models you fully own. Retrain and redeploy continuously.
Training-OnlyFragmented vendorsTrain here, serve elsewhere. Every iteration pays a migration tax.[check] Unified platform. Training completes → model is live → collect data → retrain.
Cloud-NativeAWS, GCPTraining and inference are separate silos. No open model expertise.[check] Model-agnostic. 1-click hot-loading from training to inference.
START BUILDING TODAY

Start the flywheel today

From first experiment to continuously improving production models — all on one platform.