Kimi K2 Instruct, a 1T parameter model with state of the art quality for coding, reasoning, and agentic tool use, is now available on Fireworks! Try now

Blog
Batch Api

Run bulk async workloads with Fireworks Batch API

Fireworks AI Batch API

With Fireworks’ Batch API, you can asynchronously run large volumes of requests on 1000+ open or finetuned models with no rate limits, 50% lower cost, and a 24-hour turnaround time.

This is helpful for use cases like:

  • Evaluations: Benchmark across models to identify the best model for your use case
  • Data generation: Generate bulk outputs using large models to fine-tune smaller models
  • Data augmentation: Create paraphrases, sentiment labels, or question-answer pairs at scale
  • ETL Pipelines and Daily bulk processing: Process large numbers of documents daily without worrying about rate limits

To use the Batch API, you simply upload your dataset in JSONL batch format and kick off a Batch API job. You can then check in on the status of your request, and retrieve the results once they are ready.

The Batch API has the following benefits:

  • High rate limits: There is no upper limit on the number of requests in a batch job. The dataset must be <500 MB in size.
  • Lower cost: 50% lower cost compared to typical Serverless (synchronous) API pricing
  • Fast processing: Results can be retrieved in a few hours, with a max turnaround time of 24 hours
  • Access 1000+ models: Run batch jobs on all major open models supported by Fireworks or your own fine-tuned models

Getting Started in 3 Steps

Prepare Your Dataset

Create a JSONL Batch format dataset, for example:

12
{"custom_id": "request-1", "body": {"messages": [{"role": "user", "content": "Tell me an interesting fact"}]}}
{"custom_id": "request-2", "body": {"messages": [{"role": "user", "content": "Tell me a joke"}]}}
1
firectl create dataset test-set ./my-local-folder/dataset.jsonl

Launch a Batch Job

Choose from 1,000+ models—including base models in our Model Library [link], public community models, and your own fine-tuned variants.

1234
firectl create bij \
--input-dataset-id test-set \
--model accounts/fireworks/models/qwen3-14b \
--output-dataset-id output-set

Download Results

Once the job completes, fetch your outputs in one command:

12
firectl download dataset output-set --output-dir ./output

Ready to unlock true AI scale?

Dive in with our Batch Inference Guide or API reference