We’re thrilled to announce Fireworks AI has now made Amazon SageMaker available as a Bring Your Own Compute (BYOC) deployment option.
This integration allows developers and enterprise ML teams to train models using SageMaker, and leverage Fireworks’ high-performance, low-latency inference platform for model serving — all within their existing AWS environment.
As organizations embrace Generative AI at scale, they’re hitting the same roadblocks: training and experimentation in SageMaker is made seamless, but production-grade inference requires custom platform, performance tuning, and ongoing cost management.
That’s where Fireworks comes in.
Fireworks is the fastest inference and AI platform that enables customers to build magical AI Applications. Fireworks offers:
Now, with Amazon SageMaker as a deployment option, customers can benefit all within their AWS environment.
With Fireworks' deployment on Amazon SageMaker, you can:
All of this while retaining full control over your data, compliance boundaries, and AWS resource governance.
Train models separately on SageMaker
Save the trained model in S3
Deploy Fireworks as a container on SageMaker
Expose the model via a SageMaker endpoint
We designed this feature with enterprise needs in mind. That means:
Whether you're deploying an open-source LLM or scaling a fine-tuned model to thousands of users, Fireworks BYOC ensures your workloads are fast, reliable, and cost-efficient.
Fireworks deployment on Amazon SageMaker is available now in private preview.
Want early access? Request access here.
This launch is a big step toward our mission: making world-class AI infrastructure accessible and scalable for everyone. By integrating deeply with Amazon SageMaker, we're bridging the gap between model development and high-performance inference.
We can't wait to see what you build!