Try the latest GLM-4.6 with extended context, superior coding, and refined intelligence. Now available on-demand

Blog
Fireworks Amd Ai Infrastructure Partnership

Fireworks and AMD partner to power the next gen of AI infrastructure on AMD Instinct™ GPUs

fireworks amd

Fireworks and AMD partner to power the next gen of AI infrastructure on AMD Instinct™ GPUs

Fireworks and AMD have entered into a multi-year strategic agreement to optimize AMD Instinct™ GPUs and accelerate adoption across AI-native companies, developers, and enterprises. We’re excited to share this new chapter in Fireworks’ mission to power the next generation of AI inference workloads.

Our collaboration brings together AMD’s leadership in high-performance computing and Fireworks’ advanced AI stack to deliver scalable, production-grade AI systems that run inference faster, with the best quality, for the most efficient cost. For every organization and workload, there is a sweet spot where price, performance, and speed meet a technical and business outcome. By partnering with AMD, Fireworks provides best-in-class optimization technology alongside AMD Instinct™ GPUs.

Unlocking Superior AI Performance Economics for AI Leaders

From model-serving runtimes to training frameworks, Fireworks is working closely with AMD to optimize every layer of our software stack for AMD Instinct™ MI325X and MI355X accelerators..

Tuning the Fireworks stack for AMD’s advanced architecture, customers can achieve:

  • Lower Total Cost of Ownership: Improved efficiency at the model level reduces the number of GPUs needed for production workloads.
  • Better Throughput and Latency: Optimizations at the kernel and runtime level ensure fast inference and training cycles.
  • Rapid Deployment at Scale: With pre-optimized models and tooling, customers can go from experimentation to production in a fraction of the time.

Powering the Next Generation of AI Infrastructure

Under the multi-year agreement, AMD will provide Fireworks with access to AMD Instinct MI325X and MI355X GPUs through preferred cloud service providers (CSPs). This collaboration ensures that every improvement in AMD’s hardware roadmap translates directly into better performance and efficiency for Fireworks customers who want to train, serve, and scale models on their own terms.

The Path Ahead

By combining AMD’s cutting-edge accelerators with Fireworks’ purpose-built AI software stack, we’re helping shape the next wave of faster, more efficient, and more open source AI innovation.

Stay tuned for benchmarks, developer resources, and new model integrations optimized for AMD Instinct GPUs.