Fireworks AI now integrates with AWS AgentCore, enabling developers to deploy AI agents with optimized inference on secure, serverless AWS infrastructure. Build locally, deploy globally with enterprise-grade security and automatic scaling.
AWS AgentCore Runtime provides serverless infrastructure purpose-built for AI agents. It solves the operational complexity of deploying dynamic AI agents at scale by offering:
This eliminates the need to manage containers, orchestration, or scaling infrastructure while maintaining enterprise security requirements.
Fireworks AI delivers the fastest, highest quality inference engine for agentic workloads. We provide optimizations like adaptive caching and speculative decoding that are critical for multi-turn agent interactions. Combined with AgentCore's serverless deployment, you get:
To demonstrate this integration, we built two cookbooks using AgentCore Runtime and AgentCore Code Interpreter. These agents can read files, generate python code, run the code and interpret the results. The agents use state of the art open source models Kimi K2 0905 and Qwen 3 Coder 480B respectively.
The architecture shows the complete flow: develop your agent locally with your choice of models and frameworks, configure it with a Docker file, and deploy to AgentCore Runtime via AWS CodeBuild. Once deployed, users can invoke the agent through the Runtime endpoint with all infra handled serverlessly by AWS.
See our full documentation and both cookbooks in our documentation page.
Ready to build production AI agents?
The integration supports serverless inference, fine-tuned models, and on-demand deployments. For enterprise deployments leveraging existing AWS compute please reach out to [email protected]
We will continue to expand our AWS AgentCore integration with additional cookbooks and deeper platform integrations. Stay tuned for more examples covering multi-agent systems, custom tool integration, and production deployment patterns.