Fireworks RFT now available! Fine-tune open models that outperform frontier models. Try today

Blog
Nvidia Nemotron Nano2

Unlock Advanced Reasoning with NVIDIA Nemotron Nano 2 Models on Fireworks AI

Unlock Advanced Reasoning with NVIDIA Nemotron Nano 2 Models on Fireworks AI

We're excited to collaborate with NVIDIA to bring their groundbreaking NVIDIA Nemotron Nano 2 9B models to the Fireworks AI platform. NVIDIA Nemotron is a family of open models, datasets, and technologies that unlock developers to build highly efficient and accurate specialized agents. The Nemotron models are trained from scratch by NVIDIA, designed as a unified model for both reasoning and non-reasoning tasks. It responds to user queries and tasks by first generating a reasoning trace and then concluding with a final response. Nemotron Nano 2 is built on a hybrid Mamba-Transformer architecture that delivers expert-level reasoning with unprecedented efficiency.

Deploying Nemotron models on Fireworks can unlock more powerful use cases for developers. In scientific research, Nemotron models acts as the ideal lab partner that can easily process dense papers, explain complex concepts, and rapidly generate new hypotheses. For Search and Code Understanding, its comprehensive long-context up to 128K enables developers to digest larger code repositories or pull information from many different sources, all while maintaining context across different scenarios. Ultimately, from scientific research to automated agentic workflows, Nemotron delivers the high-level complex reasoning skills necessary to provide reliable, actionable decision support where simpler models often struggle.

NVIDIA Nemotron Nano 2: A Breakthrough in Efficient Reasoning

Traditional Transformers face a fundamental bottleneck: self-attention scales quadratically with sequence length. What does this mean? Transformers models often hit a wall when dealing with longer documents or more complex large datasets. Self-Attention works by comparing every single word to every other word in the text. For example if a document is twice as long, the work doesn't just double—it quadruples. This concept is known as O(L²) scaling, and it quickly creates massive computing and memory bottlenecks for anything requiring truly long context.

NVIDIA's Nemotron models solve this challenge through an ingenious hybrid Mamba-Transformer design that provides a major leap in efficiency. The architecture rather than comparing every token, operates more like a selective human brain that’s focus is:

  • Attention on demand: Only a small fraction (around 8%) of the layers use the heavy-lifting self-attention mechanism, and they are positioned precisely where the model needs to link distant ideas.
  • Speed and stability: The majority of layers use efficient Mamba-2 and FFN blocks, which ensure that the computational work is constant for every token.

By blending these two approaches, Nemotron provides the necessary attention for accuracy while dramatically reducing computational overhead and ensuring stable, constant memory usage—a game-changer for long-context inference. This results in transformer-level accuracy with faster inference on long sequences, expert-level reasoning, and efficient scaling for production deployments.

Available in two sizes (Nemotron-Nano-9B-v2 and Nemotron-Nano-12B-v2), both models combine breakthrough architecture with industry-leading performance on Fireworks infrastructure.

Developers can unlock significant advantages by migrating to NVIDIA Nemotron Nano models on Fireworks:

  1. Unlock expert level reasoning capabilities with high efficiency and optimized accuracy, that can lead to reduced costs
  2. Go from experimentation to production quickly and seamlessly. With Fireworks, you can effortlessly scale your production workloads globally, accessing cutting-edge hardware across 18+ regions and 8 providers without managing infrastructure. Optimize for superior speed, greater capacity, and reduced costs through on-demand GPUs, auto-scaling, and pay-per-second pricing.

Read more about the Nemotron architecture in NVIDIA's Technical Whitepaper.

Getting Started on Fireworks

📓 Complete Example: To accelerate your development, we have created an extensive cookbook for your end-to-end development: here.

Below we have included some key snippets from the cookbook on how to get started.

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950
### Step 1: Create a Deployment
### Deploying Nemotron models on Fireworks is straightforward using the `firectl` CLI:
```bash
# Install firectl
brew tap fw-ai/firectl && brew install firectl
# Login to Fireworks
firectl login
# Create Nemotron 9B deployment
firectl create deployment accounts/fireworks/models/nvidia-nemotron-nano-9b-v2
### Step 2: Use the API
Once deployed, access your model through the standard OpenAI-compatible API:
```python
from fireworks import LLM
# Connect to your existing deployment
llm = LLM(
model="nvidia-nemotron-nano-9b-v2", # The model your deployment is running
deployment_type="on-demand",
id="your-deployment-id", # Your deployment ID
)
# Use OpenAI-compatible chat completions
response = llm.chat.completions.create(
messages=[{"role": "user", "content": "Say this is a test"}]
)
print(response.choices[0].message.content)
```

Hardware Requirements

NVIDIA-Nemotron-Nano-9B-v2 and NVIDIA-Nemotron-Nano-12B-v2 can both run on 1x H100 GPU with excellent performance.

How Does It Perform?

To evaluate reasoning capabilities, we tested all models on a random sample from GPQA Diamond–a rigorous benchmark of 198 graduate-level questions in biology, physics, and chemistry.

What is GPQA?

GPQA is the Graduate-Level Google-Proof Q&A Benchmark, a challenging dataset of multiple-choice questions written by domain experts. According to the original paper researchers under the following categories are looking for different levels of accuracy

- PhD experts in their field: 65-74% accuracy

- Baseline Highly Skilled Non-Experts with approximately 30 minutes to do a quick google search: 34% accuracy

- GPT-4 (at publication): 39% accuracy

- Random guessing: 25% accuracy

Our Results (50 questions):

ModelAccuracy
NVIDIA-Nemotron-Nano-12B-v262%
Qwen3 8B60%
NVIDIA-Nemotron-Nano-9B-v256%
Qwen3 14B 56%
Llama 3.1 8B24%
GPQA Diamond Benchmark Graph of All the Models
GPQA Diamond Benchmark Graph

Question Distribution

The 50-question sample provided balanced coverage across scientific domains:

Question Distribution Pie Chart
Question Distribution Chart
12345678910
**Organic Chemistry**: 26% (13 questions)
**Quantum Mechanics**: 20% (10 questions)
**Physics (general)**: 12% (6 questions)
**Astrophysics**: 10% (5 questions)
**High-energy particle physics**: 10% (5 questions)
**Molecular Biology**: 6% (3 questions)
**Chemistry (general)**: 6% (3 questions)
**Genetics**: 4% (2 questions)
**Relativistic Mechanics**: 4% (2 questions)
**Electromagnetism and Photonics**: 2% (1 question)

This diverse distribution ensures the models are tested across the full spectrum of graduate-level scientific reasoning.

Key Takeaways

The GPQA performance benchmarks demonstrated the Nemotron Nano family is perfect for scientific research, complex agentic applications, and more. The Nemotron-Nano-12B-v2 model achieves an accuracy of 62%, placing its scientific reasoning capabilities firmly within the range of PhD-level experts. Even the highly efficient Nemotron-Nano-9B-v2 model delivers a powerful 56% accuracy, which significantly exceeds the GPT-4 baseline (39%). Another important metric is that both Nemotron models significantly surpass the GPQA baseline search metric of 34%.

We are excited to bring these powerful, efficient models from NVIDIA to everyone.

The Mamba-Transformer hybrid architecture of Nemotron Nano 2 delivers the best of both worlds: expert-level reasoning (62% on GPQA Diamond) combined with efficient inference.

Deploy Nemotron models on Fireworks today with a single command and experience the future of efficient AI. We're excited to work with NVIDIA to make these breakthrough models available to developers worldwide.

Visit fireworks.ai to get started.

Questions? Join our Discord or contact [email protected]

Additional Resources