Fireworks RFT now available! Fine-tune open models that outperform frontier models. Try today

Model Library
/Meta/Llama 4 Maverick Instruct (Basic)
Meta Mark

Llama 4 Maverick Instruct (Basic)

Ready
fireworks/llama4-maverick-instruct-basic

    The Llama 4 collection of models are natively multimodal AI models that enable text and multimodal experiences. These models leverage a mixture-of-experts architecture to offer industry-leading performance in text and image understanding.

    Llama 4 Maverick Instruct (Basic) API Features

    Fine-tuning

    Docs

    Llama 4 Maverick Instruct (Basic) can be customized with your data to improve responses. Fireworks uses LoRA to efficiently train and deploy your personalized model

    On-demand Deployment

    Docs

    On-demand deployments give you dedicated GPUs for Llama 4 Maverick Instruct (Basic) using Fireworks' reliable, high-performance system with no rate limits.

    Llama 4 Maverick Instruct FAQs

    What is Llama 4 Maverick Instruct and who developed it?

    Llama 4 Maverick Instruct is the instruction-tuned variant of Llama 4 Maverick, a 17 billion-parameter mixture-of-experts (128 experts, 400B total parameters) multimodal model created by Meta.

    What applications and use cases does Llama 4 Maverick Instruct excel at?

    Llama 4 Maverick Instruct is designed for:

    • Assistant-style chat
    • Visual reasoning (image captioning, multi-image QA)
    • Multilingual generation
    • Synthetic-data/distillation workflows
    What is the maximum context length for Llama 4 Maverick Instruct?

    Llama 4 Maverick Instruct has a maximum context length of 1,048,576 tokens.

    Does Llama 4 Maverick Instruct support quantized formats (4-bit/8-bit)?

    Yes. Meta releases Maverick in BF16 and FP8 checkpoints and provides on-the-fly int4 (≈ 4-bit) quantization code; this enables single-GPU deployment.

    How many parameters does Llama 4 Maverick Instruct have?

    Llama 4 Maverick Instruct has 17 billion active parameters (one expert per token) within a 128-expert MoE totaling 400 billion parameters.

    Is fine-tuning supported for Llama 4 Maverick Instruct?

    Yes. Fine-tuning is available for Llama 4 Maverick Instruct on Fireworks via LoRA.

    What license governs commercial use of Llama 4 Maverick Instruct?

    The model is released under the Llama 4 Community License Agreement (custom commercial license by Meta).

    Metadata

    State
    Ready
    Created on
    4/5/2025
    Kind
    Base model
    Provider
    Meta
    Hugging Face
    Llama-4-Maverick-17B-128E-Instruct

    Specification

    Calibrated
    Yes
    Mixture-of-Experts
    Yes
    Parameters
    401.6B

    Supported Functionality

    Fine-tuning
    Supported
    Serverless
    Not supported
    Serverless LoRA
    Not supported
    Context Length
    1048.6k tokens
    Function Calling
    Supported
    Embeddings
    Not supported
    Rerankers
    Not supported
    Support image input
    Supported