Fireworks RFT now available! Fine-tune open models that outperform frontier models. Try today

Model Library
/Z.ai/GLM-4.5
fireworks/glm-4p5

    The GLM-4.5 series models are foundation models designed for intelligent agents. GLM-4.5 has 355 billion total parameters with 32 billion active parameters, while GLM-4.5-Air adopts a more compact design with 106 billion total parameters and 12 billion active parameters. GLM-4.5 models unify reasoning, coding, and intelligent agent capabilities to meet the complex demands of intelligent agent applications.

    GLM-4.5 API Features

    Fine-tuning

    Docs

    GLM-4.5 can be customized with your data to improve responses. Fireworks uses LoRA to efficiently train and deploy your personalized model

    Serverless

    Docs

    Immediately run model on pre-configured GPUs and pay-per-token

    On-demand Deployment

    Docs

    On-demand deployments give you dedicated GPUs for GLM-4.5 using Fireworks' reliable, high-performance system with no rate limits.

    Available Serverless

    Run queries immediately, pay only for usage

    $0.55 / $2.19
    Per 1M Tokens (input/output)

    GLM-4.5 FAQs

    What is GLM-4.5 and who developed it?

    GLM-4.5 is a hybrid MoE large language model developed by Zhipu AI. It is designed for reasoning, coding, and agentic tasks. Fireworks AI offers serverless and on-demand inference for the model.

    What applications and use cases does GLM-4.5 excel at?

    GLM-4.5 is optimized for:

    • Agentic applications (e.g., function calling, web browsing)
    • Advanced reasoning (math, logic, science)
    • Coding tasks (project creation, bug fixing, agentic coding)
    • Full-stack development and presentation generation with tool integration.
    What is the maximum context length for GLM-4.5?

    The maximum context length is 131,072 tokens.

    What is the usable context window for GLM-4.5?

    The full 128K context length is usable, given the appropriate hardware configuration (e.g., H100x32 or H200x16).

    Does GLM-4.5 support quantized formats (4-bit/8-bit)?

    Yes. GLM-4.5 is available in FP8 versions, suitable for inference-optimized deployments.

    What are known failure modes of GLM-4.5?

    While not explicitly known failure modes, the model shows lower performance on:

    • HLE (14.4%) - reasoning under human-like evaluation
    • BrowseComp (26.4%) - complex web search tasks
    Does GLM-4.5 support streaming responses and function-calling schemas?

    Yes. GLM-4.5 supports native function calling and OpenAI-style tool schemas. It also supports speculative decoding and streaming via SGLang and vLLM.

    How many parameters does GLM-4.5 have?

    GLM-4.5 has 355B total parameters, with 32B active parameters per forward pass.

    What license governs commercial use of GLM-4.5?

    GLM-4.5 is released under the MIT license, allowing commercial and derivative use.

    Metadata

    State
    Ready
    Created on
    7/29/2025
    Kind
    Base model
    Provider
    Z.ai
    Hugging Face
    GLM-4.5

    Specification

    Calibrated
    Yes
    Mixture-of-Experts
    Yes
    Parameters
    352.8B

    Supported Functionality

    Fine-tuning
    Supported
    Serverless
    Supported
    Serverless LoRA
    Not supported
    Context Length
    131.1k tokens
    Function Calling
    Supported
    Embeddings
    Not supported
    Rerankers
    Not supported
    Support image input
    Not supported