zai-org / GLM-5

huggingface.co
Total runs: 442.4K
24-hour runs: -3.7K
7-day runs: -35.3K
30-day runs: 230.6K
Model's Last Updated: April 05 2026
text-generation

Introduction of GLM-5

Model Details of GLM-5

GLM-5

👋 Join our WeChat or Discord community.
📖 Check out the GLM-5 technical blog .
📍 Use GLM-5 API services on Z.ai API Platform.
👉 One click to GLM-5 .

Introduction

We are launching GLM-5, targeting complex systems engineering and long-horizon agentic tasks. Scaling is still one of the most important ways to improve the intelligence efficiency of Artificial General Intelligence (AGI). Compared to GLM-4.5, GLM-5 scales from 355B parameters (32B active) to 744B parameters (40B active), and increases pre-training data from 23T to 28.5T tokens. GLM-5 also integrates DeepSeek Sparse Attention (DSA), largely reducing deployment cost while preserving long-context capacity.

Reinforcement learning aims to bridge the gap between competence and excellence in pre-trained models. However, deploying it at scale for LLMs is a challenge due to the RL training inefficiency. To this end, we developed slime , a novel asynchronous RL infrastructure that substantially improves training throughput and efficiency, enabling more fine-grained post-training iterations. With advances in both pre-training and post-training, GLM-5 delivers significant improvement compared to GLM-4.7 across a wide range of academic benchmarks and achieves best-in-class performance among all open-source models in the world on reasoning, coding, and agentic tasks, closing the gap with frontier models.

Benchmark
GLM-5 GLM-4.7 DeepSeek-V3.2 Kimi K2.5 Claude Opus 4.5 Gemini 3 Pro GPT-5.2 (xhigh)
HLE 30.5 24.8 25.1 31.5 28.4 37.2 35.4
HLE (w/ Tools) 50.4 42.8 40.8 51.8 43.4* 45.8* 45.5*
AIME 2026 I 92.7 92.9 92.7 92.5 93.3 90.6 -
HMMT Nov. 2025 96.9 93.5 90.2 91.1 91.7 93.0 97.1
IMOAnswerBench 82.5 82.0 78.3 81.8 78.5 83.3 86.3
GPQA-Diamond 86.0 85.7 82.4 87.6 87.0 91.9 92.4
SWE-bench Verified 77.8 73.8 73.1 76.8 80.9 76.2 80.0
SWE-bench Multilingual 73.3 66.7 70.2 73.0 77.5 65.0 72.0
Terminal-Bench 2.0 (Terminus 2) 56.2 / 60.7 † 41.0 39.3 50.8 59.3 54.2 54.0
Terminal-Bench 2.0 (Claude Code) 56.2 / 61.1 † 32.8 46.4 - 57.9 - -
CyberGym 43.2 23.5 17.3 41.3 50.6 39.9 -
BrowseComp 62.0 52.0 51.4 60.6 37.0 37.8 -
BrowseComp (w/ Context Manage) 75.9 67.5 67.6 74.9 67.8 59.2 65.8
BrowseComp-Zh 72.7 66.6 65.0 62.3 62.4 66.8 76.1
τ²-Bench 89.7 87.4 85.3 80.2 91.6 90.7 85.5
MCP-Atlas (Public Set) 67.8 52.0 62.2 63.8 65.2 66.6 68.0
Tool-Decathlon 38.0 23.8 35.2 27.8 43.5 36.4 46.3
Vending Bench 2 $4,432.12 $2,376.82 $1,034.00 $1,198.46 $4,967.06 $5,478.16 $3,591.33

*: refers to their scores of full set.

†: A verified version of Terminal-Bench 2.0 that fixes some ambiguous instructions. See footnote for more evaluation details.

Footnote
  • Humanity’s Last Exam (HLE) & other reasoning tasks : We evaluate with a maximum generation length of 131,072 tokens ( temperature=1.0, top_p=0.95, max_new_tokens=131072 ). By default, we report the text-only subset; results marked with * are from the full set. We use GPT-5.2 (medium) as the judge model. For HLE-with-tools, we use a maximum context length of 202,752 tokens.
  • SWE-bench & SWE-bench Multilingual : We run the SWE-bench suite with OpenHands using a tailored instruction prompt. Settings: temperature=0.7, top_p=0.95, max_new_tokens=16384 , with a 200K context window.
  • BrowserComp : Without context management, we retain details from the most recent 5 turns. With context management, we use the same discard-all strategy as DeepSeek-v3.2 and Kimi K2.5.
  • Terminal-Bench 2.0 (Terminus 2) : We evaluate with the Terminus framework using timeout=2h, temperature=0.7, top_p=1.0, max_new_tokens=8192 , with a 128K context window. Resource limits are capped at 16 CPUs and 32 GB RAM.
  • Terminal-Bench 2.0 (Claude Code) : We evaluate in Claude Code 2.1.14 (think mode, default effort) with temperature=1.0, top_p=0.95, max_new_tokens=65536 . We remove wall-clock time limits due to generation speed, while preserving per-task CPU and memory constraints. Scores are averaged over 5 runs. We fix environment issues introduced by Claude Code and also report results on a verified Terminal-Bench 2.0 dataset that resolves ambiguous instructions (see: https://huggingface.co/datasets/zai-org/terminal-bench-2-verified ).
  • CyberGym : We evaluate in Claude Code 2.1.18 (think mode, no web tools) with ( temperature=1.0, top_p=1.0, max_new_tokens=32000 ) and a 250-minute timeout per task. Results are single-run Pass@1 over 1,507 tasks.
  • MCP-Atlas : All models are evaluated in think mode on the 500-task public subset with a 10-minute timeout per task. We use Gemini 3 Pro as the judge model.
  • τ²-bench : We add a small prompt adjustment in Retail and Telecom to avoid failures caused by premature user termination. For Airline, we apply the domain fixes proposed in the Claude Opus 4.5 system card.
  • Vending Bench 2 : Runs are conducted independently by Andon Labs .
Serve GLM-5 Locally
Prepare environment

vLLM, SGLang, and xLLM all support local deployment of GLM-5. A simple deployment guide is provided here.

  • vLLM

    Using Docker as:

    docker pull vllm/vllm-openai:nightly 
    

    or using pip:

    pip install -U vllm --pre --index-url https://pypi.org/simple --extra-index-url https://wheels.vllm.ai/nightly
    

    then upgrade transformers:

    pip install git+https://github.com/huggingface/transformers.git
    
  • SGLang

    Using Docker as:

    docker pull lmsysorg/sglang:glm5-hopper # For Hopper GPU
    docker pull lmsysorg/sglang:glm5-blackwell # For Blackwell GPU
    
Deploy
  • vLLM

    vllm serve zai-org/GLM-5-FP8 \
         --tensor-parallel-size 8 \
         --gpu-memory-utilization 0.85 \
         --speculative-config.method mtp \
         --speculative-config.num_speculative_tokens 1 \
         --tool-call-parser glm47 \
         --reasoning-parser glm45 \
         --enable-auto-tool-choice \
         --served-model-name glm-5-fp8
    

    Check the recipes for more details.

  • SGLang

    python3 -m sglang.launch_server \
      --model-path zai-org/GLM-5-FP8 \
      --tp-size 8 \
      --tool-call-parser glm47  \
      --reasoning-parser glm45 \
      --speculative-algorithm EAGLE \
      --speculative-num-steps 3 \
      --speculative-eagle-topk 1 \
      --speculative-num-draft-tokens 4 \
      --mem-fraction-static 0.85 \
      --served-model-name glm-5-fp8
    

    Check the sglang cookbook for more details.

  • xLLM and other Ascend NPU

    Please check the deployment guide here .

Citation

Our technical report is coming soon.

Runs of zai-org GLM-5 on huggingface.co

442.4K
Total runs
-3.7K
24-hour runs
-13.3K
3-day runs
-35.3K
7-day runs
230.6K
30-day runs

More Information About GLM-5 huggingface.co Model

More GLM-5 license Visit here:

https://choosealicense.com/licenses/mit

GLM-5 huggingface.co

GLM-5 huggingface.co is an AI model on huggingface.co that provides GLM-5's model effect (), which can be used instantly with this zai-org GLM-5 model. huggingface.co supports a free trial of the GLM-5 model, and also provides paid use of the GLM-5. Support call GLM-5 model through api, including Node.js, Python, http.

zai-org GLM-5 online free

GLM-5 huggingface.co is an online trial and call api platform, which integrates GLM-5's modeling effects, including api services, and provides a free online trial of GLM-5, you can try GLM-5 online for free by clicking the link below.

zai-org GLM-5 online free url in huggingface.co:

https://huggingface.co/zai-org/GLM-5

GLM-5 install

GLM-5 is an open source model from GitHub that offers a free installation service, and any user can find GLM-5 on GitHub to install. At the same time, huggingface.co provides the effect of GLM-5 install, users can directly use GLM-5 installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

GLM-5 install url in huggingface.co:

https://huggingface.co/zai-org/GLM-5

Url of GLM-5

Provider of GLM-5 huggingface.co

zai-org
ORGANIZATIONS

Other API from zai-org

huggingface.co

Total runs: 8.0M
Run Growth: 4.2M
Growth Rate: 52.16%
Updated:April 14 2026
huggingface.co

Total runs: 1.3M
Run Growth: -2.9M
Growth Rate: -228.14%
Updated:April 05 2026
huggingface.co

Total runs: 662.4K
Run Growth: -777.3K
Growth Rate: -117.33%
Updated:January 29 2026
huggingface.co

Total runs: 606.9K
Run Growth: 600.0K
Growth Rate: 98.87%
Updated:April 16 2026
huggingface.co

Total runs: 432.9K
Run Growth: 284.7K
Growth Rate: 65.76%
Updated:March 13 2025
huggingface.co

Total runs: 416.3K
Run Growth: -30.2K
Growth Rate: -7.25%
Updated:August 04 2024
huggingface.co

Total runs: 394.5K
Run Growth: -3.5K
Growth Rate: -0.89%
Updated:August 11 2025
huggingface.co

Total runs: 237.8K
Run Growth: 172.7K
Growth Rate: 72.59%
Updated:December 23 2025
huggingface.co

Total runs: 237.4K
Run Growth: 236.2K
Growth Rate: 99.45%
Updated:April 16 2026
huggingface.co

Total runs: 89.1K
Run Growth: -57.8K
Growth Rate: -64.94%
Updated:January 29 2026
huggingface.co

Total runs: 82.8K
Run Growth: 17.2K
Growth Rate: 20.76%
Updated:August 11 2025
huggingface.co

Total runs: 76.0K
Run Growth: 18.0K
Growth Rate: 23.70%
Updated:December 05 2024
huggingface.co

Total runs: 67.5K
Run Growth: 6.1K
Growth Rate: 9.11%
Updated:March 03 2025
huggingface.co

Total runs: 58.8K
Run Growth: 12.0K
Growth Rate: 20.49%
Updated:October 25 2025
huggingface.co

Total runs: 35.9K
Run Growth: 5.9K
Growth Rate: 16.28%
Updated:November 23 2024
huggingface.co

Total runs: 33.2K
Run Growth: 10.6K
Growth Rate: 32.05%
Updated:September 30 2025
huggingface.co

Total runs: 27.5K
Run Growth: 10.1K
Growth Rate: 36.85%
Updated:November 23 2024
huggingface.co

Total runs: 20.3K
Run Growth: 2.7K
Growth Rate: 13.35%
Updated:April 14 2025
huggingface.co

Total runs: 18.6K
Run Growth: 7.9K
Growth Rate: 42.28%
Updated:October 16 2025
huggingface.co

Total runs: 16.1K
Run Growth: -10.3K
Growth Rate: -64.29%
Updated:October 25 2024
huggingface.co

Total runs: 10.2K
Run Growth: 1.5K
Growth Rate: 15.01%
Updated:April 22 2025
huggingface.co

Total runs: 7.4K
Run Growth: 3.7K
Growth Rate: 49.76%
Updated:August 04 2024
huggingface.co

Total runs: 6.5K
Run Growth: 2.3K
Growth Rate: 35.75%
Updated:December 09 2025
huggingface.co

Total runs: 6.0K
Run Growth: -121.2K
Growth Rate: -2006.59%
Updated:December 09 2025
huggingface.co

Total runs: 3.7K
Run Growth: -2.5K
Growth Rate: -66.51%
Updated:January 15 2026
huggingface.co

Total runs: 3.1K
Run Growth: 722
Growth Rate: 23.06%
Updated:March 11 2025
huggingface.co

Total runs: 2.2K
Run Growth: 875
Growth Rate: 39.13%
Updated:January 27 2025
huggingface.co

Total runs: 2.1K
Run Growth: -7.3K
Growth Rate: -344.64%
Updated:October 25 2025
huggingface.co

Total runs: 2.0K
Run Growth: 909
Growth Rate: 46.10%
Updated:January 13 2026
huggingface.co

Total runs: 1.5K
Run Growth: -1.3K
Growth Rate: -91.49%
Updated:October 25 2025
huggingface.co

Total runs: 1.0K
Run Growth: 32
Growth Rate: 3.05%
Updated:August 12 2025
huggingface.co

Total runs: 1.0K
Run Growth: -4.6K
Growth Rate: -454.22%
Updated:January 02 2025
huggingface.co

Total runs: 873
Run Growth: 842
Growth Rate: 96.45%
Updated:November 20 2023
huggingface.co

Total runs: 392
Run Growth: 76
Growth Rate: 19.39%
Updated:October 20 2023
huggingface.co

Total runs: 300
Run Growth: 244
Growth Rate: 81.33%
Updated:August 11 2025
huggingface.co

Total runs: 267
Run Growth: -15
Growth Rate: -5.62%
Updated:March 02 2023
huggingface.co

Total runs: 194
Run Growth: -22
Growth Rate: -11.34%
Updated:August 04 2024
huggingface.co

Total runs: 184
Run Growth: -130
Growth Rate: -70.65%
Updated:December 10 2024
huggingface.co

Total runs: 169
Run Growth: 26
Growth Rate: 15.38%
Updated:October 20 2023
huggingface.co

Total runs: 114
Run Growth: -38
Growth Rate: -33.33%
Updated:November 12 2025
huggingface.co

Total runs: 79
Run Growth: 1
Growth Rate: 1.27%
Updated:July 09 2025
huggingface.co

Total runs: 65
Run Growth: 0
Growth Rate: 0.00%
Updated:July 09 2025