tiny-random / glm-4.5

huggingface.co
Total runs: 11
24-hour runs: 0
7-day runs: 1
30-day runs: 7
Model's Last Updated: July 29 2025
text-generation

Introduction of glm-4.5

Model Details of glm-4.5

This tiny model is for debugging. It is randomly initialized with the config adapted from zai-org/GLM-4.5 .

Note: The transformers implementation does not have multi-token prediction (MTP) support. So you might see some "weights not loaded" warnings. This is expected.

Example usage:
  • vLLM
model_id=tiny-random/glm-4.5
vllm serve $model_id \
    --tensor-parallel-size 1 \
    --tool-call-parser glm4_moe \
    --reasoning-parser glm4_moe \
    --enable-auto-tool-choice
  • SGLang
# Multi-token prediction is supported
model_id=tiny-random/glm-4.5
python3 -m sglang.launch_server \
    --model-path $model_id \
    --tp-size 1 \
    --cuda-graph-max-bs 4 \
    --tool-call-parser glm45  \
    --reasoning-parser glm45 \
    --speculative-algorithm EAGLE \
    --speculative-num-steps 3 \
    --speculative-eagle-topk 1 \
    --speculative-num-draft-tokens 4 \
    --mem-fraction-static 0.4
  • Transformers
from transformers import pipeline
model_id = "tiny-random/glm-4.5"
pipe = pipeline(
    "text-generation", model=model_id, device="cuda",
    trust_remote_code=True, max_new_tokens=20,
)
print(pipe("Hello World!"))
Codes to create this repo:
from copy import deepcopy

import torch
import torch.nn as nn
from transformers import (
    AutoConfig,
    AutoModelForCausalLM,
    AutoTokenizer,
    GenerationConfig,
    pipeline,
    set_seed,
)
from transformers.models.glm4_moe.modeling_glm4_moe import Glm4MoeDecoderLayer, Glm4MoeRMSNorm

source_model_id = "zai-org/GLM-4.5"
save_folder = "/tmp/tiny-random/glm-4.5"

tokenizer = AutoTokenizer.from_pretrained(
    source_model_id, trust_remote_code=True,
)
tokenizer.save_pretrained(save_folder)

config = AutoConfig.from_pretrained(
    source_model_id, trust_remote_code=True,
)
config.hidden_size = 16
config.head_dim = 64
config.intermediate_size = 64
config.num_attention_heads = 4
config.num_hidden_layers = 2  # 1 dense, 1 moe
config.num_key_value_heads = 2
config.moe_intermediate_size = 64
config.n_routed_experts = 16
config.n_shared_experts = 1
config.first_k_dense_replace = 1
config.num_experts_per_tok = 8
config.num_nextn_predict_layers = 1  # after layer 0 and 1, there will be a another MTP layer
config.tie_word_embeddings = True

torch.set_default_dtype(torch.bfloat16)
model = AutoModelForCausalLM.from_config(
    config,
    torch_dtype=torch.bfloat16,
    trust_remote_code=True,
)

class SharedHead(nn.Module):
    def __init__(self, config) -> None:
        super().__init__()
        self.norm = Glm4MoeRMSNorm(config.hidden_size, eps=config.rms_norm_eps)
        # self.head = deepcopy(model.get_output_embeddings())

class Glm4MoeDecoderMTP(Glm4MoeDecoderLayer):
    def __init__(self, config, layer_idx):
        super().__init__(config, layer_idx=layer_idx)
        self.enorm = Glm4MoeRMSNorm(config.hidden_size, eps=config.rms_norm_eps)
        self.hnorm = Glm4MoeRMSNorm(config.hidden_size, eps=config.rms_norm_eps)
        self.eh_proj = nn.Linear(config.hidden_size * 2, config.hidden_size, bias=False)
        self.shared_head = SharedHead(config=config)
        # self.embed_tokens = deepcopy(model.get_input_embeddings())

last_extra_layer = Glm4MoeDecoderMTP(config, layer_idx=config.num_hidden_layers)
model.model.layers.append(last_extra_layer)
model.generation_config = GenerationConfig.from_pretrained(
    source_model_id, trust_remote_code=True,
)
set_seed(42)
with torch.no_grad():
    for name, p in sorted(model.named_parameters()):
        torch.nn.init.normal_(p, 0, 0.2)
        print(name, p.shape)
model.save_pretrained(save_folder)

Runs of tiny-random glm-4.5 on huggingface.co

11
Total runs
0
24-hour runs
0
3-day runs
1
7-day runs
7
30-day runs

More Information About glm-4.5 huggingface.co Model

glm-4.5 huggingface.co

glm-4.5 huggingface.co is an AI model on huggingface.co that provides glm-4.5's model effect (), which can be used instantly with this tiny-random glm-4.5 model. huggingface.co supports a free trial of the glm-4.5 model, and also provides paid use of the glm-4.5. Support call glm-4.5 model through api, including Node.js, Python, http.

tiny-random glm-4.5 online free

glm-4.5 huggingface.co is an online trial and call api platform, which integrates glm-4.5's modeling effects, including api services, and provides a free online trial of glm-4.5, you can try glm-4.5 online for free by clicking the link below.

tiny-random glm-4.5 online free url in huggingface.co:

https://huggingface.co/tiny-random/glm-4.5

glm-4.5 install

glm-4.5 is an open source model from GitHub that offers a free installation service, and any user can find glm-4.5 on GitHub to install. At the same time, huggingface.co provides the effect of glm-4.5 install, users can directly use glm-4.5 installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

glm-4.5 install url in huggingface.co:

https://huggingface.co/tiny-random/glm-4.5

Url of glm-4.5

Provider of glm-4.5 huggingface.co

tiny-random
ORGANIZATIONS

Other API from tiny-random

huggingface.co

Total runs: 687
Run Growth: 670
Growth Rate: 97.53%
Updated:February 20 2026
huggingface.co

Total runs: 658
Run Growth: 604
Growth Rate: 91.79%
Updated:July 11 2025
huggingface.co

Total runs: 493
Run Growth: 474
Growth Rate: 96.15%
Updated:August 06 2025
huggingface.co

Total runs: 362
Run Growth: 73
Growth Rate: 20.17%
Updated:April 23 2026
huggingface.co

Total runs: 281
Run Growth: 278
Growth Rate: 98.93%
Updated:August 21 2025
huggingface.co

Total runs: 153
Run Growth: 80
Growth Rate: 52.29%
Updated:September 06 2025
huggingface.co

Total runs: 149
Run Growth: 64
Growth Rate: 43.24%
Updated:January 12 2025
huggingface.co

Total runs: 145
Run Growth: 57
Growth Rate: 39.31%
Updated:April 27 2025
huggingface.co

Total runs: 145
Run Growth: -26
Growth Rate: -17.93%
Updated:November 23 2025
huggingface.co

Total runs: 137
Run Growth: -99
Growth Rate: -72.26%
Updated:February 27 2026
huggingface.co

Total runs: 134
Run Growth: -128
Growth Rate: -95.52%
Updated:June 25 2025
huggingface.co

Total runs: 80
Run Growth: -10
Growth Rate: -12.50%
Updated:July 08 2025
huggingface.co

Total runs: 65
Run Growth: -270
Growth Rate: -415.38%
Updated:December 16 2025
huggingface.co

Total runs: 59
Run Growth: 13
Growth Rate: 22.03%
Updated:October 18 2025
huggingface.co

Total runs: 58
Run Growth: 58
Growth Rate: 100.00%
Updated:April 03 2026
huggingface.co

Total runs: 49
Run Growth: 44
Growth Rate: 89.80%
Updated:February 14 2026
huggingface.co

Total runs: 41
Run Growth: -30
Growth Rate: -73.17%
Updated:November 23 2025
huggingface.co

Total runs: 40
Run Growth: 33
Growth Rate: 82.50%
Updated:October 18 2025
huggingface.co

Total runs: 31
Run Growth: 31
Growth Rate: 100.00%
Updated:April 12 2026
huggingface.co

Total runs: 27
Run Growth: 15
Growth Rate: 55.56%
Updated:April 12 2026
huggingface.co

Total runs: 20
Run Growth: -14
Growth Rate: -70.00%
Updated:July 22 2025
huggingface.co

Total runs: 14
Run Growth: 9
Growth Rate: 64.29%
Updated:July 22 2025
huggingface.co

Total runs: 11
Run Growth: 2
Growth Rate: 18.18%
Updated:February 13 2026
huggingface.co

Total runs: 5
Run Growth: 1
Growth Rate: 20.00%
Updated:August 11 2025
huggingface.co

Total runs: 2
Run Growth: 1
Growth Rate: 50.00%
Updated:October 05 2025