PrimeIntellect / INTELLECT-3-FP8

huggingface.co
Total runs: 183
24-hour runs: 0
7-day runs: -64
30-day runs: -359
Model's Last Updated: November 27 2025
text-generation

Introduction of INTELLECT-3-FP8

Model Details of INTELLECT-3-FP8

INTELLECT-3

Prime Intellect Logo

INTELLECT-3: A 100B+ MoE trained with large-scale RL

Trained with prime-rl and verifiers
Environments released on Environments Hub
Read the Blog & Technical Report
X | Discord | Prime Intellect Platform

Introduction

INTELLECT-3 is a 106B (A12B) parameter Mixture-of-Experts reasoning model post-trained from GLM-4.5-Air-Base using supervised fine-tuning (SFT) followed by large-scale reinforcement learning (RL).

bench

Training was performed with prime-rl using environments built with the verifiers library. All training and evaluation environments are available on the Environments Hub .

The model, training frameworks, and environments are open-sourced under fully-permissive licenses (MIT and Apache 2.0).

For more details, see the technical report .

Evaluation

INTELLECT-3 achieves best-in-class performance on math, coding, and reasoning benchmarks:

Benchmark MATH-500 AIME24 AIME25 LCB GPQA HLE MMLU-Pro
INTELLECT-3 98.1 90.8 88.0 69.3 74.4 14.6 81.9
GLM-4.5-Air 97.8 84.6 82.0 61.5 73.3 13.3 73.9
GLM-4.5 97.0 85.8 83.3 64.5 77.0 14.8 83.5
DeepSeek R1 0528 87.3 83.2 73.4 62.5 77.5 15.9 75.3
DeepSeek v3.2 96.8 88.1 84.7 71.6 81.4 17.9 84.6
GPT-O5S 120B 96.0 75.8 77.7 69.9 70.0 10.6 67.1
Model Variants
Model HuggingFace
INTELLECT-3 PrimeIntellect/INTELLECT-3
INTELLECT-3-FP8 PrimeIntellect/INTELLECT-3-FP8
Serving with vLLM

The BF16 version can be served on 2x H200s:

vllm serve PrimeIntellect/INTELLECT-3 \
    --tensor-parallel-size 2 \
    --enable-auto-tool-choice \
    --tool-call-parser qwen3_coder \
    --reasoning-parser deepseek_r1

The FP8 version can be served on a single H200:

vllm serve PrimeIntellect/INTELLECT-3-FP8 \
    --enable-auto-tool-choice \
    --tool-call-parser qwen3_coder \
    --reasoning-parser deepseek_r1
Citation
@misc{intellect3,
  title={INTELLECT-3: Technical Report},
  author={Prime Intellect Team},
  year={2025},
  url={https://huggingface.co/PrimeIntellect/INTELLECT-3}
}

Runs of PrimeIntellect INTELLECT-3-FP8 on huggingface.co

183
Total runs
0
24-hour runs
0
3-day runs
-64
7-day runs
-359
30-day runs

More Information About INTELLECT-3-FP8 huggingface.co Model

More INTELLECT-3-FP8 license Visit here:

https://choosealicense.com/licenses/mit

INTELLECT-3-FP8 huggingface.co

INTELLECT-3-FP8 huggingface.co is an AI model on huggingface.co that provides INTELLECT-3-FP8's model effect (), which can be used instantly with this PrimeIntellect INTELLECT-3-FP8 model. huggingface.co supports a free trial of the INTELLECT-3-FP8 model, and also provides paid use of the INTELLECT-3-FP8. Support call INTELLECT-3-FP8 model through api, including Node.js, Python, http.

PrimeIntellect INTELLECT-3-FP8 online free

INTELLECT-3-FP8 huggingface.co is an online trial and call api platform, which integrates INTELLECT-3-FP8's modeling effects, including api services, and provides a free online trial of INTELLECT-3-FP8, you can try INTELLECT-3-FP8 online for free by clicking the link below.

PrimeIntellect INTELLECT-3-FP8 online free url in huggingface.co:

https://huggingface.co/PrimeIntellect/INTELLECT-3-FP8

INTELLECT-3-FP8 install

INTELLECT-3-FP8 is an open source model from GitHub that offers a free installation service, and any user can find INTELLECT-3-FP8 on GitHub to install. At the same time, huggingface.co provides the effect of INTELLECT-3-FP8 install, users can directly use INTELLECT-3-FP8 installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

INTELLECT-3-FP8 install url in huggingface.co:

https://huggingface.co/PrimeIntellect/INTELLECT-3-FP8

Url of INTELLECT-3-FP8

Provider of INTELLECT-3-FP8 huggingface.co

PrimeIntellect
ORGANIZATIONS

Other API from PrimeIntellect