arcee-ai / Trinity-Large-TrueBase

huggingface.co
Total runs: 503
24-hour runs: 0
7-day runs: 57
30-day runs: 382
Model's Last Updated: April 01 2026
text-generation

Introduction of Trinity-Large-TrueBase

Model Details of Trinity-Large-TrueBase

Arcee Trinity Large

Trinity-Large-TrueBase

Introduction

Trinity-Large-TrueBase is a base pretraining checkpoint from Arcee AI's Trinity Large training run. It is a 398B-parameter sparse Mixture-of-Experts (MoE) model with approximately 13B active parameters per token. The checkpoint was captured after 10 trillion tokens of pretraining, prior to learning-rate annealing and before any instruction tuning or reinforcement learning.

This checkpoint is intended for research, probing, ablation studies, and downstream fine-tuning and comes without any pre-baked alignment, instruction formatting, or preference optimization.

More details on the training of Trinity Large are available in the technical report .

Model Variants

The Trinity Large family consists of three checkpoints from the same training run:

  • Trinity-Large-TrueBase (this release): 10T-token pre-anneal checkpoint with no instruction data
  • Trinity-Large-Base : Full 17T-token pretrained foundation model with mid-training anneals
  • Trinity-Large-Preview : Lightly post-trained, chat-ready model undergoing active RL
Architecture

Trinity-Large-TrueBase uses a sparse MoE configuration designed to maximize efficiency while maintaining large-scale capacity.

Hyperparameter Value
Total parameters ~398B
Active parameters per token ~13B
Experts 256
Active experts 4
Routing strategy 4-of-256 (1.56% sparsity)
Dense layers 6
Pretraining context length 8,192
Architecture Sparse MoE (AfmoeForCausalLM)

Note: Extended context support (e.g., 512k) was introduced after this checkpoint and is not available in TrueBase.

Benchmark Results
Benchmark N-shot Metric Score Stderr
arc_challenge_0shot 0 acc_norm,none 0.6237 ±0.0142
bbh_fewshot 3 exact_match,remove_whitespace 0.5784 ±0.0054
gpqa_diamond_5shot 5 acc_norm,none 0.4091 ±0.0350
gpqa_diamond_generative_5shot 5 exact_match,flexible-extract 0.3788 ±0.0346
gsm8k_8shot 8 exact_match,flexible-extract 0.8036 ±0.0109
gsm8k_cot 8 exact_match,flexible-extract 0.8044 ±0.0109
hellaswag_5shot 5 acc_norm,none 0.8813 ±0.0032
humaneval_plus 0 pass@1,create_test 0.5183 ±0.0391
leaderboard_math_hard 4 exact_match,none 0.2696 ±0.0113
mbpp_plus 3 pass_at_1,none 0.8095 ±0.0202
minerva_math500 4 math_verify,none 0.4820 ±0.0224
mmlu_5shot 5 acc,none 0.7845 ±0.0033
mmlu_generative_5shot 5 exact_match,get_response 0.7848 ±0.0033
mmlu_pro 5 exact_match,custom-extract 0.5160 ±0.0044
triviaqa_5shot 5 exact_match,remove_whitespace 0.8096 ±0.0029
winogrande_5shot 5 acc,none 0.8145 ±0.0109
Training Configuration
Pretraining
  • Training tokens: 10 trillion
  • Checkpoint type: Pre-anneal
  • Instruction data: None
  • RLHF or post-training: None

This checkpoint branches from the main Trinity Large run at the 10T-token mark, prior to learning-rate decay or post-training phases.

Optimizers

Optimizer learning rates after WSD warm-up:

  • Adam learning rate: 2e-4
  • Muon learning rate: 8e-4

Muon was used to support larger critical batch sizes in a highly sparse MoE regime.

Infrastructure
Powered by Datology
Powered by Datology
Intended Use
  • Studying emergent behavior from large-scale pretraining
  • Sparse MoE routing and load-balancing research
  • Interpretability, probing, and ablation studies
  • Domain-specific fine-tuning from a clean base
  • Academic and industrial foundation model research
Rationale for Release

Most base model releases include instruction data, annealed training dynamics, or early alignment stages. Trinity-Large-TrueBase excludes these, providing an opportunity to study what large-scale models learn from pretraining data alone. This checkpoint is intended as a foundation for research rather than as a finished conversational assistant.

Known Limitations
  • Not aligned for safety, helpfulness, or conversational tone
  • Requires substantial compute and expertise to fine-tune
  • May exhibit raw or unstable behaviors typical of unaligned models
  • No extended-context tuning beyond the 8K pretraining window
License

Trinity-Large-TrueBase is released under the Apache License, Version 2.0.

Runs of arcee-ai Trinity-Large-TrueBase on huggingface.co

503
Total runs
0
24-hour runs
3
3-day runs
57
7-day runs
382
30-day runs

More Information About Trinity-Large-TrueBase huggingface.co Model

More Trinity-Large-TrueBase license Visit here:

https://choosealicense.com/licenses/apache-2.0

Trinity-Large-TrueBase huggingface.co

Trinity-Large-TrueBase huggingface.co is an AI model on huggingface.co that provides Trinity-Large-TrueBase's model effect (), which can be used instantly with this arcee-ai Trinity-Large-TrueBase model. huggingface.co supports a free trial of the Trinity-Large-TrueBase model, and also provides paid use of the Trinity-Large-TrueBase. Support call Trinity-Large-TrueBase model through api, including Node.js, Python, http.

Trinity-Large-TrueBase huggingface.co Url

https://huggingface.co/arcee-ai/Trinity-Large-TrueBase

arcee-ai Trinity-Large-TrueBase online free

Trinity-Large-TrueBase huggingface.co is an online trial and call api platform, which integrates Trinity-Large-TrueBase's modeling effects, including api services, and provides a free online trial of Trinity-Large-TrueBase, you can try Trinity-Large-TrueBase online for free by clicking the link below.

arcee-ai Trinity-Large-TrueBase online free url in huggingface.co:

https://huggingface.co/arcee-ai/Trinity-Large-TrueBase

Trinity-Large-TrueBase install

Trinity-Large-TrueBase is an open source model from GitHub that offers a free installation service, and any user can find Trinity-Large-TrueBase on GitHub to install. At the same time, huggingface.co provides the effect of Trinity-Large-TrueBase install, users can directly use Trinity-Large-TrueBase installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

Trinity-Large-TrueBase install url in huggingface.co:

https://huggingface.co/arcee-ai/Trinity-Large-TrueBase

Url of Trinity-Large-TrueBase

Trinity-Large-TrueBase huggingface.co Url

Provider of Trinity-Large-TrueBase huggingface.co

arcee-ai
ORGANIZATIONS

Other API from arcee-ai

huggingface.co

Total runs: 19.9K
Run Growth: 3.9K
Growth Rate: 19.45%
Updated:September 18 2025
huggingface.co

Total runs: 7.4K
Run Growth: -348
Growth Rate: -4.72%
Updated:December 12 2025
huggingface.co

Total runs: 1.9K
Run Growth: 1.7K
Growth Rate: 89.06%
Updated:October 10 2025
huggingface.co

Total runs: 1.7K
Run Growth: 750
Growth Rate: 42.47%
Updated:September 18 2025
huggingface.co

Total runs: 788
Run Growth: 714
Growth Rate: 90.72%
Updated:January 16 2026
huggingface.co

Total runs: 783
Run Growth: 374
Growth Rate: 48.51%
Updated:July 22 2024
huggingface.co

Total runs: 656
Run Growth: 631
Growth Rate: 96.19%
Updated:September 18 2024
huggingface.co

Total runs: 181
Run Growth: 135
Growth Rate: 74.59%
Updated:June 11 2025
huggingface.co

Total runs: 173
Run Growth: 118
Growth Rate: 67.05%
Updated:September 10 2024
huggingface.co

Total runs: 133
Run Growth: 45
Growth Rate: 33.83%
Updated:July 19 2024
huggingface.co

Total runs: 82
Run Growth: 26
Growth Rate: 61.90%
Updated:August 01 2024
huggingface.co

Total runs: 40
Run Growth: 2
Growth Rate: 5.13%
Updated:February 27 2025
huggingface.co

Total runs: 28
Run Growth: 16
Growth Rate: 57.14%
Updated:June 03 2025
huggingface.co

Total runs: 25
Run Growth: 0
Growth Rate: 0.00%
Updated:June 11 2025