ReactiveAI / RxT-Alpha-Micro-Decoder-Plus

huggingface.co
Total runs: 0
24-hour runs: 0
7-day runs: 0
30-day runs: 0
Model's Last Updated: May 29 2025
text-generation

Introduction of RxT-Alpha-Micro-Decoder-Plus

Model Details of RxT-Alpha-Micro-Decoder-Plus

RxT-Alpha Micro Decoder Plus (Base)

Reactive Transformer Architecture

Experimental research model made to test our Reactive Transformer architecture and Attention-based Memory System.

Reactive Transformer has additional Short-Term Memory layers, connected to model with Memory Cross-Attention, and updated by Memory Encoder and Memory Attention. Short-Term Memory state is kept between interactions/event (single message), not between tokens in sequence - that's key difference between RxNNs and RNNs.

The goal of the architecture is to process only single messages and keep conversation history in Short-Term Memory - we believe, that this is the key requirement for awareness and AGI. Processing all the chat history on every interaction is not natural and that's not how human awareness is working. Then, Reactive Transformer architecture is a first step in transition from language models to awareness models.

This model (decoder) is a generator decoder for Reactive Transformer system and is made for first stage of training - base model pre-training.

Decoder is based on Mixture-of-Experts architecture with 12 experts and 4 active ones (for Plus version, basic has 2 active ones and little worse results).

During first stage, Memory Cross-Attention layers are frozen and STM is in default initial random state (normal distribution with 0 mean and almost 0 variance), to not disturb basic language modelling training. We are training decoder and encoder separately with shared embeddings. Then, in second stage - Memory Reinforcement Learning, they will be connected into bigger ensemble with additional Memory Norm and Memory Attention layers, and will learn how to keep and update memory.

RxT-Alpha models intentionally use very short sequence length and STM size (256 tokens for Micro), but that isn't their "full" context size - it's only for single message. "Full" context is theoretically infinite, restricted by STM size and memory abilites. That sizes are good for research, final models will handle SOTA contexts.

RxT-Alpha Micro Training

Micro models from RxT-Alpha series are first PoC for Reactive Transformer, Attention-Based Memory System and Memory Reinforcement Learning, used mainly to test library and architecture basics, before training bigger models (that are still relatively small, as it's PoC).

Decoder was trained on Autoregressive Language Modelling task with embedding from encoder pre-training , with roneneldan/TinyStories dataset, using 4B total tokens and reached ~71.4% accuracy .

Next Stage: Memory Reinforcement Learning

The model is able to generate meaningful short stories, using grammatically correct sentences, and is ready for the memory training in the next stage. More info soon.

Decoder architecture details:
  • dim: 128
  • layers: 6
  • heads: 8
  • self-attention: symmetric Sparse Query Attention
    • query/key/value groups: 4
  • memory cross-attention: Sparse Query Attention
    • query groups: 4
    • key/value groups: 2
  • Mixture-of-Experts Feed Forward
    • experts: 12
    • active experts: 4 (vs 2 in base version )
    • SwiGLU feed forward with 256 dim
  • RoPE
  • RMS Norm
  • vocab: 5k (english only)
  • message length: 256
  • STM size: 256 * 6 layers
  • size: ~8.77M
  • Library: RxNN
  • Docs: draft/in progress
Usage

Model requires RxNN framework for training/inference. It's integrated with HuggingFace Hub and libraries.

Inference:
  • Install RxNN, PyTorch and dependencies: pip install rxnn torch transformers tokenizers
  • Install Flash Attention (optional, but recommended) - details in RxNN framework docs
import torch
from rxnn.rxt.models import RxTAlphaDecoder
from rxnn.transformers.sampler import Sampler, SampleDecoder
from rxnn.training.tokenizer import load_tokenizer_from_hf_hub

model = RxTAlphaDecoder.from_pretrained('ReactiveAI/RxT-Alpha-Micro-Decoder-Plus')
tokenizer = load_tokenizer_from_hf_hub('ReactiveAI/RxT-Alpha-Micro-Decoder-Plus')
sampler = Sampler(model, torch.device('cuda' if torch.cuda.is_available() else 'cpu'), end_token_id=3)
sample = SampleDecoder(sampler, tokenizer)

# 0.1 and 0.9 are default values for temperature and top_p
generated = sample('Example model input for text generation...', temperature=0.1, top_p=0.9, max_seq_len=256)
sample('Example model input for text generation - print streamed response...', temperature=0.1, top_p=0.9, max_seq_len=256, print_stream=True)

Runs of ReactiveAI RxT-Alpha-Micro-Decoder-Plus on huggingface.co

0
Total runs
0
24-hour runs
0
3-day runs
0
7-day runs
0
30-day runs

More Information About RxT-Alpha-Micro-Decoder-Plus huggingface.co Model

More RxT-Alpha-Micro-Decoder-Plus license Visit here:

https://choosealicense.com/licenses/apache-2.0

RxT-Alpha-Micro-Decoder-Plus huggingface.co

RxT-Alpha-Micro-Decoder-Plus huggingface.co is an AI model on huggingface.co that provides RxT-Alpha-Micro-Decoder-Plus's model effect (), which can be used instantly with this ReactiveAI RxT-Alpha-Micro-Decoder-Plus model. huggingface.co supports a free trial of the RxT-Alpha-Micro-Decoder-Plus model, and also provides paid use of the RxT-Alpha-Micro-Decoder-Plus. Support call RxT-Alpha-Micro-Decoder-Plus model through api, including Node.js, Python, http.

RxT-Alpha-Micro-Decoder-Plus huggingface.co Url

https://huggingface.co/ReactiveAI/RxT-Alpha-Micro-Decoder-Plus

ReactiveAI RxT-Alpha-Micro-Decoder-Plus online free

RxT-Alpha-Micro-Decoder-Plus huggingface.co is an online trial and call api platform, which integrates RxT-Alpha-Micro-Decoder-Plus's modeling effects, including api services, and provides a free online trial of RxT-Alpha-Micro-Decoder-Plus, you can try RxT-Alpha-Micro-Decoder-Plus online for free by clicking the link below.

ReactiveAI RxT-Alpha-Micro-Decoder-Plus online free url in huggingface.co:

https://huggingface.co/ReactiveAI/RxT-Alpha-Micro-Decoder-Plus

RxT-Alpha-Micro-Decoder-Plus install

RxT-Alpha-Micro-Decoder-Plus is an open source model from GitHub that offers a free installation service, and any user can find RxT-Alpha-Micro-Decoder-Plus on GitHub to install. At the same time, huggingface.co provides the effect of RxT-Alpha-Micro-Decoder-Plus install, users can directly use RxT-Alpha-Micro-Decoder-Plus installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

RxT-Alpha-Micro-Decoder-Plus install url in huggingface.co:

https://huggingface.co/ReactiveAI/RxT-Alpha-Micro-Decoder-Plus

Url of RxT-Alpha-Micro-Decoder-Plus

RxT-Alpha-Micro-Decoder-Plus huggingface.co Url

Provider of RxT-Alpha-Micro-Decoder-Plus huggingface.co

ReactiveAI
ORGANIZATIONS

Other API from ReactiveAI

huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated:February 16 2026
huggingface.co

Total runs: 0
Run Growth: -1
Growth Rate: 0.00%
Updated:October 04 2025
huggingface.co

Total runs: 0
Run Growth: -1
Growth Rate: 0.00%
Updated:October 04 2025
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated:October 04 2025
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated:October 04 2025
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated:May 02 2025