FlameF0X / AURORA-Tiny

huggingface.co
Total runs: 0
24-hour runs: 0
7-day runs: 0
30-day runs: 0
Model's Last Updated: August 18 2025
text-generation

Introduction of AURORA-Tiny

Model Details of AURORA-Tiny

Hii!!! This is a side project, so is not the best.

image/png

AURORA-Tiny 🌅✨

Adaptive Unified Reasoning and Organized Reasoning Architecture - Tiny

A ultra-lightweight text diffusion model that generates coherent text through iterative denoising. AURORA-Tiny combines the power of transformer architectures with diffusion processes in a compact, efficient design perfect for local training and experimentation.

The model is 6M parameters.

✨ Features
  • Ultra-Compact Design : Optimized for local training with minimal hardware requirements
  • Transformer-based Architecture : Multi-head attention with time conditioning in a tiny footprint
  • Diffusion Process : Iterative denoising for high-quality text generation
  • Flexible Training : Works with any plain text dataset from Hugging Face
  • Efficient Training : Train on CPU or modest GPUs in minutes, not hours
  • Prompt-based Generation : Support for both conditional and unconditional generation
🚀 Quick Start
Installation
pip install torch torchvision torchaudio
pip install datasets matplotlib tqdm numpy
Basic Usage
from aurora import DiffusionTrainer, TextTokenizer, DiffusionTransformer, DiffusionSchedule

# Load your dataset (or use built-in loader)
texts = load_hf_dataset("rotten_tomatoes", max_samples=3000)

# Build tokenizer
tokenizer = TextTokenizer(vocab_size=2000)
tokenizer.fit(texts)

# Initialize model
model = DiffusionTransformer(
    vocab_size=len(tokenizer.word_to_id),
    d_model=256,
    n_heads=8,
    n_layers=6
)

# Train
trainer = DiffusionTrainer(model, tokenizer, schedule, device='cuda')
trainer.train(train_loader, val_loader, epochs=15)

# Generate text
generated_text = trainer.generate("This movie is", max_length=30)
print(generated_text)
🏗️ Architecture

AURORA-Tiny uses a novel combination of:

  1. Time-Conditioned Transformers : Each transformer block receives timestep embeddings
  2. Sinusoidal Time Embeddings : Continuous time representation for the diffusion process
  3. Linear Noise Schedule : Gradual noise addition during forward diffusion
  4. DDIM-style Sampling : Deterministic sampling for consistent generation
Model Components
  • Token Embedding : Maps discrete tokens to continuous space
  • Position Encoding : Learnable positional embeddings
  • Time Conditioning : Sinusoidal embeddings injected into each layer
  • Multi-Head Attention : Standard transformer attention with time modulation
  • Output Projection : Maps back to vocabulary space

Tested on RTX 3060, batch_size=16, 15 epochs. Model size: ~2.4M parameters

🎛️ Configuration
Model Hyperparameters
model_config = {
    'vocab_size': 2000,      # Vocabulary size
    'd_model': 256,          # Hidden dimension
    'n_heads': 8,            # Attention heads
    'n_layers': 6,           # Transformer layers
    'max_seq_len': 64,       # Maximum sequence length
    'timesteps': 100         # Diffusion timesteps
}
Training Parameters
training_config = {
    'batch_size': 16,        # Batch size
    'learning_rate': 1e-4,   # Learning rate
    'weight_decay': 0.01,    # L2 regularization
    'epochs': 15,            # Training epochs
    'grad_clip': 1.0         # Gradient clipping
}
📚 Supported Datasets

AURORA-Tiny works with any text dataset from Hugging Face. Pre-configured datasets include:

  • rotten_tomatoes - Movie reviews (8.5k samples)
  • imdb - Movie reviews (50k samples)
  • ag_news - News articles (120k samples)
  • poem_sentiment - Poetry (890 samples)
  • yelp_review_full - Restaurant reviews (650k samples)
🎯 Generation Strategies
Conditional Generation
# Generate from a prompt
text = trainer.generate("The movie was", max_length=50, num_steps=20)
Unconditional Generation
# Generate from scratch
text = trainer.generate("", max_length=50, num_steps=20)
Fine-tuned Sampling
# Control generation quality vs speed
text = trainer.generate(
    prompt="Breaking news",
    max_length=100,
    num_steps=50,  # More steps = higher quality
)
🔬 Technical Details
Diffusion Process

AURORA-Tiny uses a forward diffusion process that gradually adds Gaussian noise to text embeddings:

q(x_t | x_{t-1}) = N(x_t; √(1-β_t)x_{t-1}, β_t I)

The reverse process is learned by the neural network:

p_θ(x_{t-1} | x_t, t) = N(x_{t-1}; μ_θ(x_t, t), Σ_θ(x_t, t))
Training Objective

The model is trained to minimize the variational lower bound:

L = E_t,x_0,ε [||ε - ε_θ(√(ᾱ_t)x_0 + √(1-ᾱ_t)ε, t)||²]
📈 Monitoring

Training progress is automatically tracked and visualized:

  • Loss Curves : Training and validation loss over epochs
  • Vocabulary Stats : Word frequency distributions
  • Generation Samples : Example outputs during training
🛠️ Customization
Custom Tokenizer
class CustomTokenizer(TextTokenizer):
    def __init__(self, vocab_size=5000):
        super().__init__(vocab_size)
        # Add custom preprocessing
        
    def preprocess(self, text):
        # Custom text preprocessing
        return text.lower().strip()
Custom Architecture
model = DiffusionTransformer(
    vocab_size=vocab_size,
    d_model=512,       # Larger model
    n_heads=16,        # More attention heads  
    n_layers=12,       # Deeper network
    timesteps=1000     # More diffusion steps
)
🎨 Creative Applications

AURORA-Tiny excels at:

  • Story Continuation : Complete narrative fragments
  • Style Transfer : Generate text in specific styles
  • Creative Writing : Poetry, fiction, and experimental text
  • Data Augmentation : Generate synthetic training data
  • Content Variation : Create multiple versions of text
🤝 Contributing

Contributions welcome! Areas for improvement:

  • Better noise schedules (cosine, learned schedules)
  • Advanced sampling methods (DPM-Solver, PLMS)
  • Larger model architectures
  • Multi-modal extensions
  • Evaluation benchmarks

AURORA - Where text generation meets the dawn of diffusion 🌅

Runs of FlameF0X AURORA-Tiny on huggingface.co

0
Total runs
0
24-hour runs
0
3-day runs
0
7-day runs
0
30-day runs

More Information About AURORA-Tiny huggingface.co Model

More AURORA-Tiny license Visit here:

https://choosealicense.com/licenses/apache-2.0

AURORA-Tiny huggingface.co

AURORA-Tiny huggingface.co is an AI model on huggingface.co that provides AURORA-Tiny's model effect (), which can be used instantly with this FlameF0X AURORA-Tiny model. huggingface.co supports a free trial of the AURORA-Tiny model, and also provides paid use of the AURORA-Tiny. Support call AURORA-Tiny model through api, including Node.js, Python, http.

AURORA-Tiny huggingface.co Url

https://huggingface.co/FlameF0X/AURORA-Tiny

FlameF0X AURORA-Tiny online free

AURORA-Tiny huggingface.co is an online trial and call api platform, which integrates AURORA-Tiny's modeling effects, including api services, and provides a free online trial of AURORA-Tiny, you can try AURORA-Tiny online for free by clicking the link below.

FlameF0X AURORA-Tiny online free url in huggingface.co:

https://huggingface.co/FlameF0X/AURORA-Tiny

AURORA-Tiny install

AURORA-Tiny is an open source model from GitHub that offers a free installation service, and any user can find AURORA-Tiny on GitHub to install. At the same time, huggingface.co provides the effect of AURORA-Tiny install, users can directly use AURORA-Tiny installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

AURORA-Tiny install url in huggingface.co:

https://huggingface.co/FlameF0X/AURORA-Tiny

Url of AURORA-Tiny

AURORA-Tiny huggingface.co Url

Provider of AURORA-Tiny huggingface.co

FlameF0X
ORGANIZATIONS

Other API from FlameF0X

huggingface.co

Total runs: 315
Run Growth: 27
Growth Rate: 8.57%
Updated:March 13 2026
huggingface.co

Total runs: 114
Run Growth: -76
Growth Rate: -66.67%
Updated:December 01 2025
huggingface.co

Total runs: 112
Run Growth: 16
Growth Rate: 14.29%
Updated:October 31 2025
huggingface.co

Total runs: 70
Run Growth: 70
Growth Rate: 100.00%
Updated:December 03 2025
huggingface.co

Total runs: 62
Run Growth: 12
Growth Rate: 19.35%
Updated:November 29 2025
huggingface.co

Total runs: 50
Run Growth: 47
Growth Rate: 94.00%
Updated:June 29 2025
huggingface.co

Total runs: 44
Run Growth: 11
Growth Rate: 25.00%
Updated:November 25 2025
huggingface.co

Total runs: 39
Run Growth: -135
Growth Rate: -346.15%
Updated:October 23 2025
huggingface.co

Total runs: 36
Run Growth: 20
Growth Rate: 55.56%
Updated:February 23 2026
huggingface.co

Total runs: 21
Run Growth: -10
Growth Rate: -47.62%
Updated:October 17 2025
huggingface.co

Total runs: 16
Run Growth: 0
Growth Rate: 0.00%
Updated:December 19 2025
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated:December 06 2025
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated:December 29 2025
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated:December 31 2025
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated:April 19 2026
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated:April 25 2026