CohereLabs / tiny-aya-fire

huggingface.co
Total runs: 3.4K
24-hour runs: -30
7-day runs: -83
30-day runs: 2.1K
Model's Last Updated: March 23 2026
text-generation

Introduction of tiny-aya-fire

Model Details of tiny-aya-fire

Model Card for tiny-aya-fire

Tiny Aya Fire

Best for South Asian languages. For other regions, check tiny-aya-global , tiny-aya-earth , tiny-aya-water

Model Summary

Cohere Labs Tiny Aya is an open weights research release of a pretrained 3.35 billion parameter model optimized for efficient, strong, and balanced multilingual representation across 70+ languages, including many lower-resourced ones. The model is designed to support downstream adaptation, instruction tuning, and local deployment under realistic compute constraints.

Developed by: Cohere and Cohere Labs

For more details about this model family, please check out our blog post and tech report .

Try Cohere Labs Tiny Aya

You can try out Cohere Labs Tiny Aya before downloading the weights in our hosted Hugging Face Space .

Usage

from transformers import AutoTokenizer, AutoModelForCausalLM

model_id = "CohereLabs/tiny-aya-fire"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)

# Format message with the chat template
messages = [{"role": "user", "content": "Explica en español qué significa la palabra japonesa 'ikigai' y da un ejemplo práctico."}]
inputs = tokenizer.apply_chat_template(
    messages,
    tokenize=True,
    add_generation_prompt=True,
    return_tensors="pt",
    return_dict=True,
)

gen_tokens = model.generate(
    **inputs,
    max_new_tokens=4096,
    do_sample=True,
    temperature=0.3,
    top_p=0.95
)

gen_text = tokenizer.decode(gen_tokens[0])
print(gen_text)

You can also use the model directly using transformers pipeline abstraction:

from transformers import pipeline
import torch

model_id = "CohereLabs/tiny-aya-fire"

pipe = pipeline(
    "text-generation",
    model=model_id,
    torch_dtype="auto",
    device_map="auto",
)

messages = [
    {"role": "user", "content": "Explain the Transformer architecture"},
]
outputs = pipe(
    messages,
    max_new_tokens=300,
)
print(outputs[0]["generated_text"][-1])
Model Details

Input : Text only.

Output : Model generates text.

Model Architecture : This is an auto-regressive language model that uses an optimized transformer architecture. After pretraining, this model uses supervised fine-tuning (SFT) and preference training to align model behavior to human preferences for helpfulness and safety. The model features three layers with sliding window attention (window size 4096) and RoPE for efficient local context modeling and relative positional encoding. A fourth layer uses global attention without positional embeddings, enabling unrestricted token interactions across the entire sequence.

Languages covered: The model has been trained on 70+ languages, with a focus on: English, Dutch, French, Italian, Portuguese, Romanian, Spanish, Czech, Polish, Ukrainian, Russian, Greek, German, Danish, Swedish, Norwegian, Catalan, Galician, Welsh, Irish, Basque, Croatian, Latvian, Lithuanian, Slovak, Slovenian, Estonian, Finnish, Hungarian, Serbian, Bulgarian, Arabic, Persian, Urdu, Turkish, Maltese, Hebrew, Hindi, Marathi, Bengali, Gujarati, Punjabi, Tamil, Telugu, Nepali, Tagalog, Malay, Indonesian, Vietnamese, Javanese, Khmer, Thai, Lao, Chinese, Burmese, Japanese, Korean, Amharic, Hausa, Igbo, Malagasy, Shona, Swahili, Wolof, Xhosa, Yoruba, and Zulu

Context Length: Tiny Aya supports a context length of 8K & 8K output length.

Regional Performance Heatmap

Performance Comparison

Usage and Limitations
Intended Usage

Tiny Aya is a family of massively multilingual small language models built to bring capable AI to languages that are often underserved by existing models. The models support languages across Indic, East and Southeast Asian, African, European, and Middle Eastern language families, with a deliberate emphasis on low-resource language performance.

Intended applications include multilingual text generation, conversational AI, summarization, translation and cross-lingual tasks, as well as research in multilingual NLP and low-resource language modeling. The models are also suited for efficient deployment in multilingual regions, helping bridge the digital language divide for underrepresented language communities.

Strengths

Tiny Aya demonstrates strong open-ended generation quality across its full language coverage, with particularly notable performance on low-resource languages. The model performs well on translation, summarization, and cross-lingual tasks, benefiting from training signal shared across language families and scripts.

Limitations

Reasoning tasks. The model's strongest performance is on open-ended generation and conversational tasks. Chain-of-thought reasoning tasks such as multilingual math (MGSM) are comparatively weaker.

Factual knowledge. As with any language model, outputs may contain incorrect or outdated statements, particularly in lower-resource languages with thinner training data coverage.

Uneven resource distribution. High-resource languages benefit from richer training signal and tend to exhibit more consistent quality across tasks. The lowest-resource languages in the model's coverage may show greater variability, and culturally specific nuance, sarcasm, or figurative language may be less reliably handled in these languages.

Task complexity. The model performs best with clear prompts and instructions. Highly complex or open-ended reasoning, particularly in lower-resource languages, remains challenging.

Model Card Contact

For errors or additional questions about details in this model card, contact [ [email protected] ].

Terms of Use:

We hope that the release of this model will make community-based research efforts more accessible, by releasing the weights of a highly performant 111 billion parameter model to researchers all over the world. This model is governed by a CC-BY-NC License (Non-Commercial) with an acceptable use addendum, and also requires adhering to Cohere Lab's Acceptable Use Policy . If you are interested in commercial use, please contact Cohere’s Sales team .

Try it now:

You can try Tiny Aya in our dedicated Hugging Face Space .

Runs of CohereLabs tiny-aya-fire on huggingface.co

3.4K
Total runs
-30
24-hour runs
9
3-day runs
-83
7-day runs
2.1K
30-day runs

More Information About tiny-aya-fire huggingface.co Model

More tiny-aya-fire license Visit here:

https://choosealicense.com/licenses/cc-by-nc-4.0

tiny-aya-fire huggingface.co

tiny-aya-fire huggingface.co is an AI model on huggingface.co that provides tiny-aya-fire's model effect (), which can be used instantly with this CohereLabs tiny-aya-fire model. huggingface.co supports a free trial of the tiny-aya-fire model, and also provides paid use of the tiny-aya-fire. Support call tiny-aya-fire model through api, including Node.js, Python, http.

CohereLabs tiny-aya-fire online free

tiny-aya-fire huggingface.co is an online trial and call api platform, which integrates tiny-aya-fire's modeling effects, including api services, and provides a free online trial of tiny-aya-fire, you can try tiny-aya-fire online for free by clicking the link below.

CohereLabs tiny-aya-fire online free url in huggingface.co:

https://huggingface.co/CohereLabs/tiny-aya-fire

tiny-aya-fire install

tiny-aya-fire is an open source model from GitHub that offers a free installation service, and any user can find tiny-aya-fire on GitHub to install. At the same time, huggingface.co provides the effect of tiny-aya-fire install, users can directly use tiny-aya-fire installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

tiny-aya-fire install url in huggingface.co:

https://huggingface.co/CohereLabs/tiny-aya-fire

Url of tiny-aya-fire

tiny-aya-fire huggingface.co Url

Provider of tiny-aya-fire huggingface.co

CohereLabs
ORGANIZATIONS

Other API from CohereLabs

huggingface.co

Total runs: 11.4K
Run Growth: 791
Growth Rate: 6.95%
Updated:September 11 2025
huggingface.co

Total runs: 7.2K
Run Growth: 3.0K
Growth Rate: 42.04%
Updated:September 11 2025
huggingface.co

Total runs: 1.8K
Run Growth: -535
Growth Rate: -29.92%
Updated:September 11 2025