budecosystem / hex-1

huggingface.co
Total runs: 163
24-hour runs: 0
7-day runs: 3
30-day runs: 132
Model's Last Updated: May 07 2025
text-generation

Introduction of hex-1

Model Details of hex-1

India, being one of the most linguistically diverse nations in the world, faces a major roadblock in harnessing the full potential of Generative AI. With only about 10% of the population fluent in English, the remaining 90% are effectively left behind—unable to engage with GenAI tools that are predominantly built for English-speaking users.

Most leading language models today are trained using the English language, offering little to no support for Indian languages. As a result, the depth and richness of India’s linguistic and cultural heritage are being overlooked by this global AI wave—leaving billions underserved and underrepresented. To address this gap, we need language models that are;

Proficient in Indic languages Open-source, making it available to researchers, developers, and the public Offers a commercial license, allowing businesses to freely build applications, tools, and services without restrictive usage terms Hex1: Indic LLM Built for India

Hex1 is a 4B parameter language model specifically optimized for Indian languages. It is designed to bridge the linguistic AI gap in India by enabling developers to build intelligent systems that understand and respond in native Indian languages. In its first release, Hex1 supports five major Indian languages, including Hindi, Kannada, Telugu, Tamil and Malayalam. Future versions of the model are set to expand support to more languages, broadening its usability across the Indian subcontinent.

When benchmarked against leading models like Gemma-2B, LLaMA-3.2-3B, and Sarvam-1, Hex1 delivers best-in-class performance in all five supported languages for MMLU benchmark. This makes it one of the most capable models currently available for Indic language tasks.

Quickstart

The following contains a code snippet illustrating how to use the model generate content based on given inputs.

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "budecosystem/hex-1"

# load the tokenizer and the model
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto"
)

# prepare the model input
prompt = "பொங்கல் என்றால் என்ன?."
messages = [
    {"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True,
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

# conduct text completion
generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=32768
)
output_ids = generated_ids[0][len(model_inputs.input_ids[0]):].tolist() 

content = tokenizer.decode(output_ids, skip_special_tokens=True).strip("\n")

print("content:", content)
Training results - Multilingual Task Performance Comparison
Language Hellaswag ARC-c ARC-e MMLU BoolQ
Hindi 47.85 36.68 52.14 46.73 57.61
Tamil 49.45 38.65 53.45 44.71 45.87
Telugu 50.84 37.96 53.36 46.85 51.89
Kannada 52.16 38.31 53.11 46.38 52.32
Malayalam 46.32 29.60 40.86 43.63 46.69
Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • seed: 42
  • distributed_type: multi-GPU
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 3.0
Aknowledgements

Our heartfelt thanks go to the open-source community and the trailblazers in AI research whose work has paved the way for innovations. Special shout out to the Qwen3 team for the open-source model.

Runs of budecosystem hex-1 on huggingface.co

163
Total runs
0
24-hour runs
-3
3-day runs
3
7-day runs
132
30-day runs

More Information About hex-1 huggingface.co Model

hex-1 huggingface.co

hex-1 huggingface.co is an AI model on huggingface.co that provides hex-1's model effect (), which can be used instantly with this budecosystem hex-1 model. huggingface.co supports a free trial of the hex-1 model, and also provides paid use of the hex-1. Support call hex-1 model through api, including Node.js, Python, http.

budecosystem hex-1 online free

hex-1 huggingface.co is an online trial and call api platform, which integrates hex-1's modeling effects, including api services, and provides a free online trial of hex-1, you can try hex-1 online for free by clicking the link below.

budecosystem hex-1 online free url in huggingface.co:

https://huggingface.co/budecosystem/hex-1

hex-1 install

hex-1 is an open source model from GitHub that offers a free installation service, and any user can find hex-1 on GitHub to install. At the same time, huggingface.co provides the effect of hex-1 install, users can directly use hex-1 installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

hex-1 install url in huggingface.co:

https://huggingface.co/budecosystem/hex-1

Url of hex-1

Provider of hex-1 huggingface.co

budecosystem
ORGANIZATIONS

Other API from budecosystem

huggingface.co

Total runs: 1.3K
Run Growth: -302
Growth Rate: -22.97%
Updated:September 02 2023
huggingface.co

Total runs: 9
Run Growth: -1
Growth Rate: -11.11%
Updated:September 05 2023
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated:September 07 2023