INTELLECT-1
is the first collaboratively trained 10 billion parameter language model trained from scratch on 1 trillion tokens of English text and code.
INTELLECT-1
was trained on up to 14 concurrent nodes distributed across 3 continents, with contributions from 30 independent community contributors providing compute.
The training code utilizes the
prime framework
, a scalable distributed training framework designed for fault-tolerant, dynamically scaling, high-perfomance training on unreliable, globally distributed workers.
The key abstraction that allows dynamic scaling is the
ElasticDeviceMesh
which manages dynamic global process groups for fault-tolerant communication across the internet and local process groups for communication within a node.
The model was trained using the
DiLoCo
algorithms with 100 inner steps. The global all-reduce was done with custom int8 all-reduce kernels to reduce the communication payload required, greatly reducing the communication overhead by a factor 400x.
For more detailed technical insights, please refer to our
technical paper
.
Note: You must add a BOS token at the beginning of each sample. Performance may be impacted otherwise.
Usage
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
torch.set_default_device("cuda")
model = AutoModelForCausalLM.from_pretrained("PrimeIntellect/INTELLECT-1")
tokenizer = AutoTokenizer.from_pretrained("PrimeIntellect/INTELLECT-1")
input_text = "What is the Metamorphosis of Prime Intellect about?"
input_ids = tokenizer.encode(input_text, return_tensors="pt")
output_ids = model.generate(input_ids, max_length=50, num_return_sequences=1)
output_text = tokenizer.decode(output_ids[0], skip_special_tokens=True)
print(output_text)
Example text generation pipeline
import torch
from transformers import pipeline
torch.set_default_device("cuda")
pipe = pipeline("text-generation", model="PrimeIntellect/INTELLECT-1")
print(pipe("What is prime intellect ?"))
If you use this model in your research, please cite it as follows:
@article{jaghouar2024intellect,
title={INTELLECT-1 Technical Report.},
author={Jaghouar, Sami and Ong, Jack Min and Basra, Manveer and Obeid, Fares and Straube, Jannik and Keiblinger, Michael and Bakouch, Elie and Atkins, Lucas and Panahi, Maziyar and Goddard, Charles and Ryabinin, Max and Hagemann, Johannes},
journal={arXiv preprint},
year={2024}
}
Runs of PrimeIntellect INTELLECT-1 on huggingface.co
48
Total runs
0
24-hour runs
5
3-day runs
14
7-day runs
-17
30-day runs
More Information About INTELLECT-1 huggingface.co Model
INTELLECT-1 huggingface.co is an AI model on huggingface.co that provides INTELLECT-1's model effect (), which can be used instantly with this PrimeIntellect INTELLECT-1 model. huggingface.co supports a free trial of the INTELLECT-1 model, and also provides paid use of the INTELLECT-1. Support call INTELLECT-1 model through api, including Node.js, Python, http.
INTELLECT-1 huggingface.co is an online trial and call api platform, which integrates INTELLECT-1's modeling effects, including api services, and provides a free online trial of INTELLECT-1, you can try INTELLECT-1 online for free by clicking the link below.
PrimeIntellect INTELLECT-1 online free url in huggingface.co:
INTELLECT-1 is an open source model from GitHub that offers a free installation service, and any user can find INTELLECT-1 on GitHub to install. At the same time, huggingface.co provides the effect of INTELLECT-1 install, users can directly use INTELLECT-1 installed effect in huggingface.co for debugging and trial. It also supports api for free installation.