This page houses
ARC8-Encoder_multi
from three different versions of pretrained ARC-Encoders. Architectures and methods to train them are described in the paper
ARC-Encoder: learning compressed text representations for large language models
available
here
. A code to reproduce the pretraining, further fine-tune the encoders or even evaluate them on dowstream tasks is available at
ARC-Encoder repository
.
Models Details
All the encoders released here are trained on web crawl filtered using
Dactory
based on a
Llama3.2-3B
base backbone. It consists in two ARC-Encoder specifically trained for one decoder and one for two decoders in the same time:
ARC8-Encoder_Llama
, trained on 2.6B tokens on
Llama3.1-8B
base specifically with a pooling factor of 8.
ARC8-Encoder_Mistral
, trained on 2.6B tokens on
Mistral-7B
base specifically with a pooling factor of 8.
ARC8-Encoder_multi
, trained by sampling among the two decoders with a pooling factor of 8.
Uses
As described in the
paper
, the pretrained ARC-Encoders can be fine-tuned to perform various downstream tasks.
You can also adapt an ARC-Encoder to a new pooling factor (PF) by fine-tuning it on the desired PF.
For optimal results, we recommend fine-tuning toward a lower PF than the one used during pretraining.
To reproduce the results presented in the paper, you can use our released fine-tuning dataset,
ARC_finetuning
.
Licensing
ARC-Encoders are licensed under the CC-BY 4.0 license.
Terms of use: As the released models are pretrained from Llama3.2 3B backbone, ARC-Encoders are subject to the Llama Terms of Use found at
Llama license
.
Citations
If you use one of these models, please cite:
@techreport{pilchen2025arc_encoder,
title={ARC-Encoder: learning compressed text representations for large language models},
author={Pilchen, Hippolyte and Grave, Edouard and P{\'e}rez, Patrick},
year={2025}
}
Runs of kyutai ARC8_Encoder_multi on huggingface.co
19
Total runs
0
24-hour runs
0
3-day runs
3
7-day runs
13
30-day runs
More Information About ARC8_Encoder_multi huggingface.co Model
ARC8_Encoder_multi huggingface.co is an AI model on huggingface.co that provides ARC8_Encoder_multi's model effect (), which can be used instantly with this kyutai ARC8_Encoder_multi model. huggingface.co supports a free trial of the ARC8_Encoder_multi model, and also provides paid use of the ARC8_Encoder_multi. Support call ARC8_Encoder_multi model through api, including Node.js, Python, http.
ARC8_Encoder_multi huggingface.co is an online trial and call api platform, which integrates ARC8_Encoder_multi's modeling effects, including api services, and provides a free online trial of ARC8_Encoder_multi, you can try ARC8_Encoder_multi online for free by clicking the link below.
kyutai ARC8_Encoder_multi online free url in huggingface.co:
ARC8_Encoder_multi is an open source model from GitHub that offers a free installation service, and any user can find ARC8_Encoder_multi on GitHub to install. At the same time, huggingface.co provides the effect of ARC8_Encoder_multi install, users can directly use ARC8_Encoder_multi installed effect in huggingface.co for debugging and trial. It also supports api for free installation.