RUCAIBox / elmer

huggingface.co
Total runs: 41
24-hour runs: 1
7-day runs: 1
30-day runs: 5
Model's Last Updated: October 28 2022
text-generation

Introduction of elmer

Model Details of elmer

ELMER

The ELMER model was proposed in ELMER: A Non-Autoregressive Pre-trained Language Model for Efficient and Effective Text Generation by Junyi Li, Tianyi Tang, Wayne Xin Zhao, Jian-Yun Nie and Ji-Rong Wen.

The detailed information and instructions can be found https://github.com/RUCAIBox/ELMER .

Model Description

ELMER is an efficient and effective PLM for NAR text generation, which generates tokens at different layers by leveraging the early exit technique.

The architecture of ELMER is a variant of the standard Transformer encoder-decoder and poses three technical contributions:

  1. For decoder, we replace the original masked multi-head attention with bi-directional multi-head attention akin to the encoder. Therefore, ELMER dynamically adjusts the output length by emitting an end token "[EOS]" at any position.
  2. Leveraging early exit, ELMER injects "off-ramps" at each decoder layer, which make predictions with intermediate hidden states. If ELMER exits at the $l$-th layer, we copy the $l$-th hidden states to the subsequent layers.
  3. ELMER utilizes a novel pre-training objective, layer permutation language modeling (LPLM), to pre-train on the large-scale corpus. LPLM permutes the exit layer for each token from 1 to the maximum layer $L$.
Examples

To fine-tune ELMER on non-autoregressive text generation:

>>> from transformers import BartTokenizer as ElmerTokenizer
>>> from transformers import BartForConditionalGeneration as ElmerForConditionalGeneration

>>> tokenizer = ElmerTokenizer.from_pretrained("RUCAIBox/elmer")
>>> model = ElmerForConditionalGeneration.from_pretrained("RUCAIBox/elmer")
Citation
@article{lijunyi2022elmer,
  title={ELMER: A Non-Autoregressive Pre-trained Language Model for Efficient and Effective Text Generation},
  author={Li, Junyi and Tang, Tianyi and Zhao, Wayne Xin and Nie, Jian-Yun and Wen, Ji-Rong},
  booktitle={EMNLP 2022},
  year={2022}
}

Runs of RUCAIBox elmer on huggingface.co

41
Total runs
1
24-hour runs
1
3-day runs
1
7-day runs
5
30-day runs

More Information About elmer huggingface.co Model

elmer huggingface.co

elmer huggingface.co is an AI model on huggingface.co that provides elmer's model effect (), which can be used instantly with this RUCAIBox elmer model. huggingface.co supports a free trial of the elmer model, and also provides paid use of the elmer. Support call elmer model through api, including Node.js, Python, http.

RUCAIBox elmer online free

elmer huggingface.co is an online trial and call api platform, which integrates elmer's modeling effects, including api services, and provides a free online trial of elmer, you can try elmer online for free by clicking the link below.

RUCAIBox elmer online free url in huggingface.co:

https://huggingface.co/RUCAIBox/elmer

elmer install

elmer is an open source model from GitHub that offers a free installation service, and any user can find elmer on GitHub to install. At the same time, huggingface.co provides the effect of elmer install, users can directly use elmer installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

elmer install url in huggingface.co:

https://huggingface.co/RUCAIBox/elmer

Url of elmer

Provider of elmer huggingface.co

RUCAIBox
ORGANIZATIONS

Other API from RUCAIBox

huggingface.co

Total runs: 4.6K
Run Growth: 390
Growth Rate: 8.56%
Updated:June 27 2022
huggingface.co

Total runs: 30
Run Growth: 23
Growth Rate: 76.67%
Updated:August 03 2023
huggingface.co

Total runs: 12
Run Growth: 3
Growth Rate: 25.00%
Updated:June 27 2022
huggingface.co

Total runs: 7
Run Growth: 0
Growth Rate: 0.00%
Updated:June 27 2022
huggingface.co

Total runs: 6
Run Growth: -1
Growth Rate: -16.67%
Updated:August 03 2023