google / t5-efficient-small-dl4

huggingface.co
Total runs: 7
24-hour runs: 0
7-day runs: 4
30-day runs: -5
Model's Last Updated: January 25 2023
text-generation

Introduction of t5-efficient-small-dl4

Model Details of t5-efficient-small-dl4

T5-Efficient-SMALL-DL4 (Deep-Narrow version)

T5-Efficient-SMALL-DL4 is a variation of Google's original T5 following the T5 model architecture . It is a pretrained-only checkpoint and was released with the paper Scale Efficiently: Insights from Pre-training and Fine-tuning Transformers by Yi Tay, Mostafa Dehghani, Jinfeng Rao, William Fedus, Samira Abnar, Hyung Won Chung, Sharan Narang, Dani Yogatama, Ashish Vaswani, Donald Metzler .

In a nutshell, the paper indicates that a Deep-Narrow model architecture is favorable for downstream performance compared to other model architectures of similar parameter count.

To quote the paper:

We generally recommend a DeepNarrow strategy where the model’s depth is preferentially increased before considering any other forms of uniform scaling across other dimensions. This is largely due to how much depth influences the Pareto-frontier as shown in earlier sections of the paper. Specifically, a tall small (deep and narrow) model is generally more efficient compared to the base model. Likewise, a tall base model might also generally more efficient compared to a large model. We generally find that, regardless of size, even if absolute performance might increase as we continue to stack layers, the relative gain of Pareto-efficiency diminishes as we increase the layers, converging at 32 to 36 layers. Finally, we note that our notion of efficiency here relates to any one compute dimension, i.e., params, FLOPs or throughput (speed). We report all three key efficiency metrics (number of params, FLOPS and speed) and leave this decision to the practitioner to decide which compute dimension to consider.

To be more precise, model depth is defined as the number of transformer blocks that are stacked sequentially. A sequence of word embeddings is therefore processed sequentially by each transformer block.

Details model architecture

This model checkpoint - t5-efficient-small-dl4 - is of model type Small with the following variations:

  • dl is 4

It has 52.13 million parameters and thus requires ca. 208.51 MB of memory in full precision ( fp32 ) or 104.25 MB of memory in half precision ( fp16 or bf16 ).

A summary of the original T5 model architectures can be seen here:

Model nl (el/dl) ff dm kv nh #Params
Tiny 4/4 1024 256 32 4 16M
Mini 4/4 1536 384 32 8 31M
Small 6/6 2048 512 32 8 60M
Base 12/12 3072 768 64 12 220M
Large 24/24 4096 1024 64 16 738M
Xl 24/24 16384 1024 128 32 3B
XXl 24/24 65536 1024 128 128 11B

whereas the following abbreviations are used:

Abbreviation Definition
nl Number of transformer blocks (depth)
dm Dimension of embedding vector (output vector of transformers block)
kv Dimension of key/value projection matrix
nh Number of attention heads
ff Dimension of intermediate vector within transformer block (size of feed-forward projection matrix)
el Number of transformer blocks in the encoder (encoder depth)
dl Number of transformer blocks in the decoder (decoder depth)
sh Signifies that attention heads are shared
skv Signifies that key-values projection matrices are tied

If a model checkpoint has no specific, el or dl than both the number of encoder- and decoder layers correspond to nl .

Pre-Training

The checkpoint was pretrained on the Colossal, Cleaned version of Common Crawl (C4) for 524288 steps using the span-based masked language modeling (MLM) objective.

Fine-Tuning

Note : This model is a pretrained checkpoint and has to be fine-tuned for practical usage. The checkpoint was pretrained in English and is therefore only useful for English NLP tasks. You can follow on of the following examples on how to fine-tune the model:

PyTorch :

Tensorflow :

JAX/Flax :

Downstream Performance

TODO: Add table if available

Computational Complexity

TODO: Add table if available

More information

We strongly recommend the reader to go carefully through the original paper Scale Efficiently: Insights from Pre-training and Fine-tuning Transformers to get a more nuanced understanding of this model checkpoint. As explained in the following issue , checkpoints including the sh or skv model architecture variations have not been ported to Transformers as they are probably of limited practical usage and are lacking a more detailed description. Those checkpoints are kept here as they might be ported potentially in the future.

Runs of google t5-efficient-small-dl4 on huggingface.co

7
Total runs
0
24-hour runs
0
3-day runs
4
7-day runs
-5
30-day runs

More Information About t5-efficient-small-dl4 huggingface.co Model

More t5-efficient-small-dl4 license Visit here:

https://choosealicense.com/licenses/apache-2.0

t5-efficient-small-dl4 huggingface.co

t5-efficient-small-dl4 huggingface.co is an AI model on huggingface.co that provides t5-efficient-small-dl4's model effect (), which can be used instantly with this google t5-efficient-small-dl4 model. huggingface.co supports a free trial of the t5-efficient-small-dl4 model, and also provides paid use of the t5-efficient-small-dl4. Support call t5-efficient-small-dl4 model through api, including Node.js, Python, http.

t5-efficient-small-dl4 huggingface.co Url

https://huggingface.co/google/t5-efficient-small-dl4

google t5-efficient-small-dl4 online free

t5-efficient-small-dl4 huggingface.co is an online trial and call api platform, which integrates t5-efficient-small-dl4's modeling effects, including api services, and provides a free online trial of t5-efficient-small-dl4, you can try t5-efficient-small-dl4 online for free by clicking the link below.

google t5-efficient-small-dl4 online free url in huggingface.co:

https://huggingface.co/google/t5-efficient-small-dl4

t5-efficient-small-dl4 install

t5-efficient-small-dl4 is an open source model from GitHub that offers a free installation service, and any user can find t5-efficient-small-dl4 on GitHub to install. At the same time, huggingface.co provides the effect of t5-efficient-small-dl4 install, users can directly use t5-efficient-small-dl4 installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

t5-efficient-small-dl4 install url in huggingface.co:

https://huggingface.co/google/t5-efficient-small-dl4

Url of t5-efficient-small-dl4

t5-efficient-small-dl4 huggingface.co Url

Provider of t5-efficient-small-dl4 huggingface.co

google
ORGANIZATIONS

Other API from google

huggingface.co

Total runs: 2.6M
Run Growth: 129.6K
Growth Rate: 5.10%
Updated:March 22 2025
huggingface.co

Total runs: 2.0M
Run Growth: 175.3K
Growth Rate: 8.71%
Updated:March 22 2025
huggingface.co

Total runs: 1.6M
Run Growth: 456.2K
Growth Rate: 28.67%
Updated:July 17 2023
huggingface.co

Total runs: 590.5K
Run Growth: -32.7K
Growth Rate: -5.54%
Updated:February 29 2024
huggingface.co

Total runs: 587.4K
Run Growth: 7.2K
Growth Rate: 1.22%
Updated:October 11 2023
huggingface.co

Total runs: 542.4K
Run Growth: -1.6M
Growth Rate: -277.71%
Updated:April 04 2025
huggingface.co

Total runs: 491.2K
Run Growth: 95.5K
Growth Rate: 19.44%
Updated:July 17 2023
huggingface.co

Total runs: 484.4K
Run Growth: -630.0K
Growth Rate: -126.71%
Updated:March 22 2025
huggingface.co

Total runs: 431.2K
Run Growth: 54.4K
Growth Rate: 13.02%
Updated:November 27 2023
huggingface.co

Total runs: 406.2K
Run Growth: -50.0K
Growth Rate: -12.30%
Updated:August 08 2024
huggingface.co

Total runs: 404.4K
Run Growth: 129.8K
Growth Rate: 32.09%
Updated:August 28 2024
huggingface.co

Total runs: 396.6K
Run Growth: 372.3K
Growth Rate: 99.57%
Updated:April 03 2026
huggingface.co

Total runs: 367.6K
Run Growth: -199.4K
Growth Rate: -54.01%
Updated:January 25 2023
huggingface.co

Total runs: 349.4K
Run Growth: -50.4K
Growth Rate: -14.43%
Updated:August 28 2024
huggingface.co

Total runs: 287.8K
Run Growth: -6.2K
Growth Rate: -2.15%
Updated:January 25 2023
huggingface.co

Total runs: 280.5K
Run Growth: 246.0K
Growth Rate: 99.51%
Updated:April 03 2026
huggingface.co

Total runs: 271.6K
Run Growth: -71.8K
Growth Rate: -25.40%
Updated:January 25 2023
huggingface.co

Total runs: 236.7K
Run Growth: 50.3K
Growth Rate: 22.01%
Updated:October 29 2025
huggingface.co

Total runs: 203.0K
Run Growth: -2.7K
Growth Rate: -1.30%
Updated:September 18 2023
huggingface.co

Total runs: 202.6K
Run Growth: -21.3K
Growth Rate: -10.24%
Updated:January 25 2023
huggingface.co

Total runs: 198.7K
Run Growth: 172.9K
Growth Rate: 94.65%
Updated:April 03 2026
huggingface.co

Total runs: 165.0K
Run Growth: -29.3K
Growth Rate: -19.97%
Updated:September 27 2024
huggingface.co

Total runs: 137.5K
Run Growth: 104.0K
Growth Rate: 75.79%
Updated:June 27 2024
huggingface.co

Total runs: 132.7K
Run Growth: 8.5K
Growth Rate: 6.40%
Updated:March 22 2025
huggingface.co

Total runs: 119.8K
Run Growth: 58.3K
Growth Rate: 48.81%
Updated:January 25 2023
huggingface.co

Total runs: 101.4K
Run Growth: 46.5K
Growth Rate: 45.84%
Updated:January 25 2023
huggingface.co

Total runs: 99.9K
Run Growth: 13.9K
Growth Rate: 13.57%
Updated:August 14 2025
huggingface.co

Total runs: 95.2K
Run Growth: -31.2K
Growth Rate: -31.83%
Updated:May 27 2022
huggingface.co

Total runs: 90.5K
Run Growth: 29.6K
Growth Rate: 32.67%
Updated:July 03 2023
huggingface.co

Total runs: 87.4K
Run Growth: 12.5K
Growth Rate: 15.34%
Updated:March 22 2025
huggingface.co

Total runs: 77.6K
Run Growth: -62.8K
Growth Rate: -80.92%
Updated:November 28 2023
huggingface.co

Total runs: 68.8K
Run Growth: 13.5K
Growth Rate: 19.64%
Updated:August 08 2024
huggingface.co

Total runs: 68.8K
Run Growth: 14.4K
Growth Rate: 21.00%
Updated:September 27 2024
huggingface.co

Total runs: 66.9K
Run Growth: -19.6K
Growth Rate: -29.35%
Updated:January 25 2023
huggingface.co

Total runs: 50.0K
Run Growth: -24.8K
Growth Rate: -49.60%
Updated:December 19 2025
huggingface.co

Total runs: 48.7K
Run Growth: 16.7K
Growth Rate: 39.47%
Updated:January 25 2023
huggingface.co

Total runs: 43.2K
Run Growth: -17.9K
Growth Rate: -41.43%
Updated:August 14 2024
huggingface.co

Total runs: 43.1K
Run Growth: -314.4K
Growth Rate: -729.74%
Updated:August 28 2024
huggingface.co

Total runs: 40.1K
Run Growth: 16.3K
Growth Rate: 40.66%
Updated:July 10 2025