⚠️ T5 is currently only available via the
keras-hub-nightly
package. Use
pip install keras-hub-nightly
to try this model.
T5 encoder-decoder backbone model.
T5 is a LLM pretrained on a mix of unsupervised and supervised tasks,
where each task is converted to a sequence-to-sequence format.
T5 works well on a variety of tasks out-of-the-box by prepending
various prefixex to the input sequence, e.g., for translation:
"translate English to German: ..."
, for summarization:
"summarize: ..."
.
The default constructor gives a fully customizable, randomly initialized T5
model with any number of layers, heads, and embedding dimensions. To load
preset architectures and weights, use the
from_preset
constructor.
Disclaimer: Pre-trained models are provided on an "as is" basis, without
warranties or conditions of any kind.
Arguments
vocabulary_size
: int. The size of the token vocabulary.
num_layers
: int. The number of Transformer layers.
num_heads
: int. The number of attention heads for each Transformer.
The hidden size must be divisible by the number of attention heads.
hidden_dim
: int. The hidden size of the Transformer layers.
intermediate_dim
: int. The output dimension of the first Dense layer in
a two-layer feedforward network for each Transformer layer.
key_value_dim
: int. The dimension of each head of the key/value
projections in the multi-head attention layers. Defaults to
hidden_dim / num_heads.
dropout
: float. Dropout probability for the Transformer layers.
activation
: activation function (or activation string name). The
activation to be used in the inner dense blocks of the
Transformer layers. Defaults to
"relu"
.
use_gated_activation
: boolean. Whether to use activation gating in
the inner dense blocks of the Transformer layers.
The original T5 architecture didn't use gating, but more
recent versions do. Defaults to
True
.
layer_norm_epsilon
: float. Epsilon factor to be used in the
layer normalization layers in the Transformer layers.
tie_embedding_weights
: boolean. If
True
, the weights of the token
embedding and the weights projecting language model outputs from
hidden_dim
Runs of keras t5_large_multi on huggingface.co
2
Total runs
0
24-hour runs
0
3-day runs
1
7-day runs
1
30-day runs
More Information About t5_large_multi huggingface.co Model
t5_large_multi huggingface.co is an AI model on huggingface.co that provides t5_large_multi's model effect (), which can be used instantly with this keras t5_large_multi model. huggingface.co supports a free trial of the t5_large_multi model, and also provides paid use of the t5_large_multi. Support call t5_large_multi model through api, including Node.js, Python, http.
t5_large_multi huggingface.co is an online trial and call api platform, which integrates t5_large_multi's modeling effects, including api services, and provides a free online trial of t5_large_multi, you can try t5_large_multi online for free by clicking the link below.
keras t5_large_multi online free url in huggingface.co:
t5_large_multi is an open source model from GitHub that offers a free installation service, and any user can find t5_large_multi on GitHub to install. At the same time, huggingface.co provides the effect of t5_large_multi install, users can directly use t5_large_multi installed effect in huggingface.co for debugging and trial. It also supports api for free installation.