Model name: t5_local
About T5
T5 is a text-to-text transformer model that converts input text into output text. It is based on the Transformer architecture where encoder processes the input and a decoder then genrates the output. It is pre-trained on a large corpus of text data and is capable of performing a wide range of natural language processing tasks, such as chatbots, translation, summarization, and question answering.
Read more about T5 on Wikipedia.
Supported aidb operations
- encode_text
- decode_text
- encode_text_batch
- decode_text_batch
Supported models
- t5-small (default)
- t5-base
- t5-large
- t5-3b
- t5-11b
Creating the default model
Creating a specific model
Model configuration settings
The following configuration settings are available for T5 models:
model
- The T5 model to use. The default ist5-small
.revision
- The revision of the model to use. The default isrefs/pr/15
. TODOtemperature
- The temperature to use for sampling. The default is0.0
.seed
- The random seed to use for sampling. The default is1599222198345926291
.max_tokens
- The maximum number of tokens to generate. The default is16384
.repeat_penalty
- The repetition penalty to use. The default is1.1
.repeat_last_n
- The number of tokens to consider for the repetition penalty. The default is64
.
Model credentials
No credentials are required for the T5 model.
Could this page be better? Report a problem or suggest an addition!