Character string(s) specifying the
pre-trained language model(s) to be downloaded.
For a full list of options, see
HuggingFace.
Defaults to download nothing and check currently downloaded models.
"gpt2"
(50257 vocab, 768 dims, 12 layers)
"openai-gpt"
(40478 vocab, 768 dims, 12 layers)
"bert-base-uncased"
(30522 vocab, 768 dims, 12 layers)
"bert-large-uncased"
(30522 vocab, 1024 dims, 24 layers)
"bert-base-cased"
(28996 vocab, 768 dims, 12 layers)
"bert-large-cased"
(28996 vocab, 1024 dims, 24 layers)
"bert-base-chinese"
(21128 vocab, 768 dims, 12 layers)
"bert-base-multilingual-cased"
(119547 vocab, 768 dims, 12 layers)
"distilbert-base-uncased"
(30522 vocab, 768 dims, 6 layers)
"distilbert-base-cased"
(28996 vocab, 768 dims, 6 layers)
"distilbert-base-multilingual-cased"
(119547 vocab, 768 dims, 6 layers)
"albert-base-v2"
(30000 vocab, 768 dims, 12 layers)
"albert-large-v2"
(30000 vocab, 1024 dims, 24 layers)
"roberta-base"
(50265 vocab, 768 dims, 12 layers)
"roberta-large"
(50265 vocab, 1024 dims, 24 layers)
"xlm-roberta-base"
(250002 vocab, 768 dims, 12 layers)
"xlm-roberta-large"
(250002 vocab, 1024 dims, 24 layers)
"xlnet-base-cased"
(32000 vocab, 768 dims, 12 layers)
"xlnet-large-cased"
(32000 vocab, 1024 dims, 24 layers)
"microsoft/deberta-v3-base"
(128100 vocab, 768 dims, 12 layers)
"microsoft/deberta-v3-large"
(128100 vocab, 1024 dims, 24 layers)
...
(see https://huggingface.co/models)