id
stringlengths 5
118
| author
stringlengths 2
42
⌀ | lastModified
stringlengths 19
19
⌀ | downloads
float64 0
117M
⌀ | downloadsAllTime
float64 0
2.17B
⌀ | tags
stringlengths 2
28.2k
⌀ | pipeline_tag
stringclasses 52
values | createdAt
stringlengths 19
19
⌀ | dataset
stringlengths 1
5.92k
⌀ | license
stringclasses 104
values | architectures
stringlengths 2
125
⌀ | base_model
stringlengths 10
2.07k
⌀ | base_model_relation
stringclasses 11
values |
---|---|---|---|---|---|---|---|---|---|---|---|---|
bvanaken/CORe-clinical-outcome-biobert-v1
|
bvanaken
|
2025-01-17 09:30:47
| 384 | 62,729 |
['transformers', 'pytorch', 'jax', 'safetensors', 'bert', 'medical', 'clinical', 'en', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
unknown
| null |
unknown
|
bvanaken/clinical-assertion-negation-bert
|
bvanaken
|
2025-01-17 09:32:35
| 2,400 | 2,476,893 |
['transformers', 'pytorch', 'tf', 'safetensors', 'bert', 'text-classification', 'medical', 'clinical', 'assertion', 'negation', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForSequenceClassification
| null |
unknown
|
cahya/bert-base-indonesian-522M
|
cahya
|
2021-05-19 13:38:45
| 1,832 | 845,055 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'fill-mask', 'id', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
wikipedia
|
mit
|
BertForMaskedLM
| null |
unknown
|
cahya/bert-base-indonesian-NER
|
cahya
|
2023-11-03 16:02:19
| 103,212 | 1,085,761 |
['transformers', 'pytorch', 'jax', 'bert', 'token-classification', 'id', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
unknown
|
mit
|
BertForTokenClassification
| null |
unknown
|
cahya/distilbert-base-indonesian
|
cahya
|
2021-02-08 09:06:09
| 3,809 | 78,309 |
['transformers', 'pytorch', 'distilbert', 'fill-mask', 'id', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
wikipedia_id_newspapers_2018
|
mit
|
DistilBertForMaskedLM
| null |
unknown
|
cambridgeltl/SapBERT-from-PubMedBERT-fulltext
|
cambridgeltl
|
2023-06-14 19:03:02
| 2,425,242 | 33,481,497 |
['transformers', 'pytorch', 'tf', 'jax', 'safetensors', 'bert', 'feature-extraction', 'biomedical', 'lexical semantics', 'bionlp', 'biology', 'science', 'embedding', 'entity linking', 'en', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
BertModel
|
['microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract-fulltext']
|
unknown_annotated
|
almanach/camembert-large
|
almanach
|
2024-07-21 16:58:36
| 17,563 | 546,931 |
['transformers', 'pytorch', 'safetensors', 'camembert', 'fr', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
unknown
| null |
unknown
|
cammy/distilbart-cnn-12-6-finetuned-weaksup-1000
|
cammy
|
2022-02-22 08:49:00
| 191 | 349 |
['transformers', 'pytorch', 'bart', 'text2text-generation', 'generated_from_trainer', 'autotrain_compatible', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
BartForConditionalGeneration
| null |
unknown
|
cardiffnlp/bertweet-base-stance-climate
|
cardiffnlp
|
2021-05-20 14:54:22
| 119 | 2,239 |
['transformers', 'pytorch', 'tf', 'jax', 'roberta', 'text-classification', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
RobertaForSequenceClassification
| null |
unknown
|
cardiffnlp/twitter-roberta-base-2021-124m
|
cardiffnlp
|
2022-10-10 18:42:02
| 69 | 75,561 |
['transformers', 'pytorch', 'roberta', 'fill-mask', 'timelms', 'twitter', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
twitter-api
|
mit
|
RobertaForMaskedLM
| null |
unknown
|
cardiffnlp/twitter-roberta-base-emotion
|
cardiffnlp
|
2023-05-28 05:08:00
| 2,399,547 | 7,475,391 |
['transformers', 'pytorch', 'tf', 'jax', 'roberta', 'text-classification', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
RobertaForSequenceClassification
| null |
unknown
|
cardiffnlp/twitter-roberta-base-irony
|
cardiffnlp
|
2023-08-02 00:36:09
| 126,253 | 76,022,376 |
['transformers', 'pytorch', 'tf', 'jax', 'roberta', 'text-classification', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
tweet_eval
|
unknown
|
RobertaForSequenceClassification
| null |
unknown
|
cardiffnlp/twitter-roberta-base-offensive
|
cardiffnlp
|
2022-11-28 11:36:23
| 59,257 | 9,186,853 |
['transformers', 'pytorch', 'tf', 'jax', 'roberta', 'text-classification', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
RobertaForSequenceClassification
| null |
unknown
|
cardiffnlp/twitter-roberta-base-stance-climate
|
cardiffnlp
|
2022-11-28 11:45:09
| 34 | 373,417 |
['transformers', 'pytorch', 'tf', 'jax', 'roberta', 'text-classification', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
RobertaForSequenceClassification
| null |
unknown
|
cardiffnlp/twitter-xlm-roberta-base
|
cardiffnlp
|
2023-08-31 01:52:58
| 55,639 | 1,382,646 |
['transformers', 'pytorch', 'tf', 'xlm-roberta', 'fill-mask', 'multilingual', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
XLMRobertaForMaskedLM
| null |
unknown
|
carlosaguayo/cats_vs_dogs
|
carlosaguayo
|
2022-01-13 21:58:53
| 25 | 703 |
['tf-keras', 'image-classification']
|
image-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
unknown
| null |
unknown
|
castorini/afriberta_large
|
castorini
|
2023-01-13 20:01:15
| 1,407 | 458,477 |
['transformers', 'pytorch', 'tf', 'xlm-roberta', 'fill-mask', 'om', 'am', 'rw', 'rn', 'ha', 'ig', 'so', 'sw', 'ti', 'yo', 'pcm', 'multilingual', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
castorini/afriberta-corpus
|
mit
|
XLMRobertaForMaskedLM
| null |
unknown
|
castorini/t5-base-canard
|
castorini
|
2021-06-23 11:56:05
| 247 | 28,879 |
['transformers', 'pytorch', 'jax', 't5', 'text2text-generation', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
T5ForConditionalGeneration
| null |
unknown
|
chrisjay/masakhane_benchmarks
|
chrisjay
|
2021-10-20 05:55:51
| 0 | 0 |
['african-languages', 'machine-translation', 'text']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
unknown
| null |
unknown
|
ckiplab/bert-base-chinese-ner
|
ckiplab
|
2022-05-10 03:28:12
| 143,359 | 22,384,441 |
['transformers', 'pytorch', 'jax', 'bert', 'token-classification', 'zh', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
unknown
|
gpl-3.0
|
BertForTokenClassification
| null |
unknown
|
ckiplab/bert-base-chinese
|
ckiplab
|
2022-05-10 03:28:12
| 5,447 | 950,231 |
['transformers', 'pytorch', 'jax', 'bert', 'fill-mask', 'lm-head', 'zh', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
gpl-3.0
|
BertForMaskedLM
| null |
unknown
|
tohoku-nlp/bert-base-japanese-v2
|
tohoku-nlp
|
2021-09-23 13:45:31
| 15,564 | 2,691,186 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'fill-mask', 'ja', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
wikipedia
|
cc-by-sa-4.0
|
BertForMaskedLM
| null |
unknown
|
tohoku-nlp/bert-base-japanese-whole-word-masking
|
tohoku-nlp
|
2024-02-22 00:57:37
| 123,996 | 35,618,450 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'fill-mask', 'ja', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
wikipedia
|
cc-by-sa-4.0
|
BertForMaskedLM
| null |
unknown
|
tohoku-nlp/bert-base-japanese
|
tohoku-nlp
|
2024-02-22 00:57:00
| 4,118,334 | 117,009,206 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'fill-mask', 'ja', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
wikipedia
|
cc-by-sa-4.0
|
BertForMaskedLM
| null |
unknown
|
classla/bcms-bertic
|
classla
|
2021-10-29 08:20:06
| 1,608 | 237,602 |
['transformers', 'pytorch', 'electra', 'pretraining', 'hr', 'bs', 'sr', 'cnr', 'hbs', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
ElectraForPreTraining
| null |
unknown
|
climatebert/distilroberta-base-climate-f
|
climatebert
|
2023-05-04 13:05:20
| 1,865 | 173,431 |
['transformers', 'pytorch', 'safetensors', 'roberta', 'fill-mask', 'climate', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
RobertaForMaskedLM
| null |
unknown
|
climatebert/distilroberta-base-climate-sentiment
|
climatebert
|
2023-06-02 13:53:52
| 857 | 426,784 |
['transformers', 'pytorch', 'safetensors', 'roberta', 'text-classification', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
climatebert/climate_sentiment
|
apache-2.0
|
RobertaForSequenceClassification
| null |
unknown
|
clip-italian/clip-italian
|
clip-italian
|
2023-03-16 16:06:34
| 1,852 | 124,479 |
['transformers', 'pytorch', 'jax', 'vision-text-dual-encoder', 'feature-extraction', 'italian', 'bert', 'vit', 'vision', 'it', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
wit_ctl/conceptualCaptions_mscoco-it
|
gpl-3.0
|
VisionTextDualEncoderModel
| null |
unknown
|
cmarkea/distilcamembert-base-ner
|
cmarkea
|
2024-10-26 07:18:03
| 7,444 | 12,166,165 |
['transformers', 'pytorch', 'tf', 'onnx', 'safetensors', 'camembert', 'token-classification', 'fr', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
Jean-Baptiste/wikiner_fr
|
mit
|
CamembertForTokenClassification
|
['cmarkea/distilcamembert-base']
|
quantized
|
cmarkea/distilcamembert-base-sentiment
|
cmarkea
|
2024-10-26 07:15:09
| 6,737 | 723,468 |
['transformers', 'pytorch', 'tf', 'onnx', 'safetensors', 'camembert', 'text-classification', 'fr', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
amazon_reviews_multi_allocine
|
mit
|
CamembertForSequenceClassification
|
['cmarkea/distilcamembert-base']
|
quantized
|
cmarkea/distilcamembert-base
|
cmarkea
|
2023-08-01 10:05:33
| 16,296 | 960,487 |
['transformers', 'pytorch', 'tf', 'safetensors', 'camembert', 'fill-mask', 'fr', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
oscar
|
mit
|
CamembertForMaskedLM
| null |
unknown
|
cnicu/pegasus-xsum-booksum
|
cnicu
|
2022-02-26 22:13:52
| 134 | 277 |
['transformers', 'pytorch', 'pegasus', 'text2text-generation', 'autotrain_compatible', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
unknown
|
mit
|
PegasusForConditionalGeneration
| null |
unknown
|
cointegrated/LaBSE-en-ru
|
cointegrated
|
2024-03-28 13:59:30
| 5,156,430 | 8,462,395 |
['transformers', 'pytorch', 'tf', 'safetensors', 'bert', 'pretraining', 'feature-extraction', 'embeddings', 'sentence-similarity', 'ru', 'en', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForPreTraining
| null |
unknown
|
cointegrated/rubert-base-cased-nli-threeway
|
cointegrated
|
2024-04-05 09:31:57
| 153,062 | 3,021,141 |
['transformers', 'pytorch', 'onnx', 'safetensors', 'bert', 'text-classification', 'rubert', 'russian', 'nli', 'rte', 'zero-shot-classification', 'ru', 'autotrain_compatible', 'endpoints_compatible']
|
zero-shot-classification
|
2022-03-02 23:29:05
|
cointegrated/nli-rus-translated-v2021
|
unknown
|
BertForSequenceClassification
| null |
unknown
|
cointegrated/rubert-tiny-sentiment-balanced
|
cointegrated
|
2023-03-20 09:53:10
| 195,499 | 1,643,240 |
['transformers', 'pytorch', 'safetensors', 'bert', 'text-classification', 'russian', 'classification', 'sentiment', 'multiclass', 'ru', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForSequenceClassification
| null |
unknown
|
cointegrated/rubert-tiny
|
cointegrated
|
2024-02-09 18:44:51
| 8,094 | 6,751,256 |
['transformers', 'pytorch', 'safetensors', 'bert', 'pretraining', 'russian', 'fill-mask', 'embeddings', 'masked-lm', 'tiny', 'feature-extraction', 'sentence-similarity', 'ru', 'en', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
mit
|
BertForPreTraining
| null |
unknown
|
cointegrated/rubert-tiny2-cedr-emotion-detection
|
cointegrated
|
2023-03-20 09:52:11
| 11,750 | 261,498 |
['transformers', 'pytorch', 'safetensors', 'bert', 'text-classification', 'russian', 'classification', 'sentiment', 'emotion-classification', 'multiclass', 'ru', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
cedr
|
unknown
|
BertForSequenceClassification
| null |
unknown
|
cointegrated/rut5-base-absum
|
cointegrated
|
2023-03-17 10:21:13
| 54,752 | 265,350 |
['transformers', 'pytorch', 'safetensors', 't5', 'text2text-generation', 'russian', 'summarization', 'ru', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
summarization
|
2022-03-02 23:29:05
|
IlyaGusev/gazeta_csebuetnlp/xlsum_mlsum_wiki_lingua
|
mit
|
T5ForConditionalGeneration
| null |
unknown
|
cointegrated/rut5-base-multitask
|
cointegrated
|
2023-03-17 14:12:20
| 2,153 | 77,597 |
['transformers', 'pytorch', 'jax', 'safetensors', 't5', 'text2text-generation', 'russian', 'ru', 'en', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
unknown
|
mit
|
T5ForConditionalGeneration
| null |
unknown
|
cointegrated/rut5-base-paraphraser
|
cointegrated
|
2023-03-17 10:21:29
| 1,602 | 94,030 |
['transformers', 'pytorch', 'safetensors', 't5', 'text2text-generation', 'russian', 'paraphrasing', 'paraphraser', 'paraphrase', 'ru', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
cointegrated/ru-paraphrase-NMT-Leipzig
|
mit
|
T5ForConditionalGeneration
| null |
unknown
|
cointegrated/rut5-base
|
cointegrated
|
2023-03-20 09:53:23
| 3,357 | 63,261 |
['transformers', 'pytorch', 'jax', 'safetensors', 't5', 'text2text-generation', 'russian', 'ru', 'en', 'multilingual', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
unknown
|
mit
|
T5ForConditionalGeneration
| null |
unknown
|
cointegrated/rut5-small-normalizer
|
cointegrated
|
2023-03-17 10:23:17
| 108 | 5,787 |
['transformers', 'pytorch', 'jax', 'safetensors', 't5', 'text2text-generation', 'normalization', 'denoising autoencoder', 'russian', 'ru', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
unknown
|
mit
|
T5ForConditionalGeneration
| null |
unknown
|
cointegrated/rut5-small
|
cointegrated
|
2023-04-27 13:15:07
| 1,127 | 26,752 |
['transformers', 'pytorch', 'jax', 'safetensors', 'mt5', 'text2text-generation', 'paraphrasing', 'russian', 'ru', 'autotrain_compatible', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
unknown
|
mit
|
MT5ForConditionalGeneration
| null |
unknown
|
congcongwang/gpt2_medium_fine_tuned_coder
|
congcongwang
|
2021-05-21 15:06:28
| 318 | 53,826 |
['transformers', 'pytorch', 'jax', 'gpt2', 'text-generation', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
GPT2LMHeadModel
| null |
unknown
|
cross-encoder/msmarco-MiniLM-L6-en-de-v1
|
cross-encoder
|
2024-12-12 11:33:36
| 8,579 | 248,552 |
['transformers', 'pytorch', 'safetensors', 'bert', 'text-classification', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
BertForSequenceClassification
| null |
unknown
|
cross-encoder/nli-deberta-base
|
cross-encoder
|
2024-12-12 11:34:09
| 1,379 | 2,089,413 |
['transformers', 'pytorch', 'safetensors', 'deberta', 'text-classification', 'deberta-base-base', 'zero-shot-classification', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
zero-shot-classification
|
2022-03-02 23:29:05
|
multi_nli_snli
|
apache-2.0
|
DebertaForSequenceClassification
| null |
unknown
|
cross-encoder/nli-deberta-v3-base
|
cross-encoder
|
2024-12-12 11:34:21
| 135,001 | 2,108,581 |
['transformers', 'pytorch', 'safetensors', 'deberta-v2', 'text-classification', 'microsoft/deberta-v3-base', 'zero-shot-classification', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
zero-shot-classification
|
2022-03-02 23:29:05
|
multi_nli_snli
|
apache-2.0
|
DebertaV2ForSequenceClassification
| null |
unknown
|
cross-encoder/nli-deberta-v3-large
|
cross-encoder
|
2024-12-12 11:34:46
| 70,560 | 475,086 |
['transformers', 'pytorch', 'safetensors', 'deberta-v2', 'text-classification', 'microsoft/deberta-v3-large', 'zero-shot-classification', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
zero-shot-classification
|
2022-03-02 23:29:05
|
multi_nli_snli
|
apache-2.0
|
DebertaV2ForSequenceClassification
| null |
unknown
|
csebuetnlp/banglabert
|
csebuetnlp
|
2022-12-23 18:49:36
| 8,474 | 1,997,926 |
['transformers', 'pytorch', 'electra', 'pretraining', 'bn', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
ElectraForPreTraining
| null |
unknown
|
d4data/bias-detection-model
|
d4data
|
2022-08-09 02:40:59
| 9,568 | 244,743 |
['transformers', 'tf', 'distilbert', 'text-classification', 'Text Classification', 'en', 'co2_eq_emissions', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
DistilBertForSequenceClassification
| null |
unknown
|
dalle-mini/dalle-mini
|
dalle-mini
|
2023-01-11 08:53:22
| 433 | 17,415 |
['transformers', 'jax', 'dallebart', 'text-to-image', 'en', 'co2_eq_emissions']
|
text-to-image
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
eBart
| null |
unknown
|
dandelin/vilt-b32-finetuned-flickr30k
|
dandelin
|
2022-01-23 09:46:32
| 67 | 5,764 |
['transformers', 'pytorch', 'vilt', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
ViltForImageAndTextRetrieval
| null |
unknown
|
dandelin/vilt-b32-mlm-itm
|
dandelin
|
2021-11-27 10:13:10
| 537 | 16,845 |
['transformers', 'pytorch', 'vilt', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
ViltForPreTraining
| null |
unknown
|
dangvantuan/sentence-camembert-large
|
dangvantuan
|
2024-07-05 08:49:04
| 22,306 | 1,921,090 |
['sentence-transformers', 'pytorch', 'tf', 'safetensors', 'camembert', 'Text', 'Sentence Similarity', 'Sentence-Embedding', 'camembert-large', 'sentence-similarity', 'fr', 'model-index', 'autotrain_compatible', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
stsb_multi_mt
|
apache-2.0
|
CamembertModel
| null |
unknown
|
datificate/gpt2-small-spanish
|
datificate
|
2021-05-21 15:24:00
| 544,920 | 2,406,481 |
['transformers', 'pytorch', 'tf', 'jax', 'gpt2', 'text-generation', 'es', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
wikipedia
|
apache-2.0
|
GPT2LMHeadModel
| null |
unknown
|
davanstrien/book-genre-classification
|
davanstrien
|
2021-12-21 16:05:46
| 2 | 135 |
['adapter-transformers', 'bert']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
unknown
| null |
unknown
|
dbddv01/gpt2-french-small
|
dbddv01
|
2023-05-05 11:57:48
| 3,074 | 53,327 |
['transformers', 'pytorch', 'jax', 'safetensors', 'gpt2', 'text-generation', 'french', 'model', 'fr', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
GPT2LMHeadModel
| null |
unknown
|
dbmdz/bert-base-french-europeana-cased
|
dbmdz
|
2021-09-13 21:03:24
| 36,593 | 947,861 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'historic french', 'fr', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
mit
|
unknown
| null |
unknown
|
dbmdz/bert-base-german-cased
|
dbmdz
|
2023-09-06 22:19:38
| 9,249 | 1,770,013 |
['transformers', 'pytorch', 'tf', 'jax', 'safetensors', 'bert', 'fill-mask', 'de', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
mit
|
BertForMaskedLM
| null |
unknown
|
dbmdz/bert-base-german-europeana-cased
|
dbmdz
|
2024-10-28 13:47:34
| 536 | 26,425 |
['transformers', 'pytorch', 'tf', 'jax', 'safetensors', 'bert', 'historic german', 'de', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
mit
|
unknown
| null |
unknown
|
dbmdz/bert-base-german-europeana-uncased
|
dbmdz
|
2024-12-12 16:50:15
| 69 | 6,505 |
['transformers', 'pytorch', 'tf', 'jax', 'safetensors', 'bert', 'historic german', 'de', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
mit
|
unknown
| null |
unknown
|
dbmdz/bert-base-german-uncased
|
dbmdz
|
2023-09-06 22:19:33
| 12,524 | 2,864,473 |
['transformers', 'pytorch', 'tf', 'jax', 'safetensors', 'bert', 'fill-mask', 'de', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
mit
|
BertForMaskedLM
| null |
unknown
|
dbmdz/bert-base-historic-multilingual-cased
|
dbmdz
|
2023-09-06 22:15:33
| 153 | 16,590 |
['transformers', 'pytorch', 'jax', 'tensorboard', 'safetensors', 'bert', 'fill-mask', 'multilingual', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
mit
|
BertForMaskedLM
| null |
unknown
|
dbmdz/bert-base-italian-uncased
|
dbmdz
|
2021-05-19 15:00:42
| 7,484 | 498,367 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'fill-mask', 'it', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
wikipedia
|
mit
|
BertForMaskedLM
| null |
unknown
|
dbmdz/bert-base-italian-xxl-cased
|
dbmdz
|
2023-09-06 22:19:43
| 75,786 | 2,329,383 |
['transformers', 'pytorch', 'tf', 'jax', 'safetensors', 'bert', 'fill-mask', 'it', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
wikipedia
|
mit
|
BertForMaskedLM
| null |
unknown
|
dbmdz/bert-base-italian-xxl-uncased
|
dbmdz
|
2023-09-06 22:18:38
| 43,969 | 539,624 |
['transformers', 'pytorch', 'tf', 'jax', 'safetensors', 'bert', 'fill-mask', 'it', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
wikipedia
|
mit
|
BertForMaskedLM
| null |
unknown
|
dbmdz/bert-base-turkish-128k-uncased
|
dbmdz
|
2024-10-28 13:47:11
| 56,934 | 588,871 |
['transformers', 'pytorch', 'tf', 'jax', 'safetensors', 'bert', 'tr', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
mit
|
unknown
| null |
unknown
|
dbmdz/bert-base-turkish-cased
|
dbmdz
|
2024-11-16 21:32:55
| 55,076 | 5,445,332 |
['transformers', 'pytorch', 'tf', 'jax', 'safetensors', 'bert', 'tr', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
mit
|
unknown
| null |
source
|
dbmdz/bert-base-turkish-uncased
|
dbmdz
|
2024-02-20 23:10:10
| 23,863 | 790,403 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'tr', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
mit
|
unknown
| null |
unknown
|
dbmdz/bert-large-cased-finetuned-conll03-english
|
dbmdz
|
2023-09-06 22:17:56
| 1,444,675 | 29,773,454 |
['transformers', 'pytorch', 'tf', 'jax', 'rust', 'safetensors', 'bert', 'token-classification', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForTokenClassification
| null |
unknown
|
dbmdz/bert-mini-historic-multilingual-cased
|
dbmdz
|
2023-09-06 22:19:11
| 771 | 18,316 |
['transformers', 'pytorch', 'tf', 'tensorboard', 'safetensors', 'bert', 'fill-mask', 'multilingual', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
mit
|
BertForMaskedLM
| null |
unknown
|
dccuchile/bert-base-spanish-wwm-uncased
|
dccuchile
|
2024-01-18 01:46:43
| 257,371 | 7,767,126 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'fill-mask', 'masked-lm', 'es', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForMaskedLM
| null |
source
|
deepmind/language-perceiver
|
deepmind
|
2024-04-11 13:36:40
| 3,720 | 120,956 |
['transformers', 'pytorch', 'safetensors', 'perceiver', 'fill-mask', 'en', 'autotrain_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
wikipedia_c4
|
apache-2.0
|
PerceiverForMaskedLM
| null |
unknown
|
deepmind/multimodal-perceiver
|
deepmind
|
2023-09-24 08:47:37
| 365 | 33,658 |
['transformers', 'pytorch', 'perceiver', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
kinetics-700-2020
|
apache-2.0
|
PerceiverForMultimodalAutoencoding
| null |
unknown
|
deepmind/optical-flow-perceiver
|
deepmind
|
2024-08-26 08:05:14
| 1,574 | 465,767 |
['transformers', 'pytorch', 'safetensors', 'perceiver', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
autoflow
|
apache-2.0
|
PerceiverForOpticalFlow
| null |
unknown
|
deepmind/vision-perceiver-learned
|
deepmind
|
2024-08-26 08:05:30
| 1,845 | 40,654 |
['transformers', 'pytorch', 'perceiver', 'image-classification', 'autotrain_compatible', 'endpoints_compatible']
|
image-classification
|
2022-03-02 23:29:05
|
imagenet
|
apache-2.0
|
PerceiverForImageClassificationLearned
| null |
unknown
|
deepset/bert-small-mm_retrieval-question_encoder
|
deepset
|
2023-05-05 06:56:01
| 63 | 189,282 |
['transformers', 'pytorch', 'safetensors', 'dpr', 'feature-extraction', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
DPRQuestionEncoder
| null |
unknown
|
deepset/gbert-base
|
deepset
|
2024-09-26 10:59:09
| 10,335 | 2,025,221 |
['transformers', 'pytorch', 'tf', 'safetensors', 'fill-mask', 'de', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
wikipedia_OPUS_OpenLegalData
|
mit
|
BertForMaskedLM
| null |
source
|
deepset/gbert-large
|
deepset
|
2024-09-26 10:59:24
| 9,650 | 1,249,069 |
['transformers', 'pytorch', 'tf', 'safetensors', 'fill-mask', 'de', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
wikipedia_OPUS_OpenLegalData_oscar
|
mit
|
BertForMaskedLM
| null |
unknown
|
deepset/gelectra-base-germanquad
|
deepset
|
2024-09-26 07:52:17
| 2,065 | 1,257,083 |
['transformers', 'pytorch', 'tf', 'safetensors', 'electra', 'question-answering', 'exbert', 'de', 'endpoints_compatible']
|
question-answering
|
2022-03-02 23:29:05
|
deepset/germanquad
|
mit
|
ElectraForQuestionAnswering
| null |
unknown
|
deepset/gelectra-large
|
deepset
|
2024-09-26 10:57:13
| 3,172 | 73,500 |
['transformers', 'pytorch', 'tf', 'safetensors', 'electra', 'pretraining', 'de', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
wikipedia_OPUS_OpenLegalData_oscar
|
mit
|
ElectraForPreTraining
| null |
unknown
|
deepset/roberta-large-squad2
|
deepset
|
2024-09-26 08:05:59
| 43,742 | 1,720,614 |
['transformers', 'pytorch', 'jax', 'safetensors', 'roberta', 'question-answering', 'en', 'model-index', 'endpoints_compatible']
|
question-answering
|
2022-03-02 23:29:05
|
squad_v2
|
cc-by-4.0
|
RobertaForQuestionAnswering
|
['FacebookAI/roberta-large']
|
finetune
|
deepset/tinyroberta-squad2
|
deepset
|
2024-09-26 09:12:41
| 24,793 | 5,422,358 |
['transformers', 'pytorch', 'safetensors', 'roberta', 'question-answering', 'en', 'model-index', 'endpoints_compatible']
|
question-answering
|
2022-03-02 23:29:05
|
squad_v2
|
cc-by-4.0
|
RobertaForQuestionAnswering
| null |
unknown
|
deepset/xlm-roberta-large-squad2
|
deepset
|
2024-09-26 10:53:55
| 6,481 | 1,494,722 |
['transformers', 'pytorch', 'safetensors', 'xlm-roberta', 'question-answering', 'multilingual', 'model-index', 'endpoints_compatible']
|
question-answering
|
2022-03-02 23:29:05
|
squad_v2
|
cc-by-4.0
|
XLMRobertaForQuestionAnswering
| null |
unknown
|
deutsche-telekom/mt5-small-sum-de-mit-v1
|
deutsche-telekom
|
2023-05-18 05:02:05
| 488 | 8,902 |
['transformers', 'pytorch', 'safetensors', 'mt5', 'text2text-generation', 'summarization', 'de', 'autotrain_compatible', 'endpoints_compatible']
|
summarization
|
2022-03-02 23:29:05
|
swiss_text_2019
|
mit
|
MT5ForConditionalGeneration
| null |
unknown
|
dkleczek/bert-base-polish-cased-v1
|
dkleczek
|
2021-05-19 15:54:20
| 25,938 | 381,670 |
['transformers', 'pytorch', 'jax', 'bert', 'pretraining', 'pl', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForPreTraining
| null |
unknown
|
dmis-lab/biobert-large-cased-v1.1-squad
|
dmis-lab
|
2023-01-04 12:14:48
| 2,936 | 359,169 |
['transformers', 'pytorch', 'jax', 'bert', 'question-answering', 'endpoints_compatible']
|
question-answering
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForQuestionAnswering
|
['EleutherAI/gpt-neo-2.7B']
|
unknown_annotated
|
dslim/bert-large-NER
|
dslim
|
2024-10-08 07:52:08
| 155,379 | 20,379,753 |
['transformers', 'pytorch', 'tf', 'jax', 'onnx', 'safetensors', 'bert', 'token-classification', 'en', 'model-index', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
conll2003
|
mit
|
BertForTokenClassification
| null |
unknown
|
dumitrescustefan/bert-base-romanian-cased-v1
|
dumitrescustefan
|
2022-09-17 18:18:13
| 112,127 | 331,042 |
['transformers', 'pytorch', 'jax', 'bert', 'fill-mask', 'ro', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
mit
|
unknown
| null |
unknown
|
dumitrescustefan/bert-base-romanian-ner
|
dumitrescustefan
|
2022-01-24 13:23:22
| 11,111 | 89,589 |
['transformers', 'pytorch', 'bert', 'token-classification', 'ro', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
ronec
|
mit
|
BertForTokenClassification
| null |
unknown
|
edbeeching/decision_transformer_atari
|
edbeeching
|
2022-02-21 08:15:44
| 0 | 0 |
['deep-reinforcement-learning', 'reinforcement-learning']
|
reinforcement-learning
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
unknown
| null |
unknown
|
edugp/kenlm
|
edugp
|
2022-03-02 22:44:44
| 0 | 0 |
['kenlm', 'perplexity', 'n-gram', 'kneser-ney', 'bigscience', 'es', 'af', 'ar', 'arz', 'as', 'bn', 'fr', 'sw', 'eu', 'ca', 'zh', 'en', 'hi', 'ur', 'id', 'pt', 'vi', 'gu', 'kn', 'ml', 'mr', 'ta', 'te', 'yo']
|
unknown
|
2022-03-02 23:29:05
|
wikipedia_oscar
|
mit
|
unknown
| null |
unknown
|
efederici/text2tags
|
efederici
|
2023-05-21 09:34:55
| 367 | 10,622 |
['transformers', 'pytorch', 'safetensors', 't5', 'text2text-generation', 'summarization', 'tags', 'Italian', 'it', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
summarization
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
T5ForConditionalGeneration
| null |
unknown
|
ehcalabres/wav2vec2-lg-xlsr-en-speech-emotion-recognition
|
ehcalabres
|
2024-10-24 13:29:57
| 37,811 | 2,569,527 |
['transformers', 'pytorch', 'tensorboard', 'safetensors', 'wav2vec2', 'audio-classification', 'generated_from_trainer', 'endpoints_compatible']
|
audio-classification
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
Wav2Vec2ForSequenceClassification
| null |
unknown
|
elastic/distilbert-base-uncased-finetuned-conll03-english
|
elastic
|
2023-08-28 13:37:40
| 62,299 | 882,439 |
['transformers', 'pytorch', 'safetensors', 'distilbert', 'token-classification', 'en', 'model-index', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
conll2003
|
apache-2.0
|
DistilBertForTokenClassification
| null |
unknown
|
elgeish/wav2vec2-large-xlsr-53-arabic
|
elgeish
|
2022-06-04 23:37:05
| 1,268 | 55,046 |
['transformers', 'pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'audio', 'speech', 'xlsr-fine-tuning-week', 'hf-asr-leaderboard', 'ar', 'model-index', 'endpoints_compatible']
|
automatic-speech-recognition
|
2022-03-02 23:29:05
|
arabic_speech_corpus_mozilla-foundation/common_voice_6_1
|
apache-2.0
|
Wav2Vec2ForCTC
| null |
unknown
|
emilyalsentzer/Bio_ClinicalBERT
|
emilyalsentzer
|
2024-12-03 20:22:45
| 5,524,227 | 104,169,306 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'fill-mask', 'en', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
mit
|
unknown
| null |
unknown
|
emilyalsentzer/Bio_Discharge_Summary_BERT
|
emilyalsentzer
|
2022-02-27 13:59:50
| 7,715 | 2,700,584 |
['transformers', 'pytorch', 'jax', 'bert', 'fill-mask', 'en', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
mit
|
unknown
| null |
unknown
|
emrecan/bert-base-turkish-cased-mean-nli-stsb-tr
|
emrecan
|
2022-01-24 23:55:40
| 996,008 | 2,724,734 |
['sentence-transformers', 'pytorch', 'bert', 'feature-extraction', 'sentence-similarity', 'transformers', 'tr', 'autotrain_compatible', 'text-embeddings-inference', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
nli_tr_emrecan/stsb-mt-turkish
|
apache-2.0
|
BertModel
| null |
unknown
|
emrecan/convbert-base-turkish-mc4-cased-allnli_tr
|
emrecan
|
2021-12-02 14:57:01
| 431 | 1,768 |
['transformers', 'pytorch', 'convbert', 'text-classification', 'zero-shot-classification', 'nli', 'tr', 'autotrain_compatible', 'endpoints_compatible']
|
zero-shot-classification
|
2022-03-02 23:29:05
|
nli_tr
|
apache-2.0
|
ConvBertForSequenceClassification
| null |
unknown
|
Subsets and Splits
FLUX.1-dev Adapter Models
The query performs basic filtering to retrieve specific entries related to a particular base model, which provides limited analytical value.
FLUX.1-dev Adapter Models
This query retrieves a limited number of entries where the base model is from a specific vendor and relation is 'adapter', which provides basic filtering but limited analytical value.