modelId
stringlengths
5
139
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]date
2020-02-15 11:33:14
2025-08-27 06:27:59
downloads
int64
0
223M
likes
int64
0
11.7k
library_name
stringclasses
521 values
tags
listlengths
1
4.05k
pipeline_tag
stringclasses
55 values
createdAt
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-08-27 06:27:44
card
stringlengths
11
1.01M
Oragani/JimmyWong
Oragani
2022-11-27T02:55:18Z
0
0
null
[ "license:wtfpl", "region:us" ]
null
2022-11-27T02:26:18Z
--- license: wtfpl --- A model fine-tuned on 42 images of Jimmy Wong (from BONELAB) at 6000 steps using Dreambooth. The token for jimmy is jimmywong. Credit to SLZ for making BONELAB and Jimmy Wong for acting as Jimmy in BONELAB.
mfigurski80/relation-distilbert-em
mfigurski80
2022-11-27T02:36:19Z
164
0
transformers
[ "transformers", "pytorch", "distilbert", "fill-mask", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-11-24T00:15:13Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: relation-distilbert-em results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # relation-distilbert-em This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7801 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.6947 | 1.0 | 2812 | 0.7616 | | 0.6946 | 2.0 | 5624 | 0.7740 | | 0.6944 | 3.0 | 8436 | 0.7801 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
fanpu/final_model_output_subreddit-wallstreetbets_1
fanpu
2022-11-27T02:18:17Z
113
0
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "generated_from_trainer", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2022-11-26T18:35:23Z
--- tags: - generated_from_trainer model-index: - name: final_model_output_subreddit-wallstreetbets_1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # final_model_output_subreddit-wallstreetbets_1 This model was trained from scratch on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 512 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 1000 - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.24.0 - Pytorch 1.13.0+cu117 - Datasets 2.7.1 - Tokenizers 0.13.2
pglee/outputs
pglee
2022-11-27T02:01:47Z
103
1
transformers
[ "transformers", "pytorch", "deberta-v2", "text-classification", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-11-26T09:24:02Z
--- license: mit tags: - generated_from_trainer metrics: - accuracy - f1 - precision - recall model-index: - name: outputs results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # outputs This model is a fine-tuned version of [microsoft/deberta-v3-small](https://huggingface.co/microsoft/deberta-v3-small) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0926 - Accuracy: 0.8780 - F1: 0.3881 - Precision: 0.5417 - Recall: 0.3023 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 8e-05 - train_batch_size: 256 - eval_batch_size: 512 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 4 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:| | No log | 1.0 | 6 | 0.0874 | 0.8810 | 0.4118 | 0.56 | 0.3256 | | No log | 2.0 | 12 | 0.0936 | 0.8839 | 0.4000 | 0.5909 | 0.3023 | | No log | 3.0 | 18 | 0.0922 | 0.8780 | 0.3881 | 0.5417 | 0.3023 | | No log | 4.0 | 24 | 0.0926 | 0.8780 | 0.3881 | 0.5417 | 0.3023 | ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0 - Datasets 2.1.0 - Tokenizers 0.12.1
novelai-dev/DDPB-hypernetworks
novelai-dev
2022-11-27T01:51:42Z
0
13
null
[ "license:agpl-3.0", "region:us" ]
null
2022-10-21T00:13:32Z
--- license: agpl-3.0 inference: false --- # DDPB-Hypernetworks Store Hypernetwork models curated on https://tags.novelai.dev
mynameisai/jq
mynameisai
2022-11-27T00:51:12Z
7
0
diffusers
[ "diffusers", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2022-11-27T00:48:34Z
--- license: creativeml-openrail-m tags: - text-to-image - stable-diffusion --- ### JQ Dreambooth model trained by mynameisai with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) Or you can run your new concept via `diffusers` [Colab Notebook for Inference](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_dreambooth_inference.ipynb) Sample pictures of this concept: ![0](https://huggingface.co/mynameisai/jq/resolve/main/sample_images/JQ_(01).png)
jacobthebanana/vit-synthetic-asl-numbers
jacobthebanana
2022-11-26T23:49:08Z
16
0
transformers
[ "transformers", "jax", "vit", "image-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2022-11-26T20:44:22Z
Technical details: - Base model: `google/vit-base-patch16-224` - Dataset: [synthetic ASL number dataset](https://www.kaggle.com/datasets/lexset/synthetic-asl-numbers). - Learning rate: 0.0001 - Effective training batch size: 16 (2 items per TPU core over 8 cores.) > Research supported with Cloud TPUs from Google's TPU Research Cloud (TRC)
huggingtweets/autogynefiles-s4m31p4n-tyler02020202
huggingtweets
2022-11-26T22:37:13Z
114
0
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2022-11-26T22:35:40Z
--- language: en thumbnail: http://www.huggingtweets.com/autogynefiles-s4m31p4n-tyler02020202/1669502229985/predictions.png tags: - huggingtweets widget: - text: "My dream is" --- <div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1576956835268943872/iFTl7tMV_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1529675700772302848/uXtYNx_v_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1474994961896644608/um4unzmz_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">April Clark & ppigg & tyler</div> <div style="text-align: center; font-size: 14px;">@autogynefiles-s4m31p4n-tyler02020202</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from April Clark & ppigg & tyler. | Data | April Clark | ppigg | tyler | | --- | --- | --- | --- | | Tweets downloaded | 3226 | 2858 | 2788 | | Retweets | 523 | 974 | 123 | | Short tweets | 482 | 568 | 647 | | Tweets kept | 2221 | 1316 | 2018 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/6tyh8dww/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @autogynefiles-s4m31p4n-tyler02020202's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/afwulizr) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/afwulizr/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/autogynefiles-s4m31p4n-tyler02020202') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
dante97/bertweet-isarcasm
dante97
2022-11-26T22:27:46Z
0
0
null
[ "license:mit", "region:us" ]
null
2022-11-26T22:24:51Z
--- license: mit --- A BerTweet-based model trained on iSarcasm dataset from the Task 6 of the [SemEval 2022](https://codalab.lisn.upsaclay.fr/competitions/1340) competition.
huggingtweets/andruyeung-hackwithzach
huggingtweets
2022-11-26T22:11:10Z
115
0
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2022-11-26T22:09:40Z
--- language: en thumbnail: http://www.huggingtweets.com/andruyeung-hackwithzach/1669500666413/predictions.png tags: - huggingtweets widget: - text: "My dream is" --- <div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1588851257594155008/2YVuHL3Y_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1501725597646696450/zXLLzq3Y_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Zacharⓨ Eisenhauer & Andrew Yeung</div> <div style="text-align: center; font-size: 14px;">@andruyeung-hackwithzach</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Zacharⓨ Eisenhauer & Andrew Yeung. | Data | Zacharⓨ Eisenhauer | Andrew Yeung | | --- | --- | --- | | Tweets downloaded | 481 | 3221 | | Retweets | 44 | 187 | | Short tweets | 141 | 704 | | Tweets kept | 296 | 2330 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/5oljjc5v/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @andruyeung-hackwithzach's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3861ni9l) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3861ni9l/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/andruyeung-hackwithzach') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
jerpint/whisper
jerpint
2022-11-26T22:07:52Z
0
2
null
[ "translation", "speech", "audio", "automatic-speech-recognition", "en", "dataset:whisper", "license:mit", "region:us" ]
translation
2022-11-25T16:52:43Z
--- language: - en tags: - translation - speech - audio - automatic-speech-recognition datasets: - whisper metrics: - WER license: mit --- This model was forked from the original [OpenAI whisper model](https://github.com/openai/whisper). # Whisper ## Model Whisper is a multi-lingual speech-to-text model. It takes in raw audio recordings from many languages and outputs transcriptions in the language of origin or translated to english. The model first converts speech to spectrograms, then uses an auto-regressive transformer to decode the speech to text. Here is an overview of the architecture: ![model_architecure](https://github.com/jerpint/whisper/raw/main/approach.png) For more information on the technical implementations, consult the [paper](https://cdn.openai.com/papers/whisper.pdf). ## Training Data The model was trained on 680 000 hours of audio and associated transcripts trained from the internet. The majority of the audio is in english (~65%) while the remainder is in other languages. A total of 98 different languages were used in the dataset. ![image](https://user-images.githubusercontent.com/18450628/204110014-e2684385-d790-4dd7-8ce1-47168efb2726.png) ## Model Variations OpenAI has released 9 different versions of the model, trained either on english-only audio or on multilingual data. | Size | Parameters | English-only model | Multilingual model | Required VRAM | Relative speed | |:------:|:----------:|:------------------:|:------------------:|:-------------:|:--------------:| | tiny | 39 M | `tiny.en` | `tiny` | ~1 GB | ~32x | | base | 74 M | `base.en` | `base` | ~1 GB | ~16x | | small | 244 M | `small.en` | `small` | ~2 GB | ~6x | | medium | 769 M | `medium.en` | `medium` | ~5 GB | ~2x | | large | 1550 M | N/A | `large` | ~10 GB | 1x | ## Limitations and bias In the [paper](https://cdn.openai.com/papers/whisper.pdf), they find a direct corelation between performance on a given language and the amount of data available in the dataset. As such, languages that are under-represented in the scraped dataset perform less well in whisper. Because english is much more prevalent than other languages, the model will likely perform better in english. This is shown in the following figure, where a lower word error rate (WER) indicates a better performance: ![model_performance](https://github.com/jerpint/whisper/raw/main/language-breakdown.svg)
kejian/final-cond-25-0.05-again
kejian
2022-11-26T21:11:00Z
2
0
transformers
[ "transformers", "pytorch", "gpt2", "generated_from_trainer", "en", "dataset:kejian/codeparrot-train-more-filter-3.3b-cleaned", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
2022-11-25T18:28:16Z
--- language: - en license: apache-2.0 tags: - generated_from_trainer datasets: - kejian/codeparrot-train-more-filter-3.3b-cleaned model-index: - name: kejian/final-cond-25-0.05-again results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # kejian/final-cond-25-0.05-again This model was trained from scratch on the kejian/codeparrot-train-more-filter-3.3b-cleaned dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0008 - train_batch_size: 32 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.01 - training_steps: 50354 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.23.0 - Pytorch 1.13.0+cu116 - Datasets 2.0.0 - Tokenizers 0.12.1 # Full config {'dataset': {'conditional_training_config': {'aligned_prefix': '<|aligned|>', 'drop_token_fraction': 0.05, 'misaligned_prefix': '<|misaligned|>', 'threshold': 0.000475}, 'datasets': ['kejian/codeparrot-train-more-filter-3.3b-cleaned'], 'is_split_by_sentences': True}, 'generation': {'batch_size': 64, 'metrics_configs': [{}, {'n': 1}, {}], 'scenario_configs': [{'display_as_html': True, 'generate_kwargs': {'do_sample': True, 'eos_token_id': 0, 'max_length': 704, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'unconditional', 'num_samples': 512, 'prefix': '<|aligned|>', 'use_prompt_for_scoring': False}, {'display_as_html': True, 'generate_kwargs': {'do_sample': True, 'eos_token_id': 0, 'max_length': 272, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'functions', 'num_samples': 512, 'prefix': '<|aligned|>', 'prompt_before_control': True, 'prompts_path': 'resources/functions_csnet.jsonl', 'use_prompt_for_scoring': True}], 'scorer_config': {}}, 'kl_gpt3_callback': {'gpt3_kwargs': {'model_name': 'code-cushman-001'}, 'max_tokens': 64, 'num_samples': 4096, 'prefix': '<|aligned|>'}, 'model': {'from_scratch': True, 'gpt2_config_kwargs': {'reorder_and_upcast_attn': True, 'scale_attn_by': True}, 'num_additional_tokens': 2, 'path_or_name': 'codeparrot/codeparrot-small'}, 'objective': {'name': 'MLE'}, 'tokenizer': {'path_or_name': 'codeparrot/codeparrot-small', 'special_tokens': ['<|aligned|>', '<|misaligned|>']}, 'training': {'dataloader_num_workers': 0, 'effective_batch_size': 64, 'evaluation_strategy': 'no', 'fp16': True, 'hub_model_id': 'kejian/final-cond-25-0.05-again', 'hub_strategy': 'all_checkpoints', 'learning_rate': 0.0008, 'logging_first_step': True, 'logging_steps': 1, 'num_tokens': 3300000000.0, 'output_dir': 'training_output', 'per_device_train_batch_size': 16, 'push_to_hub': True, 'remove_unused_columns': False, 'save_steps': 5000, 'save_strategy': 'steps', 'seed': 42, 'warmup_ratio': 0.01, 'weight_decay': 0.1}} # Wandb URL: https://wandb.ai/kejian/uncategorized/runs/27z94dvy
research-backup/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-2
research-backup
2022-11-26T19:36:07Z
103
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T14:08:49Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-2 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3502673796791444 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.34718100890207715 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.48916064480266813 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.662 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3508771929824561 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.37037037037037035 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8389332529757421 - name: F1 (macro) type: f1_macro value: 0.8239067926315072 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.715962441314554 - name: F1 (macro) type: f1_macro value: 0.416903883488565 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.5200433369447454 - name: F1 (macro) type: f1_macro value: 0.5091748225464606 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9359393475690339 - name: F1 (macro) type: f1_macro value: 0.8331186543970518 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8166718896897525 - name: F1 (macro) type: f1_macro value: 0.8127585517857737 --- # relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-2 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-2/raw/main/analogy.json)): - Accuracy on SAT (full): 0.3502673796791444 - Accuracy on SAT: 0.34718100890207715 - Accuracy on BATS: 0.48916064480266813 - Accuracy on U2: 0.3508771929824561 - Accuracy on U4: 0.37037037037037035 - Accuracy on Google: 0.662 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-2/raw/main/classification.json)): - Micro F1 score on BLESS: 0.8389332529757421 - Micro F1 score on CogALexV: 0.715962441314554 - Micro F1 score on EVALution: 0.5200433369447454 - Micro F1 score on K&H+N: 0.9359393475690339 - Micro F1 score on ROOT09: 0.8166718896897525 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-2/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-2") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: average - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 9 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 2 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-2/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
amagzari/bart-base-finetuned-cnn_dailymail
amagzari
2022-11-26T19:34:17Z
131
0
transformers
[ "transformers", "pytorch", "tensorboard", "bart", "text2text-generation", "summarization", "generated_from_trainer", "dataset:cnn_dailymail", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
summarization
2022-11-26T11:16:15Z
--- license: apache-2.0 tags: - summarization - generated_from_trainer datasets: - cnn_dailymail metrics: - rouge model-index: - name: bart-base-finetuned-cnn_dailymail results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: cnn_dailymail type: cnn_dailymail config: 3.0.0 split: train args: 3.0.0 metrics: - name: Rouge1 type: rouge value: 0.35105989316705805 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-base-finetuned-cnn_dailymail This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on the cnn_dailymail dataset. It achieves the following results on the evaluation set: - Loss: 1.5396 - Rouge1: 0.3511 - Rouge2: 0.1925 - Rougel: 0.3086 - Rougelsum: 0.3292 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5.6e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | |:-------------:|:-----:|:------:|:---------------:|:------:|:------:|:------:|:---------:| | 1.9486 | 1.0 | 35890 | 1.5941 | 0.3498 | 0.1893 | 0.3063 | 0.3272 | | 1.6706 | 2.0 | 71780 | 1.5601 | 0.3503 | 0.1916 | 0.3079 | 0.3279 | | 1.4809 | 3.0 | 107670 | 1.5423 | 0.3520 | 0.1923 | 0.3086 | 0.3295 | | 1.3293 | 4.0 | 143560 | 1.5396 | 0.3511 | 0.1925 | 0.3086 | 0.3292 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
research-backup/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-2
research-backup
2022-11-26T19:32:41Z
108
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T14:07:12Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-2 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3609625668449198 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3649851632047478 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.44691495275152865 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.716 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.38596491228070173 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.375 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8740394756667169 - name: F1 (macro) type: f1_macro value: 0.8618347056127108 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.7246478873239436 - name: F1 (macro) type: f1_macro value: 0.4392658356578824 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.5243770314192849 - name: F1 (macro) type: f1_macro value: 0.5089556262366394 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9525631216526397 - name: F1 (macro) type: f1_macro value: 0.8724005012414945 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8564713256032591 - name: F1 (macro) type: f1_macro value: 0.8510225054091837 --- # relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-2 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-2/raw/main/analogy.json)): - Accuracy on SAT (full): 0.3609625668449198 - Accuracy on SAT: 0.3649851632047478 - Accuracy on BATS: 0.44691495275152865 - Accuracy on U2: 0.38596491228070173 - Accuracy on U4: 0.375 - Accuracy on Google: 0.716 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-2/raw/main/classification.json)): - Micro F1 score on BLESS: 0.8740394756667169 - Micro F1 score on CogALexV: 0.7246478873239436 - Micro F1 score on EVALution: 0.5243770314192849 - Micro F1 score on K&H+N: 0.9525631216526397 - Micro F1 score on ROOT09: 0.8564713256032591 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-2/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-2") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: average - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 9 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 2 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-2/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
research-backup/relbert-roberta-base-semeval2012-v6-average-prompt-c-loob-2
research-backup
2022-11-26T19:29:12Z
103
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T14:05:35Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-average-prompt-c-loob-2 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.39572192513368987 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3857566765578635 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.47304057809894384 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.724 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.37280701754385964 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3611111111111111 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8758475214705439 - name: F1 (macro) type: f1_macro value: 0.8691352141617384 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.7424882629107981 - name: F1 (macro) type: f1_macro value: 0.48330525572765204 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.5758396533044421 - name: F1 (macro) type: f1_macro value: 0.5620753826217768 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9492940112679975 - name: F1 (macro) type: f1_macro value: 0.8534235407176715 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8314008147916013 - name: F1 (macro) type: f1_macro value: 0.8245727533841717 --- # relbert/relbert-roberta-base-semeval2012-v6-average-prompt-c-loob-2 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-c-loob-2/raw/main/analogy.json)): - Accuracy on SAT (full): 0.39572192513368987 - Accuracy on SAT: 0.3857566765578635 - Accuracy on BATS: 0.47304057809894384 - Accuracy on U2: 0.37280701754385964 - Accuracy on U4: 0.3611111111111111 - Accuracy on Google: 0.724 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-c-loob-2/raw/main/classification.json)): - Micro F1 score on BLESS: 0.8758475214705439 - Micro F1 score on CogALexV: 0.7424882629107981 - Micro F1 score on EVALution: 0.5758396533044421 - Micro F1 score on K&H+N: 0.9492940112679975 - Micro F1 score on ROOT09: 0.8314008147916013 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-c-loob-2/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-average-prompt-c-loob-2") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: average - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 9 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 2 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-c-loob-2/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
research-backup/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-2
research-backup
2022-11-26T19:25:52Z
103
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T14:03:57Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-2 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3850267379679144 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.39465875370919884 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.45525291828793774 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.698 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4298245614035088 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4074074074074074 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8725327708301943 - name: F1 (macro) type: f1_macro value: 0.8613482188522172 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.7504694835680751 - name: F1 (macro) type: f1_macro value: 0.5059131213359842 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.5687973997833152 - name: F1 (macro) type: f1_macro value: 0.5544548313192087 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9481115670863184 - name: F1 (macro) type: f1_macro value: 0.8536166449236237 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8113444061422752 - name: F1 (macro) type: f1_macro value: 0.8063280760200916 --- # relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-2 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-2/raw/main/analogy.json)): - Accuracy on SAT (full): 0.3850267379679144 - Accuracy on SAT: 0.39465875370919884 - Accuracy on BATS: 0.45525291828793774 - Accuracy on U2: 0.4298245614035088 - Accuracy on U4: 0.4074074074074074 - Accuracy on Google: 0.698 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-2/raw/main/classification.json)): - Micro F1 score on BLESS: 0.8725327708301943 - Micro F1 score on CogALexV: 0.7504694835680751 - Micro F1 score on EVALution: 0.5687973997833152 - Micro F1 score on K&H+N: 0.9481115670863184 - Micro F1 score on ROOT09: 0.8113444061422752 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-2/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-2") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: average - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 9 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 2 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-2/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
research-backup/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-1
research-backup
2022-11-26T19:13:24Z
105
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T13:58:36Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-1 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4037433155080214 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.39762611275964393 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.5125069483046136 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.774 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.42105263157894735 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4375 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8912159108030737 - name: F1 (macro) type: f1_macro value: 0.8820583021414006 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.7558685446009391 - name: F1 (macro) type: f1_macro value: 0.48835340275714023 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.5644637053087758 - name: F1 (macro) type: f1_macro value: 0.5498243307249566 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9558322320372817 - name: F1 (macro) type: f1_macro value: 0.8767567777144254 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8621121905358822 - name: F1 (macro) type: f1_macro value: 0.8578220990204453 --- # relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-1 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-1/raw/main/analogy.json)): - Accuracy on SAT (full): 0.4037433155080214 - Accuracy on SAT: 0.39762611275964393 - Accuracy on BATS: 0.5125069483046136 - Accuracy on U2: 0.42105263157894735 - Accuracy on U4: 0.4375 - Accuracy on Google: 0.774 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-1/raw/main/classification.json)): - Micro F1 score on BLESS: 0.8912159108030737 - Micro F1 score on CogALexV: 0.7558685446009391 - Micro F1 score on EVALution: 0.5644637053087758 - Micro F1 score on K&H+N: 0.9558322320372817 - Micro F1 score on ROOT09: 0.8621121905358822 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-1/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-1") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: average - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 9 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 1 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-1/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
research-backup/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-1
research-backup
2022-11-26T19:06:37Z
103
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T13:55:22Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-1 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3155080213903743 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.32344213649851633 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.556976097832129 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.746 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.34649122807017546 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.42592592592592593 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8363718547536537 - name: F1 (macro) type: f1_macro value: 0.818375851898995 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.7342723004694836 - name: F1 (macro) type: f1_macro value: 0.45135672894929957 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.5075839653304443 - name: F1 (macro) type: f1_macro value: 0.480886490627412 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9342004590665647 - name: F1 (macro) type: f1_macro value: 0.8274510913559776 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8398621121905359 - name: F1 (macro) type: f1_macro value: 0.8339376287105001 --- # relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-1 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-1/raw/main/analogy.json)): - Accuracy on SAT (full): 0.3155080213903743 - Accuracy on SAT: 0.32344213649851633 - Accuracy on BATS: 0.556976097832129 - Accuracy on U2: 0.34649122807017546 - Accuracy on U4: 0.42592592592592593 - Accuracy on Google: 0.746 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-1/raw/main/classification.json)): - Micro F1 score on BLESS: 0.8363718547536537 - Micro F1 score on CogALexV: 0.7342723004694836 - Micro F1 score on EVALution: 0.5075839653304443 - Micro F1 score on K&H+N: 0.9342004590665647 - Micro F1 score on ROOT09: 0.8398621121905359 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-1/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-1") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: average - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 9 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 1 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-b-loob-1/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
research-backup/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-0
research-backup
2022-11-26T18:57:31Z
104
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T13:52:07Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-0 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.35561497326203206 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3560830860534125 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.42857142857142855 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.668 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3815789473684211 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3587962962962963 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8540003013409673 - name: F1 (macro) type: f1_macro value: 0.8415316548238166 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.6960093896713615 - name: F1 (macro) type: f1_macro value: 0.4121210944195095 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.5021668472372698 - name: F1 (macro) type: f1_macro value: 0.4940624714187321 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9437991236001948 - name: F1 (macro) type: f1_macro value: 0.8505990194719038 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.823879661548104 - name: F1 (macro) type: f1_macro value: 0.8171304236528446 --- # relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-0 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-0/raw/main/analogy.json)): - Accuracy on SAT (full): 0.35561497326203206 - Accuracy on SAT: 0.3560830860534125 - Accuracy on BATS: 0.42857142857142855 - Accuracy on U2: 0.3815789473684211 - Accuracy on U4: 0.3587962962962963 - Accuracy on Google: 0.668 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-0/raw/main/classification.json)): - Micro F1 score on BLESS: 0.8540003013409673 - Micro F1 score on CogALexV: 0.6960093896713615 - Micro F1 score on EVALution: 0.5021668472372698 - Micro F1 score on K&H+N: 0.9437991236001948 - Micro F1 score on ROOT09: 0.823879661548104 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-0/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-0") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: average - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 8 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 0 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-e-loob-0/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
research-backup/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-0
research-backup
2022-11-26T18:54:07Z
104
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T13:50:34Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-0 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3181818181818182 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3115727002967359 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.5780989438576987 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.75 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.38596491228070173 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.36574074074074076 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8646979056802773 - name: F1 (macro) type: f1_macro value: 0.8512823829061237 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.7136150234741784 - name: F1 (macro) type: f1_macro value: 0.4307168102299423 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.5563380281690141 - name: F1 (macro) type: f1_macro value: 0.5317021379158652 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9493635668080963 - name: F1 (macro) type: f1_macro value: 0.8585118986669762 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8232528987778126 - name: F1 (macro) type: f1_macro value: 0.8208236307090298 --- # relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-0 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-0/raw/main/analogy.json)): - Accuracy on SAT (full): 0.3181818181818182 - Accuracy on SAT: 0.3115727002967359 - Accuracy on BATS: 0.5780989438576987 - Accuracy on U2: 0.38596491228070173 - Accuracy on U4: 0.36574074074074076 - Accuracy on Google: 0.75 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-0/raw/main/classification.json)): - Micro F1 score on BLESS: 0.8646979056802773 - Micro F1 score on CogALexV: 0.7136150234741784 - Micro F1 score on EVALution: 0.5563380281690141 - Micro F1 score on K&H+N: 0.9493635668080963 - Micro F1 score on ROOT09: 0.8232528987778126 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-0/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-0") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: average - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 10 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 0 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-average-prompt-d-loob-0/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
research-backup/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-2
research-backup
2022-11-26T18:34:45Z
103
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T13:42:29Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-2 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.47593582887700536 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.47477744807121663 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.6753752084491385 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.868 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.43859649122807015 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4537037037037037 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9029682085279493 - name: F1 (macro) type: f1_macro value: 0.8981565352595764 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8276995305164319 - name: F1 (macro) type: f1_macro value: 0.6302262610207737 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.6522210184182016 - name: F1 (macro) type: f1_macro value: 0.6357658526272931 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9555540098768867 - name: F1 (macro) type: f1_macro value: 0.8819875185777288 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.872453776245691 - name: F1 (macro) type: f1_macro value: 0.868959677073872 --- # relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-2 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-2/raw/main/analogy.json)): - Accuracy on SAT (full): 0.47593582887700536 - Accuracy on SAT: 0.47477744807121663 - Accuracy on BATS: 0.6753752084491385 - Accuracy on U2: 0.43859649122807015 - Accuracy on U4: 0.4537037037037037 - Accuracy on Google: 0.868 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-2/raw/main/classification.json)): - Micro F1 score on BLESS: 0.9029682085279493 - Micro F1 score on CogALexV: 0.8276995305164319 - Micro F1 score on EVALution: 0.6522210184182016 - Micro F1 score on K&H+N: 0.9555540098768867 - Micro F1 score on ROOT09: 0.872453776245691 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-2/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-2") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: mask - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 9 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 2 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-2/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
research-backup/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-2
research-backup
2022-11-26T18:30:47Z
103
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T13:40:49Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-2 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3609625668449198 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.37091988130563797 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.6158977209560867 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.718 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.41228070175438597 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3958333333333333 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9059816182009944 - name: F1 (macro) type: f1_macro value: 0.9008198783874194 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8042253521126761 - name: F1 (macro) type: f1_macro value: 0.5554363820397257 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.6218851570964247 - name: F1 (macro) type: f1_macro value: 0.6041236250930277 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.966613340752591 - name: F1 (macro) type: f1_macro value: 0.8980274378539761 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8699467251645252 - name: F1 (macro) type: f1_macro value: 0.8674992783581462 --- # relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-2 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-2/raw/main/analogy.json)): - Accuracy on SAT (full): 0.3609625668449198 - Accuracy on SAT: 0.37091988130563797 - Accuracy on BATS: 0.6158977209560867 - Accuracy on U2: 0.41228070175438597 - Accuracy on U4: 0.3958333333333333 - Accuracy on Google: 0.718 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-2/raw/main/classification.json)): - Micro F1 score on BLESS: 0.9059816182009944 - Micro F1 score on CogALexV: 0.8042253521126761 - Micro F1 score on EVALution: 0.6218851570964247 - Micro F1 score on K&H+N: 0.966613340752591 - Micro F1 score on ROOT09: 0.8699467251645252 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-2/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-2") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: mask - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 9 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 2 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-2/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
research-backup/relbert-roberta-base-semeval2012-v6-mask-prompt-a-loob-2
research-backup
2022-11-26T18:24:00Z
103
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T13:37:38Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-a-loob-2 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.7953373015873015 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4197860962566845 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.41839762611275966 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.594774874930517 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.774 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.40789473684210525 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.41898148148148145 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9005574807895134 - name: F1 (macro) type: f1_macro value: 0.8957958532235768 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8077464788732395 - name: F1 (macro) type: f1_macro value: 0.5900936140399187 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.6359696641386782 - name: F1 (macro) type: f1_macro value: 0.6206497970461441 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9577797871600473 - name: F1 (macro) type: f1_macro value: 0.8748819835358477 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8633657160764651 - name: F1 (macro) type: f1_macro value: 0.8605769477843292 --- # relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-a-loob-2 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-a-loob-2/raw/main/analogy.json)): - Accuracy on SAT (full): 0.4197860962566845 - Accuracy on SAT: 0.41839762611275966 - Accuracy on BATS: 0.594774874930517 - Accuracy on U2: 0.40789473684210525 - Accuracy on U4: 0.41898148148148145 - Accuracy on Google: 0.774 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-a-loob-2/raw/main/classification.json)): - Micro F1 score on BLESS: 0.9005574807895134 - Micro F1 score on CogALexV: 0.8077464788732395 - Micro F1 score on EVALution: 0.6359696641386782 - Micro F1 score on K&H+N: 0.9577797871600473 - Micro F1 score on ROOT09: 0.8633657160764651 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-a-loob-2/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.7953373015873015 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-a-loob-2") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: mask - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 9 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 2 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-a-loob-2/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
fanpu/final_model_output_subreddit-wallstreetbets
fanpu
2022-11-26T18:22:28Z
115
0
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "generated_from_trainer", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2022-11-26T07:53:32Z
--- license: mit tags: - generated_from_trainer model-index: - name: final_model_output_subreddit-wallstreetbets results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # final_model_output_subreddit-wallstreetbets This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the None dataset. It achieves the following results on the evaluation set: - Loss: 3.5351 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 512 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 1000 - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 3.7979 | 1.25 | 5000 | 3.6293 | | 3.4998 | 2.49 | 10000 | 3.5351 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.13.0+cu117 - Datasets 2.7.1 - Tokenizers 0.13.2
research-backup/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-1
research-backup
2022-11-26T18:14:50Z
103
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T13:34:16Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-1 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.45989304812834225 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.45103857566765576 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.6381322957198443 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.854 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4649122807017544 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4513888888888889 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8866957962935063 - name: F1 (macro) type: f1_macro value: 0.8780169867685793 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8230046948356807 - name: F1 (macro) type: f1_macro value: 0.6127225487858271 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.6451787648970747 - name: F1 (macro) type: f1_macro value: 0.6389434572412788 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.952145788412047 - name: F1 (macro) type: f1_macro value: 0.8707933313910127 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8655593857724851 - name: F1 (macro) type: f1_macro value: 0.8633898253583613 --- # relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-1 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-1/raw/main/analogy.json)): - Accuracy on SAT (full): 0.45989304812834225 - Accuracy on SAT: 0.45103857566765576 - Accuracy on BATS: 0.6381322957198443 - Accuracy on U2: 0.4649122807017544 - Accuracy on U4: 0.4513888888888889 - Accuracy on Google: 0.854 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-1/raw/main/classification.json)): - Micro F1 score on BLESS: 0.8866957962935063 - Micro F1 score on CogALexV: 0.8230046948356807 - Micro F1 score on EVALution: 0.6451787648970747 - Micro F1 score on K&H+N: 0.952145788412047 - Micro F1 score on ROOT09: 0.8655593857724851 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-1/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-1") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: mask - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 9 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 1 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-d-loob-1/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
research-backup/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-1
research-backup
2022-11-26T18:11:24Z
102
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T13:32:34Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-1 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4144385026737968 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4065281899109792 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.6142301278488049 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.736 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.40789473684210525 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.37962962962962965 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8940786499924664 - name: F1 (macro) type: f1_macro value: 0.8886568208317418 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8112676056338028 - name: F1 (macro) type: f1_macro value: 0.5806627021345718 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.628385698808234 - name: F1 (macro) type: f1_macro value: 0.6195235287229023 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9612575641649858 - name: F1 (macro) type: f1_macro value: 0.8870509987826778 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8655593857724851 - name: F1 (macro) type: f1_macro value: 0.8633665155994553 --- # relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-1 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-1/raw/main/analogy.json)): - Accuracy on SAT (full): 0.4144385026737968 - Accuracy on SAT: 0.4065281899109792 - Accuracy on BATS: 0.6142301278488049 - Accuracy on U2: 0.40789473684210525 - Accuracy on U4: 0.37962962962962965 - Accuracy on Google: 0.736 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-1/raw/main/classification.json)): - Micro F1 score on BLESS: 0.8940786499924664 - Micro F1 score on CogALexV: 0.8112676056338028 - Micro F1 score on EVALution: 0.628385698808234 - Micro F1 score on K&H+N: 0.9612575641649858 - Micro F1 score on ROOT09: 0.8655593857724851 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-1/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-1") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: mask - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 9 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 1 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-1/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
research-backup/relbert-roberta-base-semeval2012-v6-mask-prompt-e-loob-0
research-backup
2022-11-26T18:01:13Z
103
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T13:27:34Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-e-loob-0 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3716577540106952 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.37388724035608306 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4324624791550862 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.72 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.35526315789473684 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3773148148148148 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.896790718698207 - name: F1 (macro) type: f1_macro value: 0.8911002385898117 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.795774647887324 - name: F1 (macro) type: f1_macro value: 0.5441132147176833 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.6235102925243771 - name: F1 (macro) type: f1_macro value: 0.6166467084580077 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9611184530847882 - name: F1 (macro) type: f1_macro value: 0.8888486397674304 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8617988091507365 - name: F1 (macro) type: f1_macro value: 0.8588467057047432 --- # relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-e-loob-0 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-e-loob-0/raw/main/analogy.json)): - Accuracy on SAT (full): 0.3716577540106952 - Accuracy on SAT: 0.37388724035608306 - Accuracy on BATS: 0.4324624791550862 - Accuracy on U2: 0.35526315789473684 - Accuracy on U4: 0.3773148148148148 - Accuracy on Google: 0.72 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-e-loob-0/raw/main/classification.json)): - Micro F1 score on BLESS: 0.896790718698207 - Micro F1 score on CogALexV: 0.795774647887324 - Micro F1 score on EVALution: 0.6235102925243771 - Micro F1 score on K&H+N: 0.9611184530847882 - Micro F1 score on ROOT09: 0.8617988091507365 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-e-loob-0/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-e-loob-0") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: mask - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 6 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 0 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-e-loob-0/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
unza/xls-r-300m-nyanja-model_v1
unza
2022-11-26T17:57:15Z
163
0
transformers
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "NyanjaSpeech", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-11-26T15:26:44Z
--- license: apache-2.0 tags: - automatic-speech-recognition - NyanjaSpeech - generated_from_trainer metrics: - wer model-index: - name: xls-r-300m-nyanja-model_v1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xls-r-300m-nyanja-model_v1 This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the NYANJASPEECH - NYA dataset. It achieves the following results on the evaluation set: - Loss: 0.2772 - Wer: 0.9074 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 5.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.7585 | 1.58 | 500 | 0.3574 | 0.9679 | | 0.4736 | 3.16 | 1000 | 0.2772 | 0.9074 | | 0.4776 | 4.75 | 1500 | 0.2853 | 0.9578 | ### Framework versions - Transformers 4.25.0.dev0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
research-backup/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-0
research-backup
2022-11-26T17:54:30Z
102
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T13:24:17Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-0 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4144385026737968 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.42136498516320475 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.6214563646470261 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.82 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.38596491228070173 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4398148148148148 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9068856411029079 - name: F1 (macro) type: f1_macro value: 0.9011104376125445 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8016431924882629 - name: F1 (macro) type: f1_macro value: 0.5650565105468204 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.6495124593716143 - name: F1 (macro) type: f1_macro value: 0.6283507845337405 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9609097864644919 - name: F1 (macro) type: f1_macro value: 0.8939368966008155 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8602319022250078 - name: F1 (macro) type: f1_macro value: 0.8561743696081581 --- # relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-0 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-0/raw/main/analogy.json)): - Accuracy on SAT (full): 0.4144385026737968 - Accuracy on SAT: 0.42136498516320475 - Accuracy on BATS: 0.6214563646470261 - Accuracy on U2: 0.38596491228070173 - Accuracy on U4: 0.4398148148148148 - Accuracy on Google: 0.82 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-0/raw/main/classification.json)): - Micro F1 score on BLESS: 0.9068856411029079 - Micro F1 score on CogALexV: 0.8016431924882629 - Micro F1 score on EVALution: 0.6495124593716143 - Micro F1 score on K&H+N: 0.9609097864644919 - Micro F1 score on ROOT09: 0.8602319022250078 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-0/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-0") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: mask - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 8 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 0 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-c-loob-0/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
research-backup/relbert-roberta-base-semeval2012-v6-mask-prompt-b-loob-0
research-backup
2022-11-26T17:51:04Z
97
0
transformers
[ "transformers", "pytorch", "roberta", "feature-extraction", "dataset:relbert/semeval2012_relational_similarity_v6", "model-index", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2022-11-26T13:22:41Z
--- datasets: - relbert/semeval2012_relational_similarity_v6 model-index: - name: relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-b-loob-0 results: - task: name: Relation Mapping type: sorting-task dataset: name: Relation Mapping args: relbert/relation_mapping type: relation-mapping metrics: - name: Accuracy type: accuracy value: 0.6438492063492064 - task: name: Analogy Questions (SAT full) type: multiple-choice-qa dataset: name: SAT full args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.37967914438502676 - task: name: Analogy Questions (SAT) type: multiple-choice-qa dataset: name: SAT args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.3827893175074184 - task: name: Analogy Questions (BATS) type: multiple-choice-qa dataset: name: BATS args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.5519733185102835 - task: name: Analogy Questions (Google) type: multiple-choice-qa dataset: name: Google args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.708 - task: name: Analogy Questions (U2) type: multiple-choice-qa dataset: name: U2 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.37280701754385964 - task: name: Analogy Questions (U4) type: multiple-choice-qa dataset: name: U4 args: relbert/analogy_questions type: analogy-questions metrics: - name: Accuracy type: accuracy value: 0.4097222222222222 - task: name: Lexical Relation Classification (BLESS) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8978454120837728 - name: F1 (macro) type: f1_macro value: 0.8949175384097002 - task: name: Lexical Relation Classification (CogALexV) type: classification dataset: name: CogALexV args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8124413145539906 - name: F1 (macro) type: f1_macro value: 0.5766409255039524 - task: name: Lexical Relation Classification (EVALution) type: classification dataset: name: BLESS args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.6484290357529794 - name: F1 (macro) type: f1_macro value: 0.6301872449931128 - task: name: Lexical Relation Classification (K&H+N) type: classification dataset: name: K&H+N args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.9574320094595534 - name: F1 (macro) type: f1_macro value: 0.8795354417385828 - task: name: Lexical Relation Classification (ROOT09) type: classification dataset: name: ROOT09 args: relbert/lexical_relation_classification type: relation-classification metrics: - name: F1 type: f1 value: 0.8784080225634597 - name: F1 (macro) type: f1_macro value: 0.8744337880528867 --- # relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-b-loob-0 RelBERT fine-tuned from [roberta-base](https://huggingface.co/roberta-base) on [relbert/semeval2012_relational_similarity_v6](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity_v6). Fine-tuning is done via [RelBERT](https://github.com/asahi417/relbert) library (see the repository for more detail). It achieves the following results on the relation understanding tasks: - Analogy Question ([dataset](https://huggingface.co/datasets/relbert/analogy_questions), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-b-loob-0/raw/main/analogy.json)): - Accuracy on SAT (full): 0.37967914438502676 - Accuracy on SAT: 0.3827893175074184 - Accuracy on BATS: 0.5519733185102835 - Accuracy on U2: 0.37280701754385964 - Accuracy on U4: 0.4097222222222222 - Accuracy on Google: 0.708 - Lexical Relation Classification ([dataset](https://huggingface.co/datasets/relbert/lexical_relation_classification), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-b-loob-0/raw/main/classification.json)): - Micro F1 score on BLESS: 0.8978454120837728 - Micro F1 score on CogALexV: 0.8124413145539906 - Micro F1 score on EVALution: 0.6484290357529794 - Micro F1 score on K&H+N: 0.9574320094595534 - Micro F1 score on ROOT09: 0.8784080225634597 - Relation Mapping ([dataset](https://huggingface.co/datasets/relbert/relation_mapping), [full result](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-b-loob-0/raw/main/relation_mapping.json)): - Accuracy on Relation Mapping: 0.6438492063492064 ### Usage This model can be used through the [relbert library](https://github.com/asahi417/relbert). Install the library via pip ```shell pip install relbert ``` and activate model as below. ```python from relbert import RelBERT model = RelBERT("relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-b-loob-0") vector = model.get_embedding(['Tokyo', 'Japan']) # shape of (1024, ) ``` ### Training hyperparameters The following hyperparameters were used during training: - model: roberta-base - max_length: 64 - mode: mask - data: relbert/semeval2012_relational_similarity_v6 - split: train - split_eval: validation - template_mode: manual - loss_function: info_loob - classification_loss: False - temperature_nce_constant: 0.05 - temperature_nce_rank: {'min': 0.01, 'max': 0.05, 'type': 'linear'} - epoch: 6 - batch: 128 - lr: 5e-06 - lr_decay: False - lr_warmup: 1 - weight_decay: 0 - random_seed: 0 - exclude_relation: None - n_sample: 320 - gradient_accumulation: 8 - relation_level: None The full configuration can be found at [fine-tuning parameter file](https://huggingface.co/relbert/relbert-roberta-base-semeval2012-v6-mask-prompt-b-loob-0/raw/main/trainer_config.json). ### Reference If you use any resource from RelBERT, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` @inproceedings{ushio-etal-2021-distilling-relation-embeddings, title = "{D}istilling {R}elation {E}mbeddings from {P}re-trained {L}anguage {M}odels", author = "Ushio, Asahi and Schockaert, Steven and Camacho-Collados, Jose", booktitle = "EMNLP 2021", year = "2021", address = "Online", publisher = "Association for Computational Linguistics", } ```
huggingtweets/davidhornik
huggingtweets
2022-11-26T16:47:20Z
115
0
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2022-11-26T16:44:53Z
--- language: en thumbnail: http://www.huggingtweets.com/davidhornik/1669481236541/predictions.png tags: - huggingtweets widget: - text: "My dream is" --- <div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1511347110041591818/qMznehUo_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">David Hornik</div> <div style="text-align: center; font-size: 14px;">@davidhornik</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from David Hornik. | Data | David Hornik | | --- | --- | | Tweets downloaded | 3176 | | Retweets | 490 | | Short tweets | 262 | | Tweets kept | 2424 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/33354cwl/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @davidhornik's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1ok2hwr0) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1ok2hwr0/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/davidhornik') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
IsaacRodgz/setfit-stance-prediction-spanish-news-headlines
IsaacRodgz
2022-11-26T15:58:43Z
3
0
sentence-transformers
[ "sentence-transformers", "pytorch", "onnx", "bert", "feature-extraction", "sentence-similarity", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
2022-11-17T01:55:08Z
--- pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity --- # {MODEL_NAME} This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search. <!--- Describe your model here --> ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer sentences = ["This is an example sentence", "Each sentence is converted"] model = SentenceTransformer('{MODEL_NAME}') embeddings = model.encode(sentences) print(embeddings) ``` ## Evaluation Results <!--- Describe how your model was evaluated --> For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME}) ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 170 with parameters: ``` {'batch_size': 16, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss` Parameters of the fit()-Method: ``` { "epochs": 1, "evaluation_steps": 0, "evaluator": "NoneType", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": 170, "warmup_steps": 17, "weight_decay": 0.01 } ``` ## Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 256, 'do_lower_case': False}) with Transformer model: BertModel (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False}) (2): Normalize() ) ``` ## Citing & Authors <!--- Describe where people can find more information -->
multimodalart/sd-fine-tunable
multimodalart
2022-11-26T14:17:28Z
59
6
diffusers
[ "diffusers", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2022-11-17T00:51:44Z
--- license: creativeml-openrail-m --- Stable Diffusion v1-5 with the fine-tuned VAE `sd-vae-ft-mse` and files with config modifications for making it better to fine-tune made by [fast-stable-diffusion by TheLastBen](https://github.com/TheLastBen/fast-stable-diffusion) to be used on [fastDreambooth Colab Notebook](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) and on the [Dreambooth Training Space](https://huggingface.co/spaces/multimodalart/dreambooth-training) Is not suited for inference and training elsewhere is under your own risk. The [model LICENSE](https://huggingface.co/spaces/CompVis/stable-diffusion-license) still applies normally for this use-case. Refer to the [original repository](https://huggingface.co/runwayml/stable-diffusion-v1-5) for the model card
justinian336/bert-salvadoran-news
justinian336
2022-11-26T13:48:52Z
70
0
transformers
[ "transformers", "tf", "distilbert", "fill-mask", "generated_from_keras_callback", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-11-26T13:42:42Z
--- license: apache-2.0 tags: - generated_from_keras_callback model-index: - name: bert-salvadoran-news results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # bert-salvadoran-news This model is a fine-tuned version of [distilbert-base-multilingual-cased](https://huggingface.co/distilbert-base-multilingual-cased) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 1.4210 - Validation Loss: 1.5456 - Epoch: 19 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 2e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01} - training_precision: float32 ### Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 2.1106 | 1.9331 | 0 | | 1.9666 | 1.7924 | 1 | | 1.8643 | 1.7777 | 2 | | 1.8090 | 1.7018 | 3 | | 1.7668 | 1.7372 | 4 | | 1.7337 | 1.6726 | 5 | | 1.6948 | 1.6624 | 6 | | 1.6584 | 1.6863 | 7 | | 1.6232 | 1.6174 | 8 | | 1.5996 | 1.6399 | 9 | | 1.5805 | 1.5340 | 10 | | 1.5622 | 1.6202 | 11 | | 1.5124 | 1.6176 | 12 | | 1.5150 | 1.5644 | 13 | | 1.5151 | 1.6002 | 14 | | 1.4730 | 1.5746 | 15 | | 1.4538 | 1.5536 | 16 | | 1.4507 | 1.5537 | 17 | | 1.4476 | 1.5450 | 18 | | 1.4210 | 1.5456 | 19 | ### Framework versions - Transformers 4.24.0 - TensorFlow 2.9.2 - Datasets 2.7.1 - Tokenizers 0.13.2
ThetaPhiPsi/distilbert-base-uncased-finetuned-imdb-finetuned-imdb
ThetaPhiPsi
2022-11-26T13:04:18Z
180
0
transformers
[ "transformers", "pytorch", "distilbert", "fill-mask", "generated_from_trainer", "dataset:imdb", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-11-26T12:09:17Z
--- tags: - generated_from_trainer datasets: - imdb model-index: - name: distilbert-base-uncased-finetuned-imdb-finetuned-imdb results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-imdb-finetuned-imdb This model was trained from scratch on the imdb dataset. It achieves the following results on the evaluation set: - eval_loss: 2.1896 - eval_runtime: 3.5892 - eval_samples_per_second: 1707.604 - eval_steps_per_second: 26.747 - epoch: 1.74 - step: 1500 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.24.0 - Pytorch 1.10.0 - Datasets 2.7.1 - Tokenizers 0.13.2
Suyash1080/model_index
Suyash1080
2022-11-26T12:53:37Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2022-11-26T12:53:37Z
--- license: creativeml-openrail-m ---
Aunsiels/ChildBERT
Aunsiels
2022-11-26T11:42:15Z
117
2
transformers
[ "transformers", "pytorch", "bert", "fill-mask", "children", "infant", "en", "dataset:Aunsiels/InfantBooks", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-10-19T08:54:08Z
--- language: - en tags: - children - infant datasets: - Aunsiels/InfantBooks --- A BERT-model finetuned on children's books. ``` Romero, J., & Razniewski, S. (2022). Do Children Texts Hold The Key To Commonsense Knowledge? In Proceedings of the 2022 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning. ```
monistar/swin-tiny-patch4-window7-224-finetuned-eurosat
monistar
2022-11-26T11:05:51Z
220
0
transformers
[ "transformers", "pytorch", "tensorboard", "swin", "image-classification", "generated_from_trainer", "dataset:imagefolder", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2022-10-11T11:04:45Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imagefolder metrics: - accuracy model-index: - name: swin-tiny-patch4-window7-224-finetuned-eurosat results: - task: name: Image Classification type: image-classification dataset: name: imagefolder type: imagefolder config: default split: train args: default metrics: - name: Accuracy type: accuracy value: 0.9814814814814815 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # swin-tiny-patch4-window7-224-finetuned-eurosat This model is a fine-tuned version of [microsoft/swin-tiny-patch4-window7-224](https://huggingface.co/microsoft/swin-tiny-patch4-window7-224) on the imagefolder dataset. It achieves the following results on the evaluation set: - Loss: 0.0525 - Accuracy: 0.9815 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.2396 | 1.0 | 190 | 0.1071 | 0.9656 | | 0.1605 | 2.0 | 380 | 0.0665 | 0.9793 | | 0.1282 | 3.0 | 570 | 0.0525 | 0.9815 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
CareerNinja/BERT_2_Labels
CareerNinja
2022-11-26T10:35:00Z
162
0
transformers
[ "transformers", "pytorch", "bert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-11-26T09:56:41Z
--- license: apache-2.0 --- Number of Epochs = 5 <br> Dataset Size = 5.5 k samples [train/validation] <br> Number of labels used = 2 <br> Thresholding = True<br> Thresholding value = 0.7<br> Below is the function to aplly thresholding to output logits. ```python def get_prediction(text): encoding = tokenizer(text, return_tensors="pt", padding="max_length", truncation=True, max_length=128) encoding = {k: v.to(trainer.model.device) for k,v in encoding.items()} outputs = model(**encoding) logits = outputs.logits sigmoid = torch.nn.Sigmoid() probs = sigmoid(logits.squeeze().cpu()) probs = probs.detach().numpy() label = np.argmax(probs, axis=-1) if label == 1: if probs[1] > 0.7: return 1 else: return 0 else: return 0 ```
kejian/final-mle-again
kejian
2022-11-26T10:32:30Z
4
0
transformers
[ "transformers", "pytorch", "gpt2", "generated_from_trainer", "en", "dataset:kejian/codeparrot-train-more-filter-3.3b-cleaned", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
2022-11-25T18:21:47Z
--- language: - en license: apache-2.0 tags: - generated_from_trainer datasets: - kejian/codeparrot-train-more-filter-3.3b-cleaned model-index: - name: kejian/final-mle-again results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # kejian/final-mle-again This model was trained from scratch on the kejian/codeparrot-train-more-filter-3.3b-cleaned dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0008 - train_batch_size: 64 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.01 - training_steps: 50354 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.23.0 - Pytorch 1.13.0+cu116 - Datasets 2.0.0 - Tokenizers 0.12.1 # Full config {'dataset': {'datasets': ['kejian/codeparrot-train-more-filter-3.3b-cleaned'], 'is_split_by_sentences': True}, 'generation': {'batch_size': 128, 'metrics_configs': [{}, {'n': 1}, {}], 'scenario_configs': [{'display_as_html': True, 'generate_kwargs': {'do_sample': True, 'eos_token_id': 0, 'max_length': 640, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'unconditional', 'num_samples': 512}, {'display_as_html': True, 'generate_kwargs': {'do_sample': True, 'eos_token_id': 0, 'max_length': 272, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'functions', 'num_samples': 512, 'prompts_path': 'resources/functions_csnet.jsonl', 'use_prompt_for_scoring': True}], 'scorer_config': {}}, 'kl_gpt3_callback': {'gpt3_kwargs': {'model_name': 'code-cushman-001'}, 'max_tokens': 64, 'num_samples': 4096}, 'model': {'from_scratch': True, 'gpt2_config_kwargs': {'reorder_and_upcast_attn': True, 'scale_attn_by': True}, 'path_or_name': 'codeparrot/codeparrot-small'}, 'objective': {'name': 'MLE'}, 'tokenizer': {'path_or_name': 'codeparrot/codeparrot-small'}, 'training': {'dataloader_num_workers': 0, 'effective_batch_size': 64, 'evaluation_strategy': 'no', 'fp16': True, 'hub_model_id': 'kejian/final-mle-again', 'hub_strategy': 'all_checkpoints', 'learning_rate': 0.0008, 'logging_first_step': True, 'logging_steps': 1, 'num_tokens': 3300000000.0, 'output_dir': 'training_output', 'per_device_train_batch_size': 16, 'push_to_hub': True, 'remove_unused_columns': False, 'save_steps': 5000, 'save_strategy': 'steps', 'seed': 42, 'warmup_ratio': 0.01, 'weight_decay': 0.1}} # Wandb URL: https://wandb.ai/kejian/uncategorized/runs/u7vbiehz
fanzru/distilbart-cnn-6-6-finetuned-xsum-intro-test
fanzru
2022-11-26T10:14:53Z
7
0
transformers
[ "transformers", "pytorch", "tensorboard", "bart", "text2text-generation", "generated_from_trainer", "dataset:xsum", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-26T08:48:38Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - xsum metrics: - rouge model-index: - name: distilbart-cnn-6-6-finetuned-xsum-intro-test results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: xsum type: xsum config: default split: train args: default metrics: - name: Rouge1 type: rouge value: 32.0474 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbart-cnn-6-6-finetuned-xsum-intro-test This model is a fine-tuned version of [sshleifer/distilbart-cnn-6-6](https://huggingface.co/sshleifer/distilbart-cnn-6-6) on the xsum dataset. It achieves the following results on the evaluation set: - Loss: 1.9036 - Rouge1: 32.0474 - Rouge2: 12.3779 - Rougel: 23.5491 - Rougelsum: 24.251 - Gen Len: 60.8594 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:| | 1.9432 | 1.0 | 12753 | 1.9036 | 32.0474 | 12.3779 | 23.5491 | 24.251 | 60.8594 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.13.0+cu117 - Datasets 2.7.1 - Tokenizers 0.13.2
Viktor111/super
Viktor111
2022-11-26T07:49:35Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2022-11-26T07:49:35Z
--- license: creativeml-openrail-m ---
TTian/deberta-classifier-feedback-1024-pseudo-final
TTian
2022-11-26T05:53:27Z
10
0
transformers
[ "transformers", "pytorch", "tensorboard", "deberta-v2", "token-classification", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-11-26T03:30:29Z
--- license: mit tags: - generated_from_trainer model-index: - name: deberta-classifier-feedback-1024-pseudo-final results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-classifier-feedback-1024-pseudo-final This model is a fine-tuned version of [TTian/deberta-classifier-feedback-1024-pseudo](https://huggingface.co/TTian/deberta-classifier-feedback-1024-pseudo) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.5263 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.5814 | 0.04 | 10 | 0.5888 | | 0.5521 | 0.08 | 20 | 0.5736 | | 0.5685 | 0.13 | 30 | 0.5809 | | 0.6052 | 0.17 | 40 | 0.5702 | | 0.5532 | 0.21 | 50 | 0.5571 | | 0.6177 | 0.25 | 60 | 0.5848 | | 0.6196 | 0.3 | 70 | 0.5464 | | 0.5772 | 0.34 | 80 | 0.5307 | | 0.5805 | 0.38 | 90 | 0.5550 | | 0.6453 | 0.42 | 100 | 0.5467 | | 0.5756 | 0.47 | 110 | 0.5587 | | 0.5901 | 0.51 | 120 | 0.5482 | | 0.568 | 0.55 | 130 | 0.5263 | | 0.5452 | 0.59 | 140 | 0.5698 | | 0.5949 | 0.64 | 150 | 0.5484 | | 0.5537 | 0.68 | 160 | 0.5783 | | 0.5327 | 0.72 | 170 | 0.5202 | | 0.5449 | 0.76 | 180 | 0.5272 | | 0.5345 | 0.81 | 190 | 0.5621 | | 0.5837 | 0.85 | 200 | 0.5501 | | 0.5969 | 0.89 | 210 | 0.5470 | | 0.5905 | 0.93 | 220 | 0.5924 | | 0.5481 | 0.97 | 230 | 0.5415 | | 0.5035 | 1.02 | 240 | 0.5321 | | 0.4508 | 1.06 | 250 | 0.5371 | | 0.4227 | 1.1 | 260 | 0.5276 | | 0.4423 | 1.14 | 270 | 0.5324 | | 0.432 | 1.19 | 280 | 0.5378 | | 0.4317 | 1.23 | 290 | 0.5302 | | 0.46 | 1.27 | 300 | 0.5302 | | 0.435 | 1.31 | 310 | 0.5326 | | 0.3813 | 1.36 | 320 | 0.5431 | | 0.4422 | 1.4 | 330 | 0.5323 | | 0.4298 | 1.44 | 340 | 0.5575 | | 0.5068 | 1.48 | 350 | 0.5529 | | 0.4619 | 1.53 | 360 | 0.5589 | | 0.4852 | 1.57 | 370 | 0.5256 | | 0.3888 | 1.61 | 380 | 0.5731 | | 0.4319 | 1.65 | 390 | 0.5335 | | 0.4422 | 1.69 | 400 | 0.5419 | | 0.4522 | 1.74 | 410 | 0.5547 | | 0.4276 | 1.78 | 420 | 0.5263 | | 0.3988 | 1.82 | 430 | 0.5481 | | 0.4063 | 1.86 | 440 | 0.5404 | | 0.4141 | 1.91 | 450 | 0.5292 | | 0.4149 | 1.95 | 460 | 0.5241 | | 0.4104 | 1.99 | 470 | 0.5263 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
Deigant/t5-base-finetuned-qg-context-dataset-2
Deigant
2022-11-26T04:59:04Z
105
0
transformers
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-24T06:53:16Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - rouge model-index: - name: t5-base-finetuned-qg-context-dataset-2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-base-finetuned-qg-context-dataset-2 This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.5958 - Rouge1: 37.1698 - Rouge2: 15.8177 - Rougel: 33.3329 - Rougelsum: 33.0872 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 16 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:| | No log | 1.0 | 73 | 1.8700 | 30.0504 | 9.8379 | 24.0514 | 24.1584 | | No log | 2.0 | 146 | 1.6264 | 34.715 | 12.71 | 27.825 | 27.7697 | | No log | 3.0 | 219 | 1.5904 | 33.9221 | 10.688 | 29.5483 | 29.5388 | | No log | 4.0 | 292 | 1.5623 | 36.5544 | 14.8003 | 31.6295 | 31.4603 | | No log | 5.0 | 365 | 1.5463 | 34.071 | 13.189 | 30.1517 | 30.3325 | | No log | 6.0 | 438 | 1.5539 | 37.7324 | 15.5312 | 33.3968 | 33.2518 | | 1.5099 | 7.0 | 511 | 1.5643 | 32.5168 | 11.2479 | 27.4951 | 27.4425 | | 1.5099 | 8.0 | 584 | 1.5653 | 39.5646 | 17.9528 | 35.4095 | 35.2042 | | 1.5099 | 9.0 | 657 | 1.5679 | 39.333 | 17.0059 | 34.9131 | 34.7696 | | 1.5099 | 10.0 | 730 | 1.5757 | 37.5046 | 16.2468 | 32.5031 | 32.4012 | | 1.5099 | 11.0 | 803 | 1.5738 | 37.601 | 16.4592 | 33.5804 | 33.1352 | | 1.5099 | 12.0 | 876 | 1.5894 | 42.1889 | 19.3169 | 37.8273 | 37.7312 | | 1.5099 | 13.0 | 949 | 1.5929 | 38.5814 | 17.0896 | 34.4696 | 34.3629 | | 1.015 | 14.0 | 1022 | 1.5922 | 36.6392 | 16.8083 | 32.6318 | 32.4199 | | 1.015 | 15.0 | 1095 | 1.5948 | 34.6707 | 15.7198 | 30.319 | 30.3403 | | 1.015 | 16.0 | 1168 | 1.5958 | 37.1698 | 15.8177 | 33.3329 | 33.0872 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
elRivx/gLWoman
elRivx
2022-11-26T04:51:07Z
0
2
null
[ "stable-diffusion", "text-to-image", "license:creativeml-openrail-m", "region:us" ]
text-to-image
2022-11-26T04:38:49Z
--- license: creativeml-openrail-m tags: - stable-diffusion - text-to-image --- # gLWoman This is my second Stable Diffusion custom model that bring to you a generic woman generated with non-licenced images. The magic word is: gLWoman If you enjoy my work, please consider supporting me: [![Buy me a coffee](https://badgen.net/badge/icon/buymeacoffee?icon=buymeacoffee&label)](https://www.buymeacoffee.com/elrivx) Examples: <img src=https://imgur.com/tKpiVEE.png width=30% height=30%> <img src=https://imgur.com/GAOJzps.png width=30% height=30%> <img src=https://imgur.com/oxI9ZQv.png width=30% height=30%> ## License This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage. The CreativeML OpenRAIL License specifies: 1. You can't use the model to deliberately produce nor share illegal or harmful outputs or content 2. The authors claims no rights on the outputs you generate, you are free to use them and are accountable for their use which must not go against the provisions set in the license 3. You may re-distribute the weights and use the model commercially and/or as a service. If you do, please be aware you have to include the same use restrictions as the ones in the license and share a copy of the CreativeML OpenRAIL-M to all your users (please read the license entirely and carefully) [Please read the full license here](https://huggingface.co/spaces/CompVis/stable-diffusion-license)
shreyasharma/sentence_eval1
shreyasharma
2022-11-26T03:38:40Z
106
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-11-26T02:16:56Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: sentence_eval1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # sentence_eval1 This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 3.2129 - Accuracy: 0.5743 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 361 | 0.1084 | 0.9657 | | 0.0982 | 2.0 | 722 | 0.0993 | 0.9816 | | 0.0195 | 3.0 | 1083 | 0.0928 | 0.9859 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
CielTempest12335/Da
CielTempest12335
2022-11-26T02:12:57Z
0
0
null
[ "license:bigscience-openrail-m", "region:us" ]
null
2022-11-26T02:12:57Z
--- license: bigscience-openrail-m ---
SiddharthaM/hasoc19-xlm-roberta-base-targinsult1
SiddharthaM
2022-11-26T01:30:36Z
124
0
transformers
[ "transformers", "pytorch", "tensorboard", "xlm-roberta", "text-classification", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-11-25T23:33:09Z
--- license: mit tags: - generated_from_trainer metrics: - accuracy - precision - recall - f1 model-index: - name: hasoc19-xlm-roberta-base-targinsult1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # hasoc19-xlm-roberta-base-targinsult1 This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7512 - Accuracy: 0.7096 - Precision: 0.6720 - Recall: 0.6675 - F1: 0.6695 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:| | No log | 1.0 | 263 | 0.5619 | 0.6996 | 0.6660 | 0.6717 | 0.6684 | | 0.5931 | 2.0 | 526 | 0.5350 | 0.7239 | 0.6880 | 0.6576 | 0.6655 | | 0.5931 | 3.0 | 789 | 0.5438 | 0.7239 | 0.6872 | 0.6644 | 0.6714 | | 0.5101 | 4.0 | 1052 | 0.5595 | 0.7196 | 0.6866 | 0.6909 | 0.6886 | | 0.5101 | 5.0 | 1315 | 0.5580 | 0.7186 | 0.6818 | 0.6743 | 0.6774 | | 0.4313 | 6.0 | 1578 | 0.6000 | 0.7039 | 0.6679 | 0.6692 | 0.6686 | | 0.4313 | 7.0 | 1841 | 0.6429 | 0.7082 | 0.6765 | 0.6841 | 0.6794 | | 0.3591 | 8.0 | 2104 | 0.6626 | 0.7115 | 0.6772 | 0.6803 | 0.6786 | | 0.3591 | 9.0 | 2367 | 0.7231 | 0.7139 | 0.6764 | 0.6700 | 0.6727 | | 0.3016 | 10.0 | 2630 | 0.7512 | 0.7096 | 0.6720 | 0.6675 | 0.6695 | ### Framework versions - Transformers 4.24.0.dev0 - Pytorch 1.11.0+cu102 - Datasets 2.6.1 - Tokenizers 0.13.1
asdff8fa7/q-FrozenLake-v1-4x4-noSlippery
asdff8fa7
2022-11-26T00:31:33Z
0
0
null
[ "FrozenLake-v1", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2022-11-25T06:49:12Z
--- tags: - FrozenLake-v1 - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-FrozenLake-v1-4x4-noSlippery results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: FrozenLake-v1 type: FrozenLake-v1 metrics: - type: mean_reward value: 8.12 +/- 2.72 name: mean_reward verified: false --- # **Q-Learning** Agent playing **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="asdff8fa7/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) evaluate_agent(env, model["max_steps"], model["n_eval_episodes"], model["qtable"], model["eval_seed"]) ```
tomekkorbak/wonderful_keller
tomekkorbak
2022-11-26T00:19:09Z
103
0
transformers
[ "transformers", "pytorch", "gpt2", "generated_from_trainer", "en", "dataset:tomekkorbak/detoxify-pile-chunk3-0-50000", "dataset:tomekkorbak/detoxify-pile-chunk3-50000-100000", "dataset:tomekkorbak/detoxify-pile-chunk3-100000-150000", "dataset:tomekkorbak/detoxify-pile-chunk3-150000-200000", "dataset:tomekkorbak/detoxify-pile-chunk3-200000-250000", "dataset:tomekkorbak/detoxify-pile-chunk3-250000-300000", "dataset:tomekkorbak/detoxify-pile-chunk3-300000-350000", "dataset:tomekkorbak/detoxify-pile-chunk3-350000-400000", "dataset:tomekkorbak/detoxify-pile-chunk3-400000-450000", "dataset:tomekkorbak/detoxify-pile-chunk3-450000-500000", "dataset:tomekkorbak/detoxify-pile-chunk3-500000-550000", "dataset:tomekkorbak/detoxify-pile-chunk3-550000-600000", "dataset:tomekkorbak/detoxify-pile-chunk3-600000-650000", "dataset:tomekkorbak/detoxify-pile-chunk3-650000-700000", "dataset:tomekkorbak/detoxify-pile-chunk3-700000-750000", "dataset:tomekkorbak/detoxify-pile-chunk3-750000-800000", "dataset:tomekkorbak/detoxify-pile-chunk3-800000-850000", "dataset:tomekkorbak/detoxify-pile-chunk3-850000-900000", "dataset:tomekkorbak/detoxify-pile-chunk3-900000-950000", "dataset:tomekkorbak/detoxify-pile-chunk3-950000-1000000", "dataset:tomekkorbak/detoxify-pile-chunk3-1000000-1050000", "dataset:tomekkorbak/detoxify-pile-chunk3-1050000-1100000", "dataset:tomekkorbak/detoxify-pile-chunk3-1100000-1150000", "dataset:tomekkorbak/detoxify-pile-chunk3-1150000-1200000", "dataset:tomekkorbak/detoxify-pile-chunk3-1200000-1250000", "dataset:tomekkorbak/detoxify-pile-chunk3-1250000-1300000", "dataset:tomekkorbak/detoxify-pile-chunk3-1300000-1350000", "dataset:tomekkorbak/detoxify-pile-chunk3-1350000-1400000", "dataset:tomekkorbak/detoxify-pile-chunk3-1400000-1450000", "dataset:tomekkorbak/detoxify-pile-chunk3-1450000-1500000", "dataset:tomekkorbak/detoxify-pile-chunk3-1500000-1550000", "dataset:tomekkorbak/detoxify-pile-chunk3-1550000-1600000", "dataset:tomekkorbak/detoxify-pile-chunk3-1600000-1650000", "dataset:tomekkorbak/detoxify-pile-chunk3-1650000-1700000", "dataset:tomekkorbak/detoxify-pile-chunk3-1700000-1750000", "dataset:tomekkorbak/detoxify-pile-chunk3-1750000-1800000", "dataset:tomekkorbak/detoxify-pile-chunk3-1800000-1850000", "dataset:tomekkorbak/detoxify-pile-chunk3-1850000-1900000", "dataset:tomekkorbak/detoxify-pile-chunk3-1900000-1950000", "license:mit", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
2022-11-25T10:18:57Z
--- language: - en license: mit tags: - generated_from_trainer datasets: - tomekkorbak/detoxify-pile-chunk3-0-50000 - tomekkorbak/detoxify-pile-chunk3-50000-100000 - tomekkorbak/detoxify-pile-chunk3-100000-150000 - tomekkorbak/detoxify-pile-chunk3-150000-200000 - tomekkorbak/detoxify-pile-chunk3-200000-250000 - tomekkorbak/detoxify-pile-chunk3-250000-300000 - tomekkorbak/detoxify-pile-chunk3-300000-350000 - tomekkorbak/detoxify-pile-chunk3-350000-400000 - tomekkorbak/detoxify-pile-chunk3-400000-450000 - tomekkorbak/detoxify-pile-chunk3-450000-500000 - tomekkorbak/detoxify-pile-chunk3-500000-550000 - tomekkorbak/detoxify-pile-chunk3-550000-600000 - tomekkorbak/detoxify-pile-chunk3-600000-650000 - tomekkorbak/detoxify-pile-chunk3-650000-700000 - tomekkorbak/detoxify-pile-chunk3-700000-750000 - tomekkorbak/detoxify-pile-chunk3-750000-800000 - tomekkorbak/detoxify-pile-chunk3-800000-850000 - tomekkorbak/detoxify-pile-chunk3-850000-900000 - tomekkorbak/detoxify-pile-chunk3-900000-950000 - tomekkorbak/detoxify-pile-chunk3-950000-1000000 - tomekkorbak/detoxify-pile-chunk3-1000000-1050000 - tomekkorbak/detoxify-pile-chunk3-1050000-1100000 - tomekkorbak/detoxify-pile-chunk3-1100000-1150000 - tomekkorbak/detoxify-pile-chunk3-1150000-1200000 - tomekkorbak/detoxify-pile-chunk3-1200000-1250000 - tomekkorbak/detoxify-pile-chunk3-1250000-1300000 - tomekkorbak/detoxify-pile-chunk3-1300000-1350000 - tomekkorbak/detoxify-pile-chunk3-1350000-1400000 - tomekkorbak/detoxify-pile-chunk3-1400000-1450000 - tomekkorbak/detoxify-pile-chunk3-1450000-1500000 - tomekkorbak/detoxify-pile-chunk3-1500000-1550000 - tomekkorbak/detoxify-pile-chunk3-1550000-1600000 - tomekkorbak/detoxify-pile-chunk3-1600000-1650000 - tomekkorbak/detoxify-pile-chunk3-1650000-1700000 - tomekkorbak/detoxify-pile-chunk3-1700000-1750000 - tomekkorbak/detoxify-pile-chunk3-1750000-1800000 - tomekkorbak/detoxify-pile-chunk3-1800000-1850000 - tomekkorbak/detoxify-pile-chunk3-1850000-1900000 - tomekkorbak/detoxify-pile-chunk3-1900000-1950000 model-index: - name: wonderful_keller results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wonderful_keller This model was trained from scratch on the tomekkorbak/detoxify-pile-chunk3-0-50000, the tomekkorbak/detoxify-pile-chunk3-50000-100000, the tomekkorbak/detoxify-pile-chunk3-100000-150000, the tomekkorbak/detoxify-pile-chunk3-150000-200000, the tomekkorbak/detoxify-pile-chunk3-200000-250000, the tomekkorbak/detoxify-pile-chunk3-250000-300000, the tomekkorbak/detoxify-pile-chunk3-300000-350000, the tomekkorbak/detoxify-pile-chunk3-350000-400000, the tomekkorbak/detoxify-pile-chunk3-400000-450000, the tomekkorbak/detoxify-pile-chunk3-450000-500000, the tomekkorbak/detoxify-pile-chunk3-500000-550000, the tomekkorbak/detoxify-pile-chunk3-550000-600000, the tomekkorbak/detoxify-pile-chunk3-600000-650000, the tomekkorbak/detoxify-pile-chunk3-650000-700000, the tomekkorbak/detoxify-pile-chunk3-700000-750000, the tomekkorbak/detoxify-pile-chunk3-750000-800000, the tomekkorbak/detoxify-pile-chunk3-800000-850000, the tomekkorbak/detoxify-pile-chunk3-850000-900000, the tomekkorbak/detoxify-pile-chunk3-900000-950000, the tomekkorbak/detoxify-pile-chunk3-950000-1000000, the tomekkorbak/detoxify-pile-chunk3-1000000-1050000, the tomekkorbak/detoxify-pile-chunk3-1050000-1100000, the tomekkorbak/detoxify-pile-chunk3-1100000-1150000, the tomekkorbak/detoxify-pile-chunk3-1150000-1200000, the tomekkorbak/detoxify-pile-chunk3-1200000-1250000, the tomekkorbak/detoxify-pile-chunk3-1250000-1300000, the tomekkorbak/detoxify-pile-chunk3-1300000-1350000, the tomekkorbak/detoxify-pile-chunk3-1350000-1400000, the tomekkorbak/detoxify-pile-chunk3-1400000-1450000, the tomekkorbak/detoxify-pile-chunk3-1450000-1500000, the tomekkorbak/detoxify-pile-chunk3-1500000-1550000, the tomekkorbak/detoxify-pile-chunk3-1550000-1600000, the tomekkorbak/detoxify-pile-chunk3-1600000-1650000, the tomekkorbak/detoxify-pile-chunk3-1650000-1700000, the tomekkorbak/detoxify-pile-chunk3-1700000-1750000, the tomekkorbak/detoxify-pile-chunk3-1750000-1800000, the tomekkorbak/detoxify-pile-chunk3-1800000-1850000, the tomekkorbak/detoxify-pile-chunk3-1850000-1900000 and the tomekkorbak/detoxify-pile-chunk3-1900000-1950000 datasets. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.01 - training_steps: 50354 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0+cu113 - Datasets 2.5.1 - Tokenizers 0.11.6 # Full config {'dataset': {'datasets': ['tomekkorbak/detoxify-pile-chunk3-0-50000', 'tomekkorbak/detoxify-pile-chunk3-50000-100000', 'tomekkorbak/detoxify-pile-chunk3-100000-150000', 'tomekkorbak/detoxify-pile-chunk3-150000-200000', 'tomekkorbak/detoxify-pile-chunk3-200000-250000', 'tomekkorbak/detoxify-pile-chunk3-250000-300000', 'tomekkorbak/detoxify-pile-chunk3-300000-350000', 'tomekkorbak/detoxify-pile-chunk3-350000-400000', 'tomekkorbak/detoxify-pile-chunk3-400000-450000', 'tomekkorbak/detoxify-pile-chunk3-450000-500000', 'tomekkorbak/detoxify-pile-chunk3-500000-550000', 'tomekkorbak/detoxify-pile-chunk3-550000-600000', 'tomekkorbak/detoxify-pile-chunk3-600000-650000', 'tomekkorbak/detoxify-pile-chunk3-650000-700000', 'tomekkorbak/detoxify-pile-chunk3-700000-750000', 'tomekkorbak/detoxify-pile-chunk3-750000-800000', 'tomekkorbak/detoxify-pile-chunk3-800000-850000', 'tomekkorbak/detoxify-pile-chunk3-850000-900000', 'tomekkorbak/detoxify-pile-chunk3-900000-950000', 'tomekkorbak/detoxify-pile-chunk3-950000-1000000', 'tomekkorbak/detoxify-pile-chunk3-1000000-1050000', 'tomekkorbak/detoxify-pile-chunk3-1050000-1100000', 'tomekkorbak/detoxify-pile-chunk3-1100000-1150000', 'tomekkorbak/detoxify-pile-chunk3-1150000-1200000', 'tomekkorbak/detoxify-pile-chunk3-1200000-1250000', 'tomekkorbak/detoxify-pile-chunk3-1250000-1300000', 'tomekkorbak/detoxify-pile-chunk3-1300000-1350000', 'tomekkorbak/detoxify-pile-chunk3-1350000-1400000', 'tomekkorbak/detoxify-pile-chunk3-1400000-1450000', 'tomekkorbak/detoxify-pile-chunk3-1450000-1500000', 'tomekkorbak/detoxify-pile-chunk3-1500000-1550000', 'tomekkorbak/detoxify-pile-chunk3-1550000-1600000', 'tomekkorbak/detoxify-pile-chunk3-1600000-1650000', 'tomekkorbak/detoxify-pile-chunk3-1650000-1700000', 'tomekkorbak/detoxify-pile-chunk3-1700000-1750000', 'tomekkorbak/detoxify-pile-chunk3-1750000-1800000', 'tomekkorbak/detoxify-pile-chunk3-1800000-1850000', 'tomekkorbak/detoxify-pile-chunk3-1850000-1900000', 'tomekkorbak/detoxify-pile-chunk3-1900000-1950000'], 'filter_threshold': 0.00078, 'is_split_by_sentences': True}, 'generation': {'force_call_on': [25354], 'metrics_configs': [{}, {'n': 1}, {'n': 2}, {'n': 5}], 'scenario_configs': [{'generate_kwargs': {'do_sample': True, 'max_length': 128, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'unconditional', 'num_samples': 2048}, {'generate_kwargs': {'do_sample': True, 'max_length': 128, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'challenging_rtp', 'num_samples': 2048, 'prompts_path': 'resources/challenging_rtp.jsonl'}], 'scorer_config': {'device': 'cuda:0'}}, 'kl_gpt3_callback': {'force_call_on': [25354], 'max_tokens': 64, 'num_samples': 4096}, 'model': {'from_scratch': True, 'gpt2_config_kwargs': {'reorder_and_upcast_attn': True, 'scale_attn_by': True}, 'path_or_name': 'gpt2'}, 'objective': {'name': 'MLE'}, 'tokenizer': {'path_or_name': 'gpt2'}, 'training': {'dataloader_num_workers': 0, 'effective_batch_size': 64, 'evaluation_strategy': 'no', 'fp16': True, 'hub_model_id': 'wonderful_keller', 'hub_strategy': 'all_checkpoints', 'learning_rate': 0.0005, 'logging_first_step': True, 'logging_steps': 1, 'num_tokens': 3300000000, 'output_dir': 'training_output104340', 'per_device_train_batch_size': 16, 'push_to_hub': True, 'remove_unused_columns': False, 'save_steps': 25354, 'save_strategy': 'steps', 'seed': 42, 'warmup_ratio': 0.01, 'weight_decay': 0.1}} # Wandb URL: https://wandb.ai/tomekkorbak/apo/runs/354p6jl4
tomekkorbak/nifty_banach
tomekkorbak
2022-11-26T00:18:25Z
103
0
transformers
[ "transformers", "pytorch", "gpt2", "generated_from_trainer", "en", "dataset:tomekkorbak/detoxify-pile-chunk3-0-50000", "dataset:tomekkorbak/detoxify-pile-chunk3-50000-100000", "dataset:tomekkorbak/detoxify-pile-chunk3-100000-150000", "dataset:tomekkorbak/detoxify-pile-chunk3-150000-200000", "dataset:tomekkorbak/detoxify-pile-chunk3-200000-250000", "dataset:tomekkorbak/detoxify-pile-chunk3-250000-300000", "dataset:tomekkorbak/detoxify-pile-chunk3-300000-350000", "dataset:tomekkorbak/detoxify-pile-chunk3-350000-400000", "dataset:tomekkorbak/detoxify-pile-chunk3-400000-450000", "dataset:tomekkorbak/detoxify-pile-chunk3-450000-500000", "dataset:tomekkorbak/detoxify-pile-chunk3-500000-550000", "dataset:tomekkorbak/detoxify-pile-chunk3-550000-600000", "dataset:tomekkorbak/detoxify-pile-chunk3-600000-650000", "dataset:tomekkorbak/detoxify-pile-chunk3-650000-700000", "dataset:tomekkorbak/detoxify-pile-chunk3-700000-750000", "dataset:tomekkorbak/detoxify-pile-chunk3-750000-800000", "dataset:tomekkorbak/detoxify-pile-chunk3-800000-850000", "dataset:tomekkorbak/detoxify-pile-chunk3-850000-900000", "dataset:tomekkorbak/detoxify-pile-chunk3-900000-950000", "dataset:tomekkorbak/detoxify-pile-chunk3-950000-1000000", "dataset:tomekkorbak/detoxify-pile-chunk3-1000000-1050000", "dataset:tomekkorbak/detoxify-pile-chunk3-1050000-1100000", "dataset:tomekkorbak/detoxify-pile-chunk3-1100000-1150000", "dataset:tomekkorbak/detoxify-pile-chunk3-1150000-1200000", "dataset:tomekkorbak/detoxify-pile-chunk3-1200000-1250000", "dataset:tomekkorbak/detoxify-pile-chunk3-1250000-1300000", "dataset:tomekkorbak/detoxify-pile-chunk3-1300000-1350000", "dataset:tomekkorbak/detoxify-pile-chunk3-1350000-1400000", "dataset:tomekkorbak/detoxify-pile-chunk3-1400000-1450000", "dataset:tomekkorbak/detoxify-pile-chunk3-1450000-1500000", "dataset:tomekkorbak/detoxify-pile-chunk3-1500000-1550000", "dataset:tomekkorbak/detoxify-pile-chunk3-1550000-1600000", "dataset:tomekkorbak/detoxify-pile-chunk3-1600000-1650000", "dataset:tomekkorbak/detoxify-pile-chunk3-1650000-1700000", "dataset:tomekkorbak/detoxify-pile-chunk3-1700000-1750000", "dataset:tomekkorbak/detoxify-pile-chunk3-1750000-1800000", "dataset:tomekkorbak/detoxify-pile-chunk3-1800000-1850000", "dataset:tomekkorbak/detoxify-pile-chunk3-1850000-1900000", "dataset:tomekkorbak/detoxify-pile-chunk3-1900000-1950000", "license:mit", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
2022-11-25T10:19:02Z
--- language: - en license: mit tags: - generated_from_trainer datasets: - tomekkorbak/detoxify-pile-chunk3-0-50000 - tomekkorbak/detoxify-pile-chunk3-50000-100000 - tomekkorbak/detoxify-pile-chunk3-100000-150000 - tomekkorbak/detoxify-pile-chunk3-150000-200000 - tomekkorbak/detoxify-pile-chunk3-200000-250000 - tomekkorbak/detoxify-pile-chunk3-250000-300000 - tomekkorbak/detoxify-pile-chunk3-300000-350000 - tomekkorbak/detoxify-pile-chunk3-350000-400000 - tomekkorbak/detoxify-pile-chunk3-400000-450000 - tomekkorbak/detoxify-pile-chunk3-450000-500000 - tomekkorbak/detoxify-pile-chunk3-500000-550000 - tomekkorbak/detoxify-pile-chunk3-550000-600000 - tomekkorbak/detoxify-pile-chunk3-600000-650000 - tomekkorbak/detoxify-pile-chunk3-650000-700000 - tomekkorbak/detoxify-pile-chunk3-700000-750000 - tomekkorbak/detoxify-pile-chunk3-750000-800000 - tomekkorbak/detoxify-pile-chunk3-800000-850000 - tomekkorbak/detoxify-pile-chunk3-850000-900000 - tomekkorbak/detoxify-pile-chunk3-900000-950000 - tomekkorbak/detoxify-pile-chunk3-950000-1000000 - tomekkorbak/detoxify-pile-chunk3-1000000-1050000 - tomekkorbak/detoxify-pile-chunk3-1050000-1100000 - tomekkorbak/detoxify-pile-chunk3-1100000-1150000 - tomekkorbak/detoxify-pile-chunk3-1150000-1200000 - tomekkorbak/detoxify-pile-chunk3-1200000-1250000 - tomekkorbak/detoxify-pile-chunk3-1250000-1300000 - tomekkorbak/detoxify-pile-chunk3-1300000-1350000 - tomekkorbak/detoxify-pile-chunk3-1350000-1400000 - tomekkorbak/detoxify-pile-chunk3-1400000-1450000 - tomekkorbak/detoxify-pile-chunk3-1450000-1500000 - tomekkorbak/detoxify-pile-chunk3-1500000-1550000 - tomekkorbak/detoxify-pile-chunk3-1550000-1600000 - tomekkorbak/detoxify-pile-chunk3-1600000-1650000 - tomekkorbak/detoxify-pile-chunk3-1650000-1700000 - tomekkorbak/detoxify-pile-chunk3-1700000-1750000 - tomekkorbak/detoxify-pile-chunk3-1750000-1800000 - tomekkorbak/detoxify-pile-chunk3-1800000-1850000 - tomekkorbak/detoxify-pile-chunk3-1850000-1900000 - tomekkorbak/detoxify-pile-chunk3-1900000-1950000 model-index: - name: nifty_banach results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # nifty_banach This model was trained from scratch on the tomekkorbak/detoxify-pile-chunk3-0-50000, the tomekkorbak/detoxify-pile-chunk3-50000-100000, the tomekkorbak/detoxify-pile-chunk3-100000-150000, the tomekkorbak/detoxify-pile-chunk3-150000-200000, the tomekkorbak/detoxify-pile-chunk3-200000-250000, the tomekkorbak/detoxify-pile-chunk3-250000-300000, the tomekkorbak/detoxify-pile-chunk3-300000-350000, the tomekkorbak/detoxify-pile-chunk3-350000-400000, the tomekkorbak/detoxify-pile-chunk3-400000-450000, the tomekkorbak/detoxify-pile-chunk3-450000-500000, the tomekkorbak/detoxify-pile-chunk3-500000-550000, the tomekkorbak/detoxify-pile-chunk3-550000-600000, the tomekkorbak/detoxify-pile-chunk3-600000-650000, the tomekkorbak/detoxify-pile-chunk3-650000-700000, the tomekkorbak/detoxify-pile-chunk3-700000-750000, the tomekkorbak/detoxify-pile-chunk3-750000-800000, the tomekkorbak/detoxify-pile-chunk3-800000-850000, the tomekkorbak/detoxify-pile-chunk3-850000-900000, the tomekkorbak/detoxify-pile-chunk3-900000-950000, the tomekkorbak/detoxify-pile-chunk3-950000-1000000, the tomekkorbak/detoxify-pile-chunk3-1000000-1050000, the tomekkorbak/detoxify-pile-chunk3-1050000-1100000, the tomekkorbak/detoxify-pile-chunk3-1100000-1150000, the tomekkorbak/detoxify-pile-chunk3-1150000-1200000, the tomekkorbak/detoxify-pile-chunk3-1200000-1250000, the tomekkorbak/detoxify-pile-chunk3-1250000-1300000, the tomekkorbak/detoxify-pile-chunk3-1300000-1350000, the tomekkorbak/detoxify-pile-chunk3-1350000-1400000, the tomekkorbak/detoxify-pile-chunk3-1400000-1450000, the tomekkorbak/detoxify-pile-chunk3-1450000-1500000, the tomekkorbak/detoxify-pile-chunk3-1500000-1550000, the tomekkorbak/detoxify-pile-chunk3-1550000-1600000, the tomekkorbak/detoxify-pile-chunk3-1600000-1650000, the tomekkorbak/detoxify-pile-chunk3-1650000-1700000, the tomekkorbak/detoxify-pile-chunk3-1700000-1750000, the tomekkorbak/detoxify-pile-chunk3-1750000-1800000, the tomekkorbak/detoxify-pile-chunk3-1800000-1850000, the tomekkorbak/detoxify-pile-chunk3-1850000-1900000 and the tomekkorbak/detoxify-pile-chunk3-1900000-1950000 datasets. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.01 - training_steps: 50354 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0+cu113 - Datasets 2.5.1 - Tokenizers 0.11.6 # Full config {'dataset': {'datasets': ['tomekkorbak/detoxify-pile-chunk3-0-50000', 'tomekkorbak/detoxify-pile-chunk3-50000-100000', 'tomekkorbak/detoxify-pile-chunk3-100000-150000', 'tomekkorbak/detoxify-pile-chunk3-150000-200000', 'tomekkorbak/detoxify-pile-chunk3-200000-250000', 'tomekkorbak/detoxify-pile-chunk3-250000-300000', 'tomekkorbak/detoxify-pile-chunk3-300000-350000', 'tomekkorbak/detoxify-pile-chunk3-350000-400000', 'tomekkorbak/detoxify-pile-chunk3-400000-450000', 'tomekkorbak/detoxify-pile-chunk3-450000-500000', 'tomekkorbak/detoxify-pile-chunk3-500000-550000', 'tomekkorbak/detoxify-pile-chunk3-550000-600000', 'tomekkorbak/detoxify-pile-chunk3-600000-650000', 'tomekkorbak/detoxify-pile-chunk3-650000-700000', 'tomekkorbak/detoxify-pile-chunk3-700000-750000', 'tomekkorbak/detoxify-pile-chunk3-750000-800000', 'tomekkorbak/detoxify-pile-chunk3-800000-850000', 'tomekkorbak/detoxify-pile-chunk3-850000-900000', 'tomekkorbak/detoxify-pile-chunk3-900000-950000', 'tomekkorbak/detoxify-pile-chunk3-950000-1000000', 'tomekkorbak/detoxify-pile-chunk3-1000000-1050000', 'tomekkorbak/detoxify-pile-chunk3-1050000-1100000', 'tomekkorbak/detoxify-pile-chunk3-1100000-1150000', 'tomekkorbak/detoxify-pile-chunk3-1150000-1200000', 'tomekkorbak/detoxify-pile-chunk3-1200000-1250000', 'tomekkorbak/detoxify-pile-chunk3-1250000-1300000', 'tomekkorbak/detoxify-pile-chunk3-1300000-1350000', 'tomekkorbak/detoxify-pile-chunk3-1350000-1400000', 'tomekkorbak/detoxify-pile-chunk3-1400000-1450000', 'tomekkorbak/detoxify-pile-chunk3-1450000-1500000', 'tomekkorbak/detoxify-pile-chunk3-1500000-1550000', 'tomekkorbak/detoxify-pile-chunk3-1550000-1600000', 'tomekkorbak/detoxify-pile-chunk3-1600000-1650000', 'tomekkorbak/detoxify-pile-chunk3-1650000-1700000', 'tomekkorbak/detoxify-pile-chunk3-1700000-1750000', 'tomekkorbak/detoxify-pile-chunk3-1750000-1800000', 'tomekkorbak/detoxify-pile-chunk3-1800000-1850000', 'tomekkorbak/detoxify-pile-chunk3-1850000-1900000', 'tomekkorbak/detoxify-pile-chunk3-1900000-1950000'], 'is_split_by_sentences': True}, 'generation': {'force_call_on': [25354], 'metrics_configs': [{}, {'n': 1}, {'n': 2}, {'n': 5}], 'scenario_configs': [{'generate_kwargs': {'do_sample': True, 'max_length': 128, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'unconditional', 'num_samples': 2048}, {'generate_kwargs': {'do_sample': True, 'max_length': 128, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'challenging_rtp', 'num_samples': 2048, 'prompts_path': 'resources/challenging_rtp.jsonl'}], 'scorer_config': {'device': 'cuda:0'}}, 'kl_gpt3_callback': {'force_call_on': [25354], 'max_tokens': 64, 'num_samples': 4096}, 'model': {'from_scratch': True, 'gpt2_config_kwargs': {'reorder_and_upcast_attn': True, 'scale_attn_by': True}, 'path_or_name': 'gpt2'}, 'objective': {'alpha': 1, 'name': 'Unlikelihood', 'score_threshold': 0.00078}, 'tokenizer': {'path_or_name': 'gpt2'}, 'training': {'dataloader_num_workers': 0, 'effective_batch_size': 64, 'evaluation_strategy': 'no', 'fp16': True, 'hub_model_id': 'nifty_banach', 'hub_strategy': 'all_checkpoints', 'learning_rate': 0.0005, 'logging_first_step': True, 'logging_steps': 1, 'num_tokens': 3300000000, 'output_dir': 'training_output104340', 'per_device_train_batch_size': 16, 'push_to_hub': True, 'remove_unused_columns': False, 'save_steps': 25354, 'save_strategy': 'steps', 'seed': 42, 'warmup_ratio': 0.01, 'weight_decay': 0.1}} # Wandb URL: https://wandb.ai/tomekkorbak/apo/runs/1adikmkl
SiddharthaM/hasoc19-xlm-roberta-base-profane
SiddharthaM
2022-11-25T23:26:02Z
104
0
transformers
[ "transformers", "pytorch", "tensorboard", "xlm-roberta", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-11-23T00:26:47Z
# hasoc19-xlm-roberta-base-hatespeech-3 This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.379585 - Accuracy: 0.919202 - Precision: 0.888890 - Recall: 0.832338 - F1: 0.856625 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results ### Framework versions - Transformers 4.24.0.dev0 - Pytorch 1.11.0+cu102 - Datasets 2.6.1 - Tokenizers 0.13.1
TUMxudashuai/ppo-CartPole-v1
TUMxudashuai
2022-11-25T21:01:17Z
2
0
stable-baselines3
[ "stable-baselines3", "CartPole-v1", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-11-25T19:38:16Z
--- library_name: stable-baselines3 tags: - CartPole-v1 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 metrics: - type: mean_reward value: 500.00 +/- 0.00 name: mean_reward verified: false --- # **PPO** Agent playing **CartPole-v1** This is a trained model of a **PPO** agent playing **CartPole-v1** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
arbinger5217/arbinger
arbinger5217
2022-11-25T20:21:54Z
0
0
null
[ "region:us" ]
null
2022-11-25T19:31:17Z
# Stable Diffusion web UI A browser interface based on Gradio library for Stable Diffusion. ![](txt2img_Screenshot.png) Check the [custom scripts](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Custom-Scripts) wiki page for extra scripts developed by users. ## Features [Detailed feature showcase with images](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Features): - Original txt2img and img2img modes - One click install and run script (but you still must install python and git) - Outpainting - Inpainting - Color Sketch - Prompt Matrix - Stable Diffusion Upscale - Attention, specify parts of text that the model should pay more attention to - a man in a ((tuxedo)) - will pay more attention to tuxedo - a man in a (tuxedo:1.21) - alternative syntax - select text and press ctrl+up or ctrl+down to automatically adjust attention to selected text (code contributed by anonymous user) - Loopback, run img2img processing multiple times - X/Y plot, a way to draw a 2 dimensional plot of images with different parameters - Textual Inversion - have as many embeddings as you want and use any names you like for them - use multiple embeddings with different numbers of vectors per token - works with half precision floating point numbers - train embeddings on 8GB (also reports of 6GB working) - Extras tab with: - GFPGAN, neural network that fixes faces - CodeFormer, face restoration tool as an alternative to GFPGAN - RealESRGAN, neural network upscaler - ESRGAN, neural network upscaler with a lot of third party models - SwinIR and Swin2SR([see here](https://github.com/AUTOMATIC1111/stable-diffusion-webui/pull/2092)), neural network upscalers - LDSR, Latent diffusion super resolution upscaling - Resizing aspect ratio options - Sampling method selection - Adjust sampler eta values (noise multiplier) - More advanced noise setting options - Interrupt processing at any time - 4GB video card support (also reports of 2GB working) - Correct seeds for batches - Live prompt token length validation - Generation parameters - parameters you used to generate images are saved with that image - in PNG chunks for PNG, in EXIF for JPEG - can drag the image to PNG info tab to restore generation parameters and automatically copy them into UI - can be disabled in settings - drag and drop an image/text-parameters to promptbox - Read Generation Parameters Button, loads parameters in promptbox to UI - Settings page - Running arbitrary python code from UI (must run with --allow-code to enable) - Mouseover hints for most UI elements - Possible to change defaults/mix/max/step values for UI elements via text config - Random artist button - Tiling support, a checkbox to create images that can be tiled like textures - Progress bar and live image generation preview - Negative prompt, an extra text field that allows you to list what you don't want to see in generated image - Styles, a way to save part of prompt and easily apply them via dropdown later - Variations, a way to generate same image but with tiny differences - Seed resizing, a way to generate same image but at slightly different resolution - CLIP interrogator, a button that tries to guess prompt from an image - Prompt Editing, a way to change prompt mid-generation, say to start making a watermelon and switch to anime girl midway - Batch Processing, process a group of files using img2img - Img2img Alternative, reverse Euler method of cross attention control - Highres Fix, a convenience option to produce high resolution pictures in one click without usual distortions - Reloading checkpoints on the fly - Checkpoint Merger, a tab that allows you to merge up to 3 checkpoints into one - [Custom scripts](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Custom-Scripts) with many extensions from community - [Composable-Diffusion](https://energy-based-model.github.io/Compositional-Visual-Generation-with-Composable-Diffusion-Models/), a way to use multiple prompts at once - separate prompts using uppercase `AND` - also supports weights for prompts: `a cat :1.2 AND a dog AND a penguin :2.2` - No token limit for prompts (original stable diffusion lets you use up to 75 tokens) - DeepDanbooru integration, creates danbooru style tags for anime prompts - [xformers](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Xformers), major speed increase for select cards: (add --xformers to commandline args) - via extension: [History tab](https://github.com/yfszzx/stable-diffusion-webui-images-browser): view, direct and delete images conveniently within the UI - Generate forever option - Training tab - hypernetworks and embeddings options - Preprocessing images: cropping, mirroring, autotagging using BLIP or deepdanbooru (for anime) - Clip skip - Use Hypernetworks - Use VAEs - Estimated completion time in progress bar - API - Support for dedicated [inpainting model](https://github.com/runwayml/stable-diffusion#inpainting-with-stable-diffusion) by RunwayML. - via extension: [Aesthetic Gradients](https://github.com/AUTOMATIC1111/stable-diffusion-webui-aesthetic-gradients), a way to generate images with a specific aesthetic by using clip images embds (implementation of [https://github.com/vicgalle/stable-diffusion-aesthetic-gradients](https://github.com/vicgalle/stable-diffusion-aesthetic-gradients)) ## Where are Aesthetic Gradients?!?! Aesthetic Gradients are now an extension. You can install it using git: ```commandline git clone https://github.com/AUTOMATIC1111/stable-diffusion-webui-aesthetic-gradients extensions/aesthetic-gradients ``` After running this command, make sure that you have `aesthetic-gradients` dir in webui's `extensions` directory and restart the UI. The interface for Aesthetic Gradients should appear exactly the same as it was. ## Where is History/Image browser?!?! Image browser is now an extension. You can install it using git: ```commandline git clone https://github.com/yfszzx/stable-diffusion-webui-images-browser extensions/images-browser ``` After running this command, make sure that you have `images-browser` dir in webui's `extensions` directory and restart the UI. The interface for Image browser should appear exactly the same as it was. ## Installation and Running Make sure the required [dependencies](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Dependencies) are met and follow the instructions available for both [NVidia](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Install-and-Run-on-NVidia-GPUs) (recommended) and [AMD](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Install-and-Run-on-AMD-GPUs) GPUs. Alternatively, use online services (like Google Colab): - [List of Online Services](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Online-Services) ### Automatic Installation on Windows 1. Install [Python 3.10.6](https://www.python.org/downloads/windows/), checking "Add Python to PATH" 2. Install [git](https://git-scm.com/download/win). 3. Download the stable-diffusion-webui repository, for example by running `git clone https://github.com/AUTOMATIC1111/stable-diffusion-webui.git`. 4. Place `model.ckpt` in the `models` directory (see [dependencies](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Dependencies) for where to get it). 5. _*(Optional)*_ Place `GFPGANv1.4.pth` in the base directory, alongside `webui.py` (see [dependencies](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Dependencies) for where to get it). 6. Run `webui-user.bat` from Windows Explorer as normal, non-administrator, user. ### Automatic Installation on Linux 1. Install the dependencies: ```bash # Debian-based: sudo apt install wget git python3 python3-venv # Red Hat-based: sudo dnf install wget git python3 # Arch-based: sudo pacman -S wget git python3 ``` 2. To install in `/home/$(whoami)/stable-diffusion-webui/`, run: ```bash bash <(wget -qO- https://raw.githubusercontent.com/AUTOMATIC1111/stable-diffusion-webui/master/webui.sh) ``` ### Installation on Apple Silicon Find the instructions [here](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Installation-on-Apple-Silicon). ## Contributing Here's how to add code to this repo: [Contributing](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Contributing) ## Documentation The documentation was moved from this README over to the project's [wiki](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki). ## Credits - Stable Diffusion - https://github.com/CompVis/stable-diffusion, https://github.com/CompVis/taming-transformers - k-diffusion - https://github.com/crowsonkb/k-diffusion.git - GFPGAN - https://github.com/TencentARC/GFPGAN.git - CodeFormer - https://github.com/sczhou/CodeFormer - ESRGAN - https://github.com/xinntao/ESRGAN - SwinIR - https://github.com/JingyunLiang/SwinIR - Swin2SR - https://github.com/mv-lab/swin2sr - LDSR - https://github.com/Hafiidz/latent-diffusion - Ideas for optimizations - https://github.com/basujindal/stable-diffusion - Cross Attention layer optimization - Doggettx - https://github.com/Doggettx/stable-diffusion, original idea for prompt editing. - Cross Attention layer optimization - InvokeAI, lstein - https://github.com/invoke-ai/InvokeAI (originally http://github.com/lstein/stable-diffusion) - Textual Inversion - Rinon Gal - https://github.com/rinongal/textual_inversion (we're not using his code, but we are using his ideas). - Idea for SD upscale - https://github.com/jquesnelle/txt2imghd - Noise generation for outpainting mk2 - https://github.com/parlance-zz/g-diffuser-bot - CLIP interrogator idea and borrowing some code - https://github.com/pharmapsychotic/clip-interrogator - Idea for Composable Diffusion - https://github.com/energy-based-model/Compositional-Visual-Generation-with-Composable-Diffusion-Models-PyTorch - xformers - https://github.com/facebookresearch/xformers - DeepDanbooru - interrogator for anime diffusers https://github.com/KichangKim/DeepDanbooru - Security advice - RyotaK - Initial Gradio script - posted on 4chan by an Anonymous user. Thank you Anonymous user. - (You)
kbalde/mt5-small-finetuned-amazon-en-es
kbalde
2022-11-25T20:18:33Z
110
0
transformers
[ "transformers", "pytorch", "tensorboard", "mt5", "text2text-generation", "summarization", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
summarization
2022-11-25T19:38:24Z
--- license: apache-2.0 tags: - summarization - generated_from_trainer metrics: - rouge model-index: - name: mt5-small-finetuned-amazon-en-es results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mt5-small-finetuned-amazon-en-es This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on the None dataset. It achieves the following results on the evaluation set: - Loss: 3.0329 - Rouge1: 16.3034 - Rouge2: 7.8192 - Rougel: 16.0316 - Rougelsum: 15.9173 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5.6e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 8 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | |:-------------:|:-----:|:----:|:---------------:|:-------:|:------:|:-------:|:---------:| | 7.0891 | 1.0 | 1209 | 3.2989 | 13.8686 | 6.1132 | 13.3657 | 13.3454 | | 3.9283 | 2.0 | 2418 | 3.1443 | 16.3537 | 7.9374 | 15.8565 | 15.7281 | | 3.5985 | 3.0 | 3627 | 3.1004 | 17.9042 | 9.1908 | 17.5268 | 17.385 | | 3.4285 | 4.0 | 4836 | 3.0578 | 16.3118 | 8.4563 | 15.9252 | 15.9109 | | 3.3222 | 5.0 | 6045 | 3.0587 | 17.5106 | 8.6579 | 17.2096 | 17.1079 | | 3.2554 | 6.0 | 7254 | 3.0497 | 16.9153 | 8.0973 | 16.5874 | 16.4807 | | 3.2085 | 7.0 | 8463 | 3.0309 | 16.3789 | 7.9306 | 16.1233 | 16.0097 | | 3.1856 | 8.0 | 9672 | 3.0329 | 16.3034 | 7.8192 | 16.0316 | 15.9173 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
bnunticha/t5-small-en-to-th
bnunticha
2022-11-25T20:17:29Z
102
0
transformers
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-25T19:32:31Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - bleu model-index: - name: t5-small-en-to-th results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-en-to-th This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0527 - Bleu: 0.0 - Gen Len: 17.5726 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len | |:-------------:|:-----:|:-----:|:---------------:|:----:|:-------:| | 0.0414 | 1.0 | 17810 | 0.0527 | 0.0 | 17.5726 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
phqlong/xlmr-zacqa-finetuned
phqlong
2022-11-25T19:14:51Z
105
0
transformers
[ "transformers", "pytorch", "xlm-roberta", "question-answering", "generated_from_trainer", "license:mit", "endpoints_compatible", "region:us" ]
question-answering
2022-11-25T18:28:59Z
--- license: mit tags: - generated_from_trainer model-index: - name: xlmr-zacqa-finetuned results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlmr-zacqa-finetuned This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.22.2 - Pytorch 1.8.1+cu101 - Datasets 2.5.2 - Tokenizers 0.12.1
laverdes/donut-base-sroie
laverdes
2022-11-25T18:53:16Z
48
1
transformers
[ "transformers", "pytorch", "tensorboard", "vision-encoder-decoder", "image-text-to-text", "generated_from_trainer", "dataset:imagefolder", "license:mit", "endpoints_compatible", "region:us" ]
image-text-to-text
2022-11-25T15:52:17Z
--- license: mit tags: - generated_from_trainer datasets: - imagefolder model-index: - name: donut-base-sroie results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # donut-base-sroie This model is a fine-tuned version of [naver-clova-ix/donut-base](https://huggingface.co/naver-clova-ix/donut-base) on the imagefolder dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.23.1 - Pytorch 1.12.1 - Datasets 2.7.0 - Tokenizers 0.11.0
kejian/final-cond-10-0.01
kejian
2022-11-25T18:15:12Z
105
0
transformers
[ "transformers", "pytorch", "gpt2", "generated_from_trainer", "en", "dataset:kejian/codeparrot-train-more-filter-3.3b-cleaned", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
2022-11-25T09:15:03Z
--- language: - en license: apache-2.0 tags: - generated_from_trainer datasets: - kejian/codeparrot-train-more-filter-3.3b-cleaned model-index: - name: kejian/final-cond-10-0.01 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # kejian/final-cond-10-0.01 This model was trained from scratch on the kejian/codeparrot-train-more-filter-3.3b-cleaned dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0008 - train_batch_size: 64 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.01 - training_steps: 50354 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.23.0 - Pytorch 1.13.0+cu116 - Datasets 2.0.0 - Tokenizers 0.12.1 # Full config {'dataset': {'conditional_training_config': {'aligned_prefix': '<|aligned|>', 'drop_token_fraction': 0.01, 'misaligned_prefix': '<|misaligned|>', 'threshold': 0}, 'datasets': ['kejian/codeparrot-train-more-filter-3.3b-cleaned'], 'is_split_by_sentences': True}, 'generation': {'batch_size': 64, 'metrics_configs': [{}, {'n': 1}, {}], 'scenario_configs': [{'display_as_html': True, 'generate_kwargs': {'do_sample': True, 'eos_token_id': 0, 'max_length': 704, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'unconditional', 'num_samples': 512, 'prefix': '<|aligned|>', 'use_prompt_for_scoring': False}, {'display_as_html': True, 'generate_kwargs': {'do_sample': True, 'eos_token_id': 0, 'max_length': 272, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'functions', 'num_samples': 512, 'prefix': '<|aligned|>', 'prompt_before_control': True, 'prompts_path': 'resources/functions_csnet.jsonl', 'use_prompt_for_scoring': True}], 'scorer_config': {}}, 'kl_gpt3_callback': {'gpt3_kwargs': {'model_name': 'code-cushman-001'}, 'max_tokens': 64, 'num_samples': 4096, 'prefix': '<|aligned|>'}, 'model': {'from_scratch': True, 'gpt2_config_kwargs': {'reorder_and_upcast_attn': True, 'scale_attn_by': True}, 'num_additional_tokens': 2, 'path_or_name': 'codeparrot/codeparrot-small'}, 'objective': {'name': 'MLE'}, 'tokenizer': {'path_or_name': 'codeparrot/codeparrot-small', 'special_tokens': ['<|aligned|>', '<|misaligned|>']}, 'training': {'dataloader_num_workers': 0, 'effective_batch_size': 64, 'evaluation_strategy': 'no', 'fp16': True, 'hub_model_id': 'kejian/final-cond-10-0.01', 'hub_strategy': 'all_checkpoints', 'learning_rate': 0.0008, 'logging_first_step': True, 'logging_steps': 1, 'num_tokens': 3300000000.0, 'output_dir': 'training_output', 'per_device_train_batch_size': 16, 'push_to_hub': True, 'remove_unused_columns': False, 'save_steps': 5000, 'save_strategy': 'steps', 'seed': 42, 'warmup_ratio': 0.01, 'weight_decay': 0.1}} # Wandb URL: https://wandb.ai/kejian/uncategorized/runs/1wgqepja
kejian/final-cond-25-0.05
kejian
2022-11-25T18:14:24Z
103
0
transformers
[ "transformers", "pytorch", "gpt2", "generated_from_trainer", "en", "dataset:kejian/codeparrot-train-more-filter-3.3b-cleaned", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
2022-11-25T09:05:02Z
--- language: - en license: apache-2.0 tags: - generated_from_trainer datasets: - kejian/codeparrot-train-more-filter-3.3b-cleaned model-index: - name: kejian/final-cond-25-0.05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # kejian/final-cond-25-0.05 This model was trained from scratch on the kejian/codeparrot-train-more-filter-3.3b-cleaned dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0008 - train_batch_size: 32 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.01 - training_steps: 50354 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.23.0 - Pytorch 1.13.0+cu116 - Datasets 2.0.0 - Tokenizers 0.12.1 # Full config {'dataset': {'conditional_training_config': {'aligned_prefix': '<|aligned|>', 'drop_token_fraction': 0.05, 'misaligned_prefix': '<|misaligned|>', 'threshold': 0.000475}, 'datasets': ['kejian/codeparrot-train-more-filter-3.3b-cleaned'], 'is_split_by_sentences': True}, 'generation': {'batch_size': 64, 'metrics_configs': [{}, {'n': 1}, {}], 'scenario_configs': [{'display_as_html': True, 'generate_kwargs': {'do_sample': True, 'eos_token_id': 0, 'max_length': 704, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'unconditional', 'num_samples': 512, 'prefix': '<|aligned|>', 'use_prompt_for_scoring': False}, {'display_as_html': True, 'generate_kwargs': {'do_sample': True, 'eos_token_id': 0, 'max_length': 272, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'functions', 'num_samples': 512, 'prefix': '<|aligned|>', 'prompt_before_control': True, 'prompts_path': 'resources/functions_csnet.jsonl', 'use_prompt_for_scoring': True}], 'scorer_config': {}}, 'kl_gpt3_callback': {'gpt3_kwargs': {'model_name': 'code-cushman-001'}, 'max_tokens': 64, 'num_samples': 4096, 'prefix': '<|aligned|>'}, 'model': {'from_scratch': True, 'gpt2_config_kwargs': {'reorder_and_upcast_attn': True, 'scale_attn_by': True}, 'num_additional_tokens': 2, 'path_or_name': 'codeparrot/codeparrot-small'}, 'objective': {'name': 'MLE'}, 'tokenizer': {'path_or_name': 'codeparrot/codeparrot-small', 'special_tokens': ['<|aligned|>', '<|misaligned|>']}, 'training': {'dataloader_num_workers': 0, 'effective_batch_size': 64, 'evaluation_strategy': 'no', 'fp16': True, 'hub_model_id': 'kejian/final-cond-25-0.05', 'hub_strategy': 'all_checkpoints', 'learning_rate': 0.0008, 'logging_first_step': True, 'logging_steps': 1, 'num_tokens': 3300000000.0, 'output_dir': 'training_output', 'per_device_train_batch_size': 16, 'push_to_hub': True, 'remove_unused_columns': False, 'save_steps': 5000, 'save_strategy': 'steps', 'seed': 42, 'warmup_ratio': 0.01, 'weight_decay': 0.1}} # Wandb URL: https://wandb.ai/kejian/uncategorized/runs/vf26d3rh
valhalla/m2m100_tiny_random
valhalla
2022-11-25T17:22:06Z
140
0
transformers
[ "transformers", "pytorch", "m2m_100", "text2text-generation", "text-2-text-generation", "multilingual", "arxiv:2010.11125", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2022-03-02T23:29:05Z
--- language: - multilingual tags: - text-2-text-generation - m2m_100 --- # Model Card for KeywordIdentifier # Model Details ## Model Description More information needed - **Developed by:** Facebook - **Shared by [Optional]:** Suraj Patil - **Model type:** Text2Text Generation - **Language(s) (NLP):** More information needed - **License:** More information needed - **Parent Model:** [M2M100]https://huggingface.co/facebook/m2m100_418M) - **Resources for more information:** - [M2M100 Associated Paper](https://arxiv.org/abs/2010.11125) # Uses ## Direct Use This model can be used for the task of Text2Text Generation. ## Downstream Use [Optional] More information needed. ## Out-of-Scope Use The model should not be used to intentionally create hostile or alienating environments for people. # Bias, Risks, and Limitations Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups. ## Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. # Training Details ## Training Data More information needed ## Training Procedure ### Preprocessing More information needed ### Speeds, Sizes, Times More information needed # Evaluation ## Testing Data, Factors & Metrics ### Testing Data More information needed ### Factors More information needed ### Metrics More information needed ## Results More information needed # Model Examination More information needed # Environmental Impact Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** More information needed - **Hours used:** More information needed - **Cloud Provider:** More information needed - **Compute Region:** More information needed - **Carbon Emitted:** More information needed # Technical Specifications [optional] ## Model Architecture and Objective More information needed ## Compute Infrastructure More information needed ### Hardware More information needed ### Software More information needed. # Citation **BibTeX:** More information needed ```bibtex @misc{fan2020englishcentric, title={Beyond English-Centric Multilingual Machine Translation}, author={Angela Fan and Shruti Bhosale and Holger Schwenk and Zhiyi Ma and Ahmed El-Kishky and Siddharth Goyal and Mandeep Baines and Onur Celebi and Guillaume Wenzek and Vishrav Chaudhary and Naman Goyal and Tom Birch and Vitaliy Liptchinsky and Sergey Edunov and Edouard Grave and Michael Auli and Armand Joulin}, year={2020}, eprint={2010.11125}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` **APA:** More information needed # Glossary [optional] More information needed # More Information [optional] See the [model hub](https://huggingface.co/models?filter=m2m_100) for more fine-tuned versions. # Model Card Authors [optional] Suraj Patil in collaboration with Ezi Ozoani and the Hugging Face team # Model Card Contact More information needed # How to Get Started with the Model Use the code below to get started with the model. <details> <summary> Click to expand </summary> ```python from transformers import AutoTokenizer, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("valhalla/m2m100_tiny_random") model = AutoModelForSeq2SeqLM.from_pretrained("valhalla/m2m100_tiny_random") ``` </details>
CompVis/stable-diffusion-safety-checker
CompVis
2022-11-25T17:21:38Z
1,333,514
125
transformers
[ "transformers", "pytorch", "clip", "arxiv:2103.00020", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2022-08-22T10:22:34Z
--- tags: - clip --- # Model Card for stable-diffusion-safety-checker # Model Details ## Model Description More information needed - **Developed by:** More information needed - **Shared by [Optional]:** CompVis - **Model type:** Image Identification - **Language(s) (NLP):** More information needed - **License:** More information needed - **Parent Model:** [CLIP](https://huggingface.co/openai/clip-vit-large-patch14) - **Resources for more information:** - [CLIP Paper](https://arxiv.org/abs/2103.00020) - [Stable Diffusion Model Card](https://github.com/CompVis/stable-diffusion/blob/main/Stable_Diffusion_v1_Model_Card.md) # Uses ## Direct Use This model can be used for identifying NSFW image The CLIP model devlopers note in their [model card](https://huggingface.co/openai/clip-vit-large-patch14) : >The primary intended users of these models are AI researchers. We primarily imagine the model will be used by researchers to better understand robustness, generalization, and other capabilities, biases, and constraints of computer vision models. ## Downstream Use [Optional] More information needed. ## Out-of-Scope Use The model is not intended to be used with transformers but with diffusers. This model should also not be used to intentionally create hostile or alienating environments for people. # Bias, Risks, and Limitations Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups. The CLIP model devlopers note in their [model card](https://huggingface.co/openai/clip-vit-large-patch14) : > We find that the performance of CLIP - and the specific biases it exhibits - can depend significantly on class design and the choices one makes for categories to include and exclude. We tested the risk of certain kinds of denigration with CLIP by classifying images of people from Fairface into crime-related and non-human animal categories. We found significant disparities with respect to race and gender. Additionally, we found that these disparities could shift based on how the classes were constructed. > We also tested the performance of CLIP on gender, race and age classification using the Fairface dataset (We default to using race categories as they are constructed in the Fairface dataset.) in order to assess quality of performance across different demographics. We found accuracy >96% across all races for gender classification with ‘Middle Eastern’ having the highest accuracy (98.4%) and ‘White’ having the lowest (96.5%). Additionally, CLIP averaged ~93% for racial classification and ~63% for age classification ## Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. # Training Details ## Training Data More information needed ## Training Procedure ### Preprocessing More information needed ### Speeds, Sizes, Times More information needed # Evaluation ## Testing Data, Factors & Metrics ### Testing Data More information needed ### Factors More information needed ### Metrics More information needed ## Results More information needed # Model Examination More information needed # Environmental Impact Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** More information needed - **Hours used:** More information needed - **Cloud Provider:** More information needed - **Compute Region:** More information needed - **Carbon Emitted:** More information needed # Technical Specifications [optional] ## Model Architecture and Objective The CLIP model devlopers note in their [model card](https://huggingface.co/openai/clip-vit-large-patch14) : > The base model uses a ViT-L/14 Transformer architecture as an image encoder and uses a masked self-attention Transformer as a text encoder. These encoders are trained to maximize the similarity of (image, text) pairs via a contrastive loss. ## Compute Infrastructure More information needed ### Hardware More information needed ### Software More information needed. # Citation **BibTeX:** More information needed **APA:** More information needed # Glossary [optional] More information needed # More Information [optional] More information needed # Model Card Authors [optional] CompVis in collaboration with Ezi Ozoani and the Hugging Face team # Model Card Contact More information needed # How to Get Started with the Model Use the code below to get started with the model. <details> <summary> Click to expand </summary> ```python from transformers import AutoProcessor, SafetyChecker processor = AutoProcessor.from_pretrained("CompVis/stable-diffusion-safety-checker") safety_checker = SafetyChecker.from_pretrained("CompVis/stable-diffusion-safety-checker") ``` </details>
akmmsr/marian-finetuned-kde4-en-to-fr_akm
akmmsr
2022-11-25T17:15:54Z
59
0
transformers
[ "transformers", "tf", "marian", "text2text-generation", "generated_from_keras_callback", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-25T14:18:17Z
--- license: apache-2.0 tags: - generated_from_keras_callback model-index: - name: akmmsr/marian-finetuned-kde4-en-to-fr_akm results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # akmmsr/marian-finetuned-kde4-en-to-fr_akm This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.6861 - Validation Loss: 0.8049 - Epoch: 2 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5e-05, 'decay_steps': 17736, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01} - training_precision: mixed_float16 ### Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 1.0625 | 0.8779 | 0 | | 0.7985 | 0.8221 | 1 | | 0.6861 | 0.8049 | 2 | ### Framework versions - Transformers 4.24.0 - TensorFlow 2.9.2 - Datasets 2.7.1 - Tokenizers 0.13.2
gmsarti/setfit-ethos-multilabel-example
gmsarti
2022-11-25T16:17:05Z
5
0
sentence-transformers
[ "sentence-transformers", "pytorch", "mpnet", "feature-extraction", "sentence-similarity", "transformers", "autotrain_compatible", "endpoints_compatible", "region:us" ]
sentence-similarity
2022-11-25T16:16:56Z
--- pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity - transformers --- # {MODEL_NAME} This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. <!--- Describe your model here --> ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer sentences = ["This is an example sentence", "Each sentence is converted"] model = SentenceTransformer('{MODEL_NAME}') embeddings = model.encode(sentences) print(embeddings) ``` ## Usage (HuggingFace Transformers) Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings. ```python from transformers import AutoTokenizer, AutoModel import torch #Mean Pooling - Take attention mask into account for correct averaging def mean_pooling(model_output, attention_mask): token_embeddings = model_output[0] #First element of model_output contains all token embeddings input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float() return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9) # Sentences we want sentence embeddings for sentences = ['This is an example sentence', 'Each sentence is converted'] # Load model from HuggingFace Hub tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}') model = AutoModel.from_pretrained('{MODEL_NAME}') # Tokenize sentences encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt') # Compute token embeddings with torch.no_grad(): model_output = model(**encoded_input) # Perform pooling. In this case, mean pooling. sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask']) print("Sentence embeddings:") print(sentence_embeddings) ``` ## Evaluation Results <!--- Describe how your model was evaluated --> For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME}) ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 248 with parameters: ``` {'batch_size': 16, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss` Parameters of the fit()-Method: ``` { "epochs": 1, "evaluation_steps": 0, "evaluator": "NoneType", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": 248, "warmup_steps": 25, "weight_decay": 0.01 } ``` ## Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: MPNetModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False}) ) ``` ## Citing & Authors <!--- Describe where people can find more information -->
Hasanmurad/address_parsing_test_1
Hasanmurad
2022-11-25T16:02:33Z
116
1
transformers
[ "transformers", "pytorch", "tensorboard", "electra", "token-classification", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-11-25T15:45:40Z
--- tags: - generated_from_trainer metrics: - precision - recall - f1 - accuracy model-index: - name: address_parsing_test_1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # address_parsing_test_1 This model is a fine-tuned version of [csebuetnlp/banglabert](https://huggingface.co/csebuetnlp/banglabert) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0207 - Precision: 0.9953 - Recall: 0.9957 - F1: 0.9955 - Accuracy: 0.9963 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0669 | 1.0 | 1266 | 0.0168 | 0.9890 | 0.9913 | 0.9902 | 0.9937 | | 0.02 | 2.0 | 2532 | 0.0132 | 0.9953 | 0.9961 | 0.9957 | 0.9967 | | 0.0122 | 3.0 | 3798 | 0.0177 | 0.9949 | 0.9955 | 0.9952 | 0.9962 | | 0.0068 | 4.0 | 5064 | 0.0181 | 0.9951 | 0.9957 | 0.9954 | 0.9963 | | 0.0048 | 5.0 | 6330 | 0.0207 | 0.9953 | 0.9957 | 0.9955 | 0.9963 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
deepBiz/ddpm-butterflies-128
deepBiz
2022-11-25T15:44:35Z
0
0
diffusers
[ "diffusers", "tensorboard", "en", "dataset:handbags", "license:apache-2.0", "diffusers:DDPMPipeline", "region:us" ]
null
2022-11-24T18:40:30Z
--- language: en license: apache-2.0 library_name: diffusers tags: [] datasets: handbags metrics: [] --- <!-- This model card has been generated automatically according to the information the training script had access to. You should probably proofread and complete it, then remove this comment. --> # ddpm-butterflies-128 ## Model description This diffusion model is trained with the [🤗 Diffusers](https://github.com/huggingface/diffusers) library on the `handbags` dataset. ## Intended uses & limitations #### How to use ```python # TODO: add an example code snippet for running this diffusion pipeline ``` #### Limitations and bias [TODO: provide examples of latent issues and potential remediations] ## Training data [TODO: describe the data used to train the model] ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 16 - eval_batch_size: 16 - gradient_accumulation_steps: 1 - optimizer: AdamW with betas=(None, None), weight_decay=None and epsilon=None - lr_scheduler: None - lr_warmup_steps: 500 - ema_inv_gamma: None - ema_inv_gamma: None - ema_inv_gamma: None - mixed_precision: fp16 ### Training results 📈 [TensorBoard logs](https://huggingface.co/deepBiz/ddpm-butterflies-128/tensorboard?#scalars)
gary109/ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-rbma13-2_7k
gary109
2022-11-25T15:15:24Z
134
0
transformers
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "gary109/AI_Light_Dance", "generated_from_trainer", "dataset:ai_light_dance", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-11-25T14:54:29Z
--- tags: - automatic-speech-recognition - gary109/AI_Light_Dance - generated_from_trainer datasets: - ai_light_dance metrics: - wer model-index: - name: ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-rbma13-2_7k results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-rbma13-2_7k This model is a fine-tuned version of [gary109/ai-light-dance_drums_pretrain_wav2vec2-base-new-7k](https://huggingface.co/gary109/ai-light-dance_drums_pretrain_wav2vec2-base-new-7k) on the GARY109/AI_LIGHT_DANCE - ONSET-RBMA13-2 dataset. It achieves the following results on the evaluation set: - Loss: 2.3330 - Wer: 1.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 30 - num_epochs: 100.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:----:| | No log | 1.0 | 1 | 68.1358 | 1.0 | | No log | 2.0 | 2 | 68.1358 | 1.0 | | No log | 3.0 | 3 | 68.1358 | 1.0 | | No log | 4.0 | 4 | 68.0245 | 1.0 | | No log | 5.0 | 5 | 67.7874 | 1.0 | | No log | 6.0 | 6 | 67.4535 | 1.0 | | No log | 7.0 | 7 | 67.0142 | 1.0 | | No log | 8.0 | 8 | 67.0142 | 1.0 | | No log | 9.0 | 9 | 66.4335 | 1.0 | | 38.4011 | 10.0 | 10 | 65.7100 | 1.0 | | 38.4011 | 11.0 | 11 | 64.8206 | 1.0 | | 38.4011 | 12.0 | 12 | 63.8239 | 1.0 | | 38.4011 | 13.0 | 13 | 62.6489 | 1.0 | | 38.4011 | 14.0 | 14 | 61.3071 | 1.0 | | 38.4011 | 15.0 | 15 | 59.7427 | 1.0 | | 38.4011 | 16.0 | 16 | 58.0256 | 0.98 | | 38.4011 | 17.0 | 17 | 56.0327 | 1.0 | | 38.4011 | 18.0 | 18 | 53.7724 | 1.0 | | 38.4011 | 19.0 | 19 | 51.2556 | 1.0 | | 33.2554 | 20.0 | 20 | 48.4956 | 1.0 | | 33.2554 | 21.0 | 21 | 45.4038 | 1.0 | | 33.2554 | 22.0 | 22 | 41.9980 | 1.0 | | 33.2554 | 23.0 | 23 | 41.9980 | 1.0 | | 33.2554 | 24.0 | 24 | 38.2281 | 1.0 | | 33.2554 | 25.0 | 25 | 34.1577 | 1.0 | | 33.2554 | 26.0 | 26 | 29.7985 | 1.0 | | 33.2554 | 27.0 | 27 | 25.1146 | 1.0 | | 33.2554 | 28.0 | 28 | 20.2287 | 1.0 | | 33.2554 | 29.0 | 29 | 15.3406 | 1.0 | | 15.1206 | 30.0 | 30 | 10.7693 | 1.0 | | 15.1206 | 31.0 | 31 | 6.8998 | 1.0 | | 15.1206 | 32.0 | 32 | 4.5907 | 1.0 | | 15.1206 | 33.0 | 33 | 3.3596 | 1.0 | | 15.1206 | 34.0 | 34 | 2.7711 | 1.0 | | 15.1206 | 35.0 | 35 | 2.5962 | 1.0 | | 15.1206 | 36.0 | 36 | 2.9002 | 1.0 | | 15.1206 | 37.0 | 37 | 3.0061 | 1.0 | | 15.1206 | 38.0 | 38 | 2.8175 | 1.0 | | 15.1206 | 39.0 | 39 | 2.4512 | 1.0 | | 2.4298 | 40.0 | 40 | 2.3330 | 1.0 | | 2.4298 | 41.0 | 41 | 2.3766 | 1.0 | | 2.4298 | 42.0 | 42 | 2.5626 | 1.0 | | 2.4298 | 43.0 | 43 | 2.9632 | 1.0 | | 2.4298 | 44.0 | 44 | 3.2796 | 1.0 | | 2.4298 | 45.0 | 45 | 3.4015 | 1.0 | | 2.4298 | 46.0 | 46 | 3.2808 | 1.0 | | 2.4298 | 47.0 | 47 | 3.2373 | 1.0 | | 2.4298 | 48.0 | 48 | 3.2462 | 1.0 | | 2.4298 | 49.0 | 49 | 3.6168 | 1.0 | | 1.6143 | 50.0 | 50 | 3.6625 | 1.0 | | 1.6143 | 51.0 | 51 | 3.7593 | 1.0 | | 1.6143 | 52.0 | 52 | 3.9327 | 1.0 | | 1.6143 | 53.0 | 53 | 3.7185 | 1.0 | | 1.6143 | 54.0 | 54 | 3.9100 | 1.0 | | 1.6143 | 55.0 | 55 | 4.3123 | 1.0 | | 1.6143 | 56.0 | 56 | 4.2904 | 1.0 | | 1.6143 | 57.0 | 57 | 3.9519 | 1.0 | | 1.6143 | 58.0 | 58 | 3.4518 | 1.0 | | 1.6143 | 59.0 | 59 | 3.0197 | 1.0 | | 1.4054 | 60.0 | 60 | 2.8863 | 1.0 | | 1.4054 | 61.0 | 61 | 2.9754 | 1.0 | | 1.4054 | 62.0 | 62 | 3.2998 | 1.0 | | 1.4054 | 63.0 | 63 | 3.8715 | 1.0 | | 1.4054 | 64.0 | 64 | 4.1898 | 1.0 | | 1.4054 | 65.0 | 65 | 4.1813 | 1.0 | | 1.4054 | 66.0 | 66 | 3.9025 | 1.0 | | 1.4054 | 67.0 | 67 | 3.4319 | 1.0 | | 1.4054 | 68.0 | 68 | 3.2755 | 1.0 | | 1.4054 | 69.0 | 69 | 3.3349 | 1.0 | | 1.3121 | 70.0 | 70 | 3.5485 | 1.0 | | 1.3121 | 71.0 | 71 | 3.9019 | 1.0 | | 1.3121 | 72.0 | 72 | 4.0819 | 1.0 | | 1.3121 | 73.0 | 73 | 3.9955 | 1.0 | | 1.3121 | 74.0 | 74 | 3.7088 | 1.0 | | 1.3121 | 75.0 | 75 | 3.2957 | 1.0 | | 1.3121 | 76.0 | 76 | 3.1141 | 1.0 | | 1.3121 | 77.0 | 77 | 3.0852 | 1.0 | | 1.3121 | 78.0 | 78 | 3.1871 | 1.0 | | 1.3121 | 79.0 | 79 | 3.4127 | 1.0 | | 1.2576 | 80.0 | 80 | 3.6913 | 1.0 | | 1.2576 | 81.0 | 81 | 3.8286 | 1.0 | | 1.2576 | 82.0 | 82 | 3.8157 | 1.0 | | 1.2576 | 83.0 | 83 | 3.6814 | 1.0 | | 1.2576 | 84.0 | 84 | 3.4496 | 1.0 | | 1.2576 | 85.0 | 85 | 3.2844 | 1.0 | | 1.2576 | 86.0 | 86 | 3.2254 | 1.0 | | 1.2576 | 87.0 | 87 | 3.2683 | 1.0 | | 1.2576 | 88.0 | 88 | 3.3791 | 1.0 | | 1.2576 | 89.0 | 89 | 3.5501 | 1.0 | | 1.2373 | 90.0 | 90 | 3.6622 | 1.0 | | 1.2373 | 91.0 | 91 | 3.7207 | 1.0 | | 1.2373 | 92.0 | 92 | 3.6961 | 1.0 | | 1.2373 | 93.0 | 93 | 3.6099 | 1.0 | | 1.2373 | 94.0 | 94 | 3.5336 | 1.0 | | 1.2373 | 95.0 | 95 | 3.4342 | 1.0 | | 1.2373 | 96.0 | 96 | 3.3170 | 1.0 | | 1.2373 | 97.0 | 97 | 3.2624 | 1.0 | | 1.2373 | 98.0 | 98 | 3.2437 | 1.0 | | 1.2373 | 99.0 | 99 | 3.2591 | 1.0 | | 1.1952 | 100.0 | 100 | 3.2927 | 1.0 | ### Framework versions - Transformers 4.25.0.dev0 - Pytorch 1.8.1+cu111 - Datasets 2.7.1.dev0 - Tokenizers 0.13.2
scikit-learn/transformers-imdb
scikit-learn
2022-11-25T14:20:49Z
0
3
sklearn
[ "sklearn", "skops", "text-classification", "region:us" ]
text-classification
2022-11-25T14:10:26Z
--- library_name: sklearn tags: - sklearn - skops - text-classification --- # Model description This is a logistic regression model trained with GPT-2 embeddings on imdb dataset. The notebook to generate this model is in this repository and in this [kaggle link](https://www.kaggle.com/code/unofficialmerve/scikit-learn-with-transformers-with-skops/notebook). ## Intended uses & limitations This model is trained for educational purposes. ## Training Procedure ### Hyperparameters The model is trained with below hyperparameters. <details> <summary> Click to expand </summary> | Hyperparameter | Value | |-------------------------------|-------------------------------------------------------------------------------------------------------------------| | memory | | | steps | [('embedding', HFTransformersLanguage(model_name_or_path='facebook/bart-base')), ('model', LogisticRegression())] | | verbose | False | | embedding | HFTransformersLanguage(model_name_or_path='facebook/bart-base') | | model | LogisticRegression() | | embedding__model_name_or_path | facebook/bart-base | | model__C | 1.0 | | model__class_weight | | | model__dual | False | | model__fit_intercept | True | | model__intercept_scaling | 1 | | model__l1_ratio | | | model__max_iter | 100 | | model__multi_class | auto | | model__n_jobs | | | model__penalty | l2 | | model__random_state | | | model__solver | lbfgs | | model__tol | 0.0001 | | model__verbose | 0 | | model__warm_start | False | </details> ### Model Plot The model plot is below. <style>#sk-c17251a9-68a0-4b34-a80a-a89592893866 {color: black;background-color: white;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 pre{padding: 0;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-toggleable {background-color: white;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 label.sk-toggleable__label {cursor: pointer;display: block;width: 100%;margin-bottom: 0;padding: 0.3em;box-sizing: border-box;text-align: center;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 label.sk-toggleable__label-arrow:before {content: "▸";float: left;margin-right: 0.25em;color: #696969;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 label.sk-toggleable__label-arrow:hover:before {color: black;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-estimator:hover label.sk-toggleable__label-arrow:before {color: black;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-toggleable__content {max-height: 0;max-width: 0;overflow: hidden;text-align: left;background-color: #f0f8ff;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-toggleable__content pre {margin: 0.2em;color: black;border-radius: 0.25em;background-color: #f0f8ff;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 input.sk-toggleable__control:checked~div.sk-toggleable__content {max-height: 200px;max-width: 100%;overflow: auto;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 input.sk-toggleable__control:checked~label.sk-toggleable__label-arrow:before {content: "▾";}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-estimator input.sk-toggleable__control:checked~label.sk-toggleable__label {background-color: #d4ebff;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-label input.sk-toggleable__control:checked~label.sk-toggleable__label {background-color: #d4ebff;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 input.sk-hidden--visually {border: 0;clip: rect(1px 1px 1px 1px);clip: rect(1px, 1px, 1px, 1px);height: 1px;margin: -1px;overflow: hidden;padding: 0;position: absolute;width: 1px;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-estimator {font-family: monospace;background-color: #f0f8ff;border: 1px dotted black;border-radius: 0.25em;box-sizing: border-box;margin-bottom: 0.5em;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-estimator:hover {background-color: #d4ebff;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-parallel-item::after {content: "";width: 100%;border-bottom: 1px solid gray;flex-grow: 1;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-label:hover label.sk-toggleable__label {background-color: #d4ebff;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-serial::before {content: "";position: absolute;border-left: 1px solid gray;box-sizing: border-box;top: 2em;bottom: 0;left: 50%;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-serial {display: flex;flex-direction: column;align-items: center;background-color: white;padding-right: 0.2em;padding-left: 0.2em;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-item {z-index: 1;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-parallel {display: flex;align-items: stretch;justify-content: center;background-color: white;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-parallel::before {content: "";position: absolute;border-left: 1px solid gray;box-sizing: border-box;top: 2em;bottom: 0;left: 50%;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-parallel-item {display: flex;flex-direction: column;position: relative;background-color: white;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-parallel-item:first-child::after {align-self: flex-end;width: 50%;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-parallel-item:last-child::after {align-self: flex-start;width: 50%;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-parallel-item:only-child::after {width: 0;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-dashed-wrapped {border: 1px dashed gray;margin: 0 0.4em 0.5em 0.4em;box-sizing: border-box;padding-bottom: 0.4em;background-color: white;position: relative;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-label label {font-family: monospace;font-weight: bold;background-color: white;display: inline-block;line-height: 1.2em;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-label-container {position: relative;z-index: 2;text-align: center;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-container {/* jupyter's `normalize.less` sets `[hidden] { display: none; }` but bootstrap.min.css set `[hidden] { display: none !important; }` so we also need the `!important` here to be able to override the default hidden behavior on the sphinx rendered scikit-learn.org. See: https://github.com/scikit-learn/scikit-learn/issues/21755 */display: inline-block !important;position: relative;}#sk-c17251a9-68a0-4b34-a80a-a89592893866 div.sk-text-repr-fallback {display: none;}</style><div id="sk-c17251a9-68a0-4b34-a80a-a89592893866" class="sk-top-container"><div class="sk-text-repr-fallback"><pre>Pipeline(steps=[(&#x27;embedding&#x27;,HFTransformersLanguage(model_name_or_path=&#x27;facebook/bart-base&#x27;)),(&#x27;model&#x27;, LogisticRegression())])</pre><b>Please rerun this cell to show the HTML repr or trust the notebook.</b></div><div class="sk-container" hidden><div class="sk-item sk-dashed-wrapped"><div class="sk-label-container"><div class="sk-label sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="cae9454b-9d5c-424d-bbf8-8256c92c6733" type="checkbox" ><label for="cae9454b-9d5c-424d-bbf8-8256c92c6733" class="sk-toggleable__label sk-toggleable__label-arrow">Pipeline</label><div class="sk-toggleable__content"><pre>Pipeline(steps=[(&#x27;embedding&#x27;,HFTransformersLanguage(model_name_or_path=&#x27;facebook/bart-base&#x27;)),(&#x27;model&#x27;, LogisticRegression())])</pre></div></div></div><div class="sk-serial"><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="0e6ffcb1-dfbc-44ef-9a7c-c15d496369c7" type="checkbox" ><label for="0e6ffcb1-dfbc-44ef-9a7c-c15d496369c7" class="sk-toggleable__label sk-toggleable__label-arrow">HFTransformersLanguage</label><div class="sk-toggleable__content"><pre>HFTransformersLanguage(model_name_or_path=&#x27;facebook/bart-base&#x27;)</pre></div></div></div><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="b8eb1956-cf67-40f5-96f2-c2f0a0a41704" type="checkbox" ><label for="b8eb1956-cf67-40f5-96f2-c2f0a0a41704" class="sk-toggleable__label sk-toggleable__label-arrow">LogisticRegression</label><div class="sk-toggleable__content"><pre>LogisticRegression()</pre></div></div></div></div></div></div></div> ## Evaluation Results You can find the details about evaluation process and the evaluation results. | Metric | Value | |----------|----------| | f1_score | 0.867535 | # How to Get Started with the Model Use the code below to get started with the model. <details> <summary> Click to expand </summary> ```python [More Information Needed] ``` </details> ``` # Additional Content ## Confusion matrix ![Confusion matrix](confusion_matrix.png)
gary109/ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-idmt-2_8k
gary109
2022-11-25T13:51:19Z
138
0
transformers
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "gary109/AI_Light_Dance", "generated_from_trainer", "dataset:ai_light_dance", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-11-25T12:48:23Z
--- tags: - automatic-speech-recognition - gary109/AI_Light_Dance - generated_from_trainer datasets: - ai_light_dance metrics: - wer model-index: - name: ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-idmt-2_8k results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-idmt-2_8k This model is a fine-tuned version of [gary109/ai-light-dance_drums_pretrain_wav2vec2-base-new](https://huggingface.co/gary109/ai-light-dance_drums_pretrain_wav2vec2-base-new) on the GARY109/AI_LIGHT_DANCE - ONSET-IDMT-2 dataset. It achieves the following results on the evaluation set: - Loss: 0.5029 - Wer: 0.3178 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 30 - num_epochs: 100.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | No log | 1.0 | 9 | 101.8046 | 0.98 | | 17.4958 | 2.0 | 18 | 82.4920 | 1.0 | | 16.2087 | 3.0 | 27 | 36.1388 | 1.0 | | 6.2942 | 4.0 | 36 | 8.3267 | 1.0 | | 2.0411 | 5.0 | 45 | 6.8215 | 1.0 | | 1.554 | 6.0 | 54 | 5.3847 | 1.0 | | 1.6215 | 7.0 | 63 | 4.4645 | 1.0 | | 1.4962 | 8.0 | 72 | 3.2211 | 1.0 | | 1.3825 | 9.0 | 81 | 2.5513 | 1.0 | | 1.3443 | 10.0 | 90 | 2.8582 | 1.0 | | 1.3443 | 11.0 | 99 | 2.5446 | 1.0 | | 1.3096 | 12.0 | 108 | 2.0211 | 0.9956 | | 1.3361 | 13.0 | 117 | 1.8110 | 0.9944 | | 1.2862 | 14.0 | 126 | 1.7796 | 0.9933 | | 1.2556 | 15.0 | 135 | 1.7301 | 0.9922 | | 1.1959 | 16.0 | 144 | 1.4245 | 0.9989 | | 1.1161 | 17.0 | 153 | 1.1932 | 0.5678 | | 0.8853 | 18.0 | 162 | 1.2726 | 0.4922 | | 0.7996 | 19.0 | 171 | 1.0841 | 0.5511 | | 0.8165 | 20.0 | 180 | 1.4062 | 0.4411 | | 0.8165 | 21.0 | 189 | 1.4219 | 0.3367 | | 0.6807 | 22.0 | 198 | 1.2107 | 0.3344 | | 0.7315 | 23.0 | 207 | 1.1420 | 0.3189 | | 0.6203 | 24.0 | 216 | 1.0770 | 0.3778 | | 0.6552 | 25.0 | 225 | 1.1095 | 0.3789 | | 0.5618 | 26.0 | 234 | 1.0004 | 0.3478 | | 0.5311 | 27.0 | 243 | 0.8811 | 0.3311 | | 0.5391 | 28.0 | 252 | 0.8163 | 0.3678 | | 0.5275 | 29.0 | 261 | 1.0000 | 0.3311 | | 0.4965 | 30.0 | 270 | 0.7320 | 0.37 | | 0.4965 | 31.0 | 279 | 0.9643 | 0.3389 | | 0.4909 | 32.0 | 288 | 0.7663 | 0.3589 | | 0.5218 | 33.0 | 297 | 0.9004 | 0.3489 | | 0.4991 | 34.0 | 306 | 0.7342 | 0.38 | | 0.4883 | 35.0 | 315 | 0.7959 | 0.3389 | | 0.4902 | 36.0 | 324 | 0.6892 | 0.3378 | | 0.4447 | 37.0 | 333 | 0.6480 | 0.3333 | | 0.4458 | 38.0 | 342 | 0.6198 | 0.3333 | | 0.4607 | 39.0 | 351 | 0.6081 | 0.3111 | | 0.4352 | 40.0 | 360 | 0.6748 | 0.3156 | | 0.4352 | 41.0 | 369 | 0.6885 | 0.3256 | | 0.4286 | 42.0 | 378 | 0.6806 | 0.3333 | | 0.4314 | 43.0 | 387 | 0.7855 | 0.3222 | | 0.4476 | 44.0 | 396 | 0.6569 | 0.3144 | | 0.4815 | 45.0 | 405 | 0.5389 | 0.3033 | | 0.36 | 46.0 | 414 | 0.5550 | 0.3011 | | 0.4516 | 47.0 | 423 | 0.5924 | 0.3144 | | 0.3682 | 48.0 | 432 | 0.7275 | 0.3056 | | 0.4371 | 49.0 | 441 | 0.7051 | 0.3089 | | 0.4004 | 50.0 | 450 | 0.5669 | 0.3078 | | 0.4004 | 51.0 | 459 | 0.5029 | 0.3178 | | 0.3298 | 52.0 | 468 | 0.6150 | 0.32 | | 0.4083 | 53.0 | 477 | 0.5882 | 0.33 | | 0.4022 | 54.0 | 486 | 0.7253 | 0.3144 | | 0.4465 | 55.0 | 495 | 0.6808 | 0.3111 | | 0.3955 | 56.0 | 504 | 0.6002 | 0.3133 | | 0.3877 | 57.0 | 513 | 0.7593 | 0.3056 | | 0.3486 | 58.0 | 522 | 0.6764 | 0.3189 | | 0.3782 | 59.0 | 531 | 0.6772 | 0.3133 | | 0.3599 | 60.0 | 540 | 0.8846 | 0.3111 | | 0.3599 | 61.0 | 549 | 0.9458 | 0.3233 | | 0.3424 | 62.0 | 558 | 0.8399 | 0.3233 | | 0.3652 | 63.0 | 567 | 0.8266 | 0.3133 | | 0.3327 | 64.0 | 576 | 0.7813 | 0.3078 | | 0.3603 | 65.0 | 585 | 0.8066 | 0.3156 | | 0.3401 | 66.0 | 594 | 0.7960 | 0.3067 | | 0.3797 | 67.0 | 603 | 0.8513 | 0.2989 | | 0.3353 | 68.0 | 612 | 0.8319 | 0.2722 | | 0.3909 | 69.0 | 621 | 0.8244 | 0.2878 | | 0.3263 | 70.0 | 630 | 0.9539 | 0.3022 | | 0.3263 | 71.0 | 639 | 1.0030 | 0.2922 | | 0.3102 | 72.0 | 648 | 0.9875 | 0.3044 | | 0.3577 | 73.0 | 657 | 0.9030 | 0.2978 | | 0.2953 | 74.0 | 666 | 0.9392 | 0.2889 | | 0.3644 | 75.0 | 675 | 0.9089 | 0.2878 | | 0.3231 | 76.0 | 684 | 0.9264 | 0.2844 | | 0.3078 | 77.0 | 693 | 1.0536 | 0.2911 | | 0.4503 | 78.0 | 702 | 0.9473 | 0.2967 | | 0.3492 | 79.0 | 711 | 0.8909 | 0.3089 | | 0.347 | 80.0 | 720 | 0.8532 | 0.3067 | | 0.347 | 81.0 | 729 | 0.9553 | 0.2833 | | 0.2949 | 82.0 | 738 | 1.0111 | 0.2867 | | 0.3447 | 83.0 | 747 | 0.9160 | 0.3011 | | 0.2878 | 84.0 | 756 | 0.8401 | 0.2989 | | 0.3229 | 85.0 | 765 | 0.8815 | 0.2911 | | 0.276 | 86.0 | 774 | 0.8802 | 0.2911 | | 0.3469 | 87.0 | 783 | 0.9121 | 0.29 | | 0.3044 | 88.0 | 792 | 0.8934 | 0.2933 | | 0.2885 | 89.0 | 801 | 0.8806 | 0.2967 | | 0.3365 | 90.0 | 810 | 0.9037 | 0.2844 | | 0.3365 | 91.0 | 819 | 0.9218 | 0.2867 | | 0.3239 | 92.0 | 828 | 0.9228 | 0.2844 | | 0.3219 | 93.0 | 837 | 0.9167 | 0.2844 | | 0.2736 | 94.0 | 846 | 0.9495 | 0.2878 | | 0.3587 | 95.0 | 855 | 0.9997 | 0.2844 | | 0.3386 | 96.0 | 864 | 0.9977 | 0.2856 | | 0.2895 | 97.0 | 873 | 0.9964 | 0.2889 | | 0.3496 | 98.0 | 882 | 0.9765 | 0.2889 | | 0.2789 | 99.0 | 891 | 0.9713 | 0.2878 | | 0.3284 | 100.0 | 900 | 0.9687 | 0.2889 | ### Framework versions - Transformers 4.25.0.dev0 - Pytorch 1.8.1+cu111 - Datasets 2.7.1.dev0 - Tokenizers 0.13.2
GinaYang/xlm-roberta-base-finetuned-panx-en
GinaYang
2022-11-25T12:30:32Z
133
0
transformers
[ "transformers", "pytorch", "xlm-roberta", "token-classification", "generated_from_trainer", "dataset:xtreme", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-11-25T12:10:17Z
--- license: mit tags: - generated_from_trainer datasets: - xtreme metrics: - f1 model-index: - name: xlm-roberta-base-finetuned-panx-en results: - task: name: Token Classification type: token-classification dataset: name: xtreme type: xtreme config: PAN-X.en split: train args: PAN-X.en metrics: - name: F1 type: f1 value: 0.6886160714285715 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-panx-en This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the xtreme dataset. It achieves the following results on the evaluation set: - Loss: 0.4043 - F1: 0.6886 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 24 - eval_batch_size: 24 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 1.1347 | 1.0 | 50 | 0.5771 | 0.4880 | | 0.5066 | 2.0 | 100 | 0.4209 | 0.6582 | | 0.3631 | 3.0 | 150 | 0.4043 | 0.6886 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
ConvLab/t5-small-nlu-multiwoz21_sgd_tm1_tm2_tm3
ConvLab
2022-11-25T12:19:34Z
131
0
transformers
[ "transformers", "pytorch", "t5", "text2text-generation", "t5-small", "natural language understanding", "conversational system", "task-oriented dialog", "en", "dataset:ConvLab/multiwoz21", "dataset:ConvLab/sgd", "dataset:ConvLab/tm1", "dataset:ConvLab/tm2", "dataset:ConvLab/tm3", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-25T04:30:10Z
--- language: - en license: apache-2.0 tags: - t5-small - text2text-generation - natural language understanding - conversational system - task-oriented dialog datasets: - ConvLab/multiwoz21 - ConvLab/sgd - ConvLab/tm1 - ConvLab/tm2 - ConvLab/tm3 metrics: - Slot Error Rate - sacrebleu model-index: - name: t5-small-nlu-multiwoz21_sgd_tm1_tm2_tm3 results: - task: type: text2text-generation name: natural language understanding dataset: type: ConvLab/multiwoz21 name: MultiWOZ 2.1 split: test revision: 5f55375edbfe0270c20bcf770751ad982c0e6614 metrics: - type: Dialog acts Accuracy value: 77.5 name: Accuracy - type: Dialog acts F1 value: 86.4 name: F1 - task: type: text2text-generation name: natural language understanding dataset: type: ConvLab/sgd name: SGD split: test revision: 6e8c79b888b21cc658cf9c0ce128d263241cf70f metrics: - type: Dialog acts Accuracy value: 45.2 name: Accuracy - type: Dialog acts F1 value: 58.6 name: F1 - task: type: text2text-generation name: natural language understanding dataset: type: ConvLab/tm1, ConvLab/tm2, ConvLab/tm3 name: TM1+TM2+TM3 split: test metrics: - type: Dialog acts Accuracy value: 81.8 name: Accuracy - type: Dialog acts F1 value: 73.0 name: F1 widget: - text: "multiwoz21: user: I would like a taxi from Saint John's college to Pizza Hut Fen Ditton." example_title: "MultiWOZ 2.1" - text: "sgd: user: Could you get me a reservation at P.f. Chang's in Corte Madera at afternoon 12?" example_title: "Schema-Guided Dialog" - text: "tm1: user: I would like to order a pizza from Domino's." example_title: "Taskmaster-1" - text: "tm2: user: I would like help getting a flight from LA to Amsterdam." example_title: "Taskmaster-2" - text: "tm3: user: Well, I need a kids friendly movie. I was thinking about seeing Mulan." example_title: "Taskmaster-3" inference: parameters: max_length: 100 --- # t5-small-nlu-multiwoz21_sgd_tm1_tm2_tm3 This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on [MultiWOZ 2.1](https://huggingface.co/datasets/ConvLab/multiwoz21), [Schema-Guided Dialog](https://huggingface.co/datasets/ConvLab/sgd), [Taskmaster-1](https://huggingface.co/datasets/ConvLab/tm1), [Taskmaster-2](https://huggingface.co/datasets/ConvLab/tm2), and [Taskmaster-3](https://huggingface.co/datasets/ConvLab/tm3). Refer to [ConvLab-3](https://github.com/ConvLab/ConvLab-3) for model description and usage. ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 128 - eval_batch_size: 64 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 256 - optimizer: Adafactor - lr_scheduler_type: linear - num_epochs: 10.0 ### Framework versions - Transformers 4.18.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.3 - Tokenizers 0.11.0
GinaYang/xlm-roberta-base-finetuned-panx-it
GinaYang
2022-11-25T12:10:00Z
108
0
transformers
[ "transformers", "pytorch", "xlm-roberta", "token-classification", "generated_from_trainer", "dataset:xtreme", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-11-25T11:49:15Z
--- license: mit tags: - generated_from_trainer datasets: - xtreme metrics: - f1 model-index: - name: xlm-roberta-base-finetuned-panx-it results: - task: name: Token Classification type: token-classification dataset: name: xtreme type: xtreme config: PAN-X.it split: train args: PAN-X.it metrics: - name: F1 type: f1 value: 0.8124233755619126 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-panx-it This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the xtreme dataset. It achieves the following results on the evaluation set: - Loss: 0.2630 - F1: 0.8124 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 24 - eval_batch_size: 24 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.8193 | 1.0 | 70 | 0.3200 | 0.7356 | | 0.2773 | 2.0 | 140 | 0.2841 | 0.7882 | | 0.1807 | 3.0 | 210 | 0.2630 | 0.8124 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
ConvLab/t5-small-dst-sgd
ConvLab
2022-11-25T12:09:28Z
102
0
transformers
[ "transformers", "pytorch", "t5", "text2text-generation", "t5-small", "dialog state tracking", "conversational system", "task-oriented dialog", "en", "dataset:ConvLab/sgd", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-25T07:33:55Z
--- language: - en license: apache-2.0 tags: - t5-small - text2text-generation - dialog state tracking - conversational system - task-oriented dialog datasets: - ConvLab/sgd metrics: - Joint Goal Accuracy - Slot F1 model-index: - name: t5-small-dst-sgd results: - task: type: text2text-generation name: dialog state tracking dataset: type: ConvLab/sgd name: SGD split: test revision: 6e8c79b888b21cc658cf9c0ce128d263241cf70f metrics: - type: Joint Goal Accuracy value: 20.1 name: JGA - type: Slot F1 value: 58.5 name: Slot F1 widget: - text: "user: Hi, could you get me a restaurant booking on the 8th please?\nsystem: Any preference on the restaurant, location and time?\nuser: Could you get me a reservation at P.f. Chang's in Corte Madera at afternoon 12?" - text: "user: I need to book a dinner reservation for a date. Help me reserve a table at a restaurant.\nsystem: What time and location do you have in mind?\nuser: Something around 8 in the night should be fine. Oh, and look in the San Jose area." inference: parameters: max_length: 100 --- # t5-small-dst-sgd This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on [Schema-Guided Dialog](https://huggingface.co/datasets/ConvLab/sgd). Refer to [ConvLab-3](https://github.com/ConvLab/ConvLab-3) for model description and usage. ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 128 - optimizer: Adafactor - lr_scheduler_type: linear - num_epochs: 10.0 ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0+cu113 - Datasets 2.3.2 - Tokenizers 0.12.1
Sotaro0124/ai-ja
Sotaro0124
2022-11-25T11:40:01Z
0
0
null
[ "translation", "ja", "ai", "region:us" ]
translation
2022-11-10T07:09:39Z
--- language: - ja - ai tags: - translation widget: - text: "ari hawki = an konno" example_title: "と 言う と " ---
gary109/ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-idmt-mdb-enst-2
gary109
2022-11-25T11:30:51Z
134
0
transformers
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "gary109/AI_Light_Dance", "generated_from_trainer", "dataset:ai_light_dance", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-11-25T08:01:54Z
--- tags: - automatic-speech-recognition - gary109/AI_Light_Dance - generated_from_trainer datasets: - ai_light_dance metrics: - wer model-index: - name: ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-idmt-mdb-enst-2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-idmt-mdb-enst-2 This model is a fine-tuned version of [gary109/ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-idmt-mdb-2](https://huggingface.co/gary109/ai-light-dance_drums_ft_pretrain_wav2vec2-base-new_onset-idmt-mdb-2) on the GARY109/AI_LIGHT_DANCE - ONSET-IDMT-MDB-ENST-2 dataset. It achieves the following results on the evaluation set: - Loss: 0.6379 - Wer: 0.3792 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 30 - num_epochs: 100.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 1.4226 | 0.99 | 35 | 2.0435 | 0.4154 | | 0.8744 | 1.99 | 70 | 1.7193 | 0.4382 | | 0.9474 | 2.99 | 105 | 1.7853 | 0.4374 | | 0.8316 | 3.99 | 140 | 1.2827 | 0.4306 | | 0.8336 | 4.99 | 175 | 1.0676 | 0.4040 | | 0.7345 | 5.99 | 210 | 1.5364 | 0.4264 | | 0.6666 | 6.99 | 245 | 1.4284 | 0.4585 | | 0.6677 | 7.99 | 280 | 0.9475 | 0.4003 | | 0.6779 | 8.99 | 315 | 1.1172 | 0.4209 | | 0.6503 | 9.99 | 350 | 0.8999 | 0.3834 | | 0.6159 | 10.99 | 385 | 1.1501 | 0.4386 | | 0.6831 | 11.99 | 420 | 1.0860 | 0.3825 | | 0.5959 | 12.99 | 455 | 0.9410 | 0.4045 | | 0.7154 | 13.99 | 490 | 1.0463 | 0.3821 | | 0.6094 | 14.99 | 525 | 0.8598 | 0.3965 | | 0.6929 | 15.99 | 560 | 0.9494 | 0.3931 | | 0.7627 | 16.99 | 595 | 0.8060 | 0.3948 | | 0.601 | 17.99 | 630 | 0.9890 | 0.3965 | | 0.546 | 18.99 | 665 | 0.8059 | 0.3990 | | 0.5222 | 19.99 | 700 | 0.6379 | 0.3792 | | 0.5802 | 20.99 | 735 | 0.6995 | 0.3661 | | 0.5731 | 21.99 | 770 | 0.8405 | 0.3606 | | 0.5462 | 22.99 | 805 | 0.6667 | 0.3965 | | 0.6057 | 23.99 | 840 | 0.8396 | 0.3762 | | 0.5323 | 24.99 | 875 | 0.9054 | 0.3952 | | 0.683 | 25.99 | 910 | 0.6898 | 0.4062 | | 0.525 | 26.99 | 945 | 0.7245 | 0.3884 | | 0.4885 | 27.99 | 980 | 0.8076 | 0.4049 | | 0.4653 | 28.99 | 1015 | 0.8100 | 0.3838 | | 0.4827 | 29.99 | 1050 | 0.7247 | 0.3863 | | 0.4839 | 30.99 | 1085 | 0.7009 | 0.3817 | | 0.4982 | 31.99 | 1120 | 0.7637 | 0.3914 | | 0.6105 | 32.99 | 1155 | 0.7343 | 0.3914 | | 0.4936 | 33.99 | 1190 | 0.7390 | 0.3762 | | 0.4674 | 34.99 | 1225 | 0.6724 | 0.3581 | | 0.4677 | 35.99 | 1260 | 0.6730 | 0.3488 | | 0.516 | 36.99 | 1295 | 0.6956 | 0.3728 | | 0.4507 | 37.99 | 1330 | 0.6483 | 0.3615 | | 0.4207 | 38.99 | 1365 | 0.7718 | 0.3484 | | 0.4803 | 39.99 | 1400 | 0.8316 | 0.3775 | | 0.3946 | 40.99 | 1435 | 0.8322 | 0.3568 | | 0.411 | 41.99 | 1470 | 0.9933 | 0.3707 | | 0.4405 | 42.99 | 1505 | 0.8789 | 0.3943 | | 0.5124 | 43.99 | 1540 | 0.9030 | 0.3707 | | 0.5959 | 44.99 | 1575 | 0.7809 | 0.3948 | | 0.3841 | 45.99 | 1610 | 0.7716 | 0.3965 | | 0.3975 | 46.99 | 1645 | 0.7064 | 0.3931 | | 1.4983 | 47.99 | 1680 | 3.2866 | 0.3627 | | 0.3962 | 48.99 | 1715 | 0.6486 | 0.3648 | | 0.4422 | 49.99 | 1750 | 0.8450 | 0.3779 | | 0.4198 | 50.99 | 1785 | 0.7628 | 0.3564 | | 0.3577 | 51.99 | 1820 | 0.7553 | 0.3678 | | 0.4425 | 52.99 | 1855 | 0.7566 | 0.3716 | | 0.3492 | 53.99 | 1890 | 0.7710 | 0.3631 | | 0.3731 | 54.99 | 1925 | 0.7737 | 0.3627 | | 0.3868 | 55.99 | 1960 | 0.7021 | 0.3572 | | 0.3311 | 56.99 | 1995 | 0.6603 | 0.3518 | | 0.3993 | 57.99 | 2030 | 0.6664 | 0.3581 | | 0.4226 | 58.99 | 2065 | 0.6813 | 0.3551 | | 0.4143 | 59.99 | 2100 | 0.6567 | 0.3568 | | 0.3623 | 60.99 | 2135 | 0.6568 | 0.3454 | | 0.3228 | 61.99 | 2170 | 0.7326 | 0.3568 | | 0.3204 | 62.99 | 2205 | 0.7277 | 0.3640 | | 0.377 | 63.99 | 2240 | 0.7145 | 0.3585 | | 0.3487 | 64.99 | 2275 | 0.6943 | 0.3505 | | 0.343 | 65.99 | 2310 | 0.7461 | 0.3395 | | 0.3251 | 66.99 | 2345 | 0.7442 | 0.3564 | | 0.3135 | 67.99 | 2380 | 0.7331 | 0.3530 | | 0.381 | 68.99 | 2415 | 0.7306 | 0.3513 | | 0.3319 | 69.99 | 2450 | 0.8495 | 0.3484 | | 0.3552 | 70.99 | 2485 | 0.7546 | 0.3551 | | 0.3292 | 71.99 | 2520 | 0.7483 | 0.3450 | | 0.3041 | 72.99 | 2555 | 0.7305 | 0.3522 | | 0.3606 | 73.99 | 2590 | 0.7358 | 0.3484 | | 0.3629 | 74.99 | 2625 | 0.7709 | 0.3446 | | 0.3409 | 75.99 | 2660 | 0.7568 | 0.3585 | | 0.3315 | 76.99 | 2695 | 0.7466 | 0.3475 | | 0.2934 | 77.99 | 2730 | 0.7351 | 0.3496 | | 0.3366 | 78.99 | 2765 | 0.8014 | 0.3484 | | 0.3176 | 79.99 | 2800 | 0.8014 | 0.3420 | | 0.3319 | 80.99 | 2835 | 0.7996 | 0.3437 | | 0.2967 | 81.99 | 2870 | 0.8156 | 0.3412 | | 0.3137 | 82.99 | 2905 | 0.8025 | 0.3361 | | 0.3133 | 83.99 | 2940 | 0.7784 | 0.3416 | | 0.3134 | 84.99 | 2975 | 0.7894 | 0.3336 | | 0.3216 | 85.99 | 3010 | 0.8331 | 0.3395 | | 0.365 | 86.99 | 3045 | 0.7980 | 0.3353 | | 0.2962 | 87.99 | 3080 | 0.7965 | 0.3404 | | 0.3126 | 88.99 | 3115 | 0.7470 | 0.3420 | | 0.2843 | 89.99 | 3150 | 0.7788 | 0.3404 | | 0.2967 | 90.99 | 3185 | 0.7902 | 0.3374 | | 0.3171 | 91.99 | 3220 | 0.8022 | 0.3404 | | 0.3069 | 92.99 | 3255 | 0.7999 | 0.3345 | | 0.3571 | 93.99 | 3290 | 0.7896 | 0.3404 | | 0.2805 | 94.99 | 3325 | 0.7831 | 0.3391 | | 0.3099 | 95.99 | 3360 | 0.7909 | 0.3366 | | 0.2868 | 96.99 | 3395 | 0.7918 | 0.3395 | | 0.2626 | 97.99 | 3430 | 0.7766 | 0.3429 | | 0.2634 | 98.99 | 3465 | 0.7770 | 0.3437 | | 0.288 | 99.99 | 3500 | 0.7762 | 0.3429 | ### Framework versions - Transformers 4.25.0.dev0 - Pytorch 1.8.1+cu111 - Datasets 2.7.1.dev0 - Tokenizers 0.13.2
ConvLab/t5-small-nlu-tm1_tm2_tm3
ConvLab
2022-11-25T11:26:12Z
101
0
transformers
[ "transformers", "pytorch", "t5", "text2text-generation", "t5-small", "natural language understanding", "conversational system", "task-oriented dialog", "en", "dataset:ConvLab/tm1", "dataset:ConvLab/tm2", "dataset:ConvLab/tm3", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-25T04:29:07Z
--- language: - en license: apache-2.0 tags: - t5-small - text2text-generation - natural language understanding - conversational system - task-oriented dialog datasets: - ConvLab/tm1 - ConvLab/tm2 - ConvLab/tm3 metrics: - Dialog acts Accuracy - Dialog acts F1 model-index: - name: t5-small-nlu-tm1_tm2_tm3 results: - task: type: text2text-generation name: natural language understanding dataset: type: ConvLab/tm1, ConvLab/tm2, ConvLab/tm3 name: TM1+TM2+TM3 split: test metrics: - type: Dialog acts Accuracy value: 81.8 name: Accuracy - type: Dialog acts F1 value: 73.0 name: F1 widget: - text: "tm1: user: I would like to order a pizza from Domino's." - text: "tm2: user: I would like help getting a flight from LA to Amsterdam." - text: "tm3: user: Well, I need a kids friendly movie. I was thinking about seeing Mulan." inference: parameters: max_length: 100 --- # t5-small-nlu-tm1_tm2_tm3 This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on [Taskmaster-1](https://huggingface.co/datasets/ConvLab/tm1), [Taskmaster-2](https://huggingface.co/datasets/ConvLab/tm2), and [Taskmaster-3](https://huggingface.co/datasets/ConvLab/tm3). Refer to [ConvLab-3](https://github.com/ConvLab/ConvLab-3) for model description and usage. ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 128 - eval_batch_size: 64 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 256 - optimizer: Adafactor - lr_scheduler_type: linear - num_epochs: 10.0 ### Framework versions - Transformers 4.18.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.3 - Tokenizers 0.11.0
GinaYang/xlr-roberta-base-finetuned-panx-de-fr
GinaYang
2022-11-25T11:25:46Z
109
0
transformers
[ "transformers", "pytorch", "xlm-roberta", "token-classification", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-11-25T09:06:16Z
--- license: mit tags: - generated_from_trainer metrics: - f1 model-index: - name: xlr-roberta-base-finetuned-panx-de-fr results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlr-roberta-base-finetuned-panx-de-fr This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.1608 - F1: 0.8593 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 24 - eval_batch_size: 24 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.2888 | 1.0 | 715 | 0.1779 | 0.8233 | | 0.1437 | 2.0 | 1430 | 0.1570 | 0.8497 | | 0.0931 | 3.0 | 2145 | 0.1608 | 0.8593 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
ConvLab/t5-small-nlu-sgd
ConvLab
2022-11-25T11:18:08Z
103
0
transformers
[ "transformers", "pytorch", "t5", "text2text-generation", "t5-small", "natural language understanding", "conversational system", "task-oriented dialog", "en", "dataset:ConvLab/sgd", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-25T04:26:44Z
--- language: - en license: apache-2.0 tags: - t5-small - text2text-generation - natural language understanding - conversational system - task-oriented dialog datasets: - ConvLab/sgd metrics: - Dialog acts Accuracy - Dialog acts F1 model-index: - name: t5-small-nlu-sgd results: - task: type: text2text-generation name: natural language understanding dataset: type: ConvLab/sgd name: SGD split: test revision: 6e8c79b888b21cc658cf9c0ce128d263241cf70f metrics: - type: Dialog acts Accuracy value: 45.0 name: Accuracy - type: Dialog acts F1 value: 58.6 name: F1 widget: - text: "user: Could you get me a reservation at P.f. Chang's in Corte Madera at afternoon 12?" - text: "user: Sure, may I know if they have vegetarian options and how expensive is their food?" inference: parameters: max_length: 100 --- # t5-small-nlu-sgd This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on [Schema-Guided Dialog](https://huggingface.co/datasets/ConvLab/sgd). Refer to [ConvLab-3](https://github.com/ConvLab/ConvLab-3) for model description and usage. ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 128 - eval_batch_size: 64 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 256 - optimizer: Adafactor - lr_scheduler_type: linear - num_epochs: 10.0 ### Framework versions - Transformers 4.18.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.3 - Tokenizers 0.11.0
channotte/gpt2-Georges-sand
channotte
2022-11-25T10:40:45Z
59
1
transformers
[ "transformers", "tf", "gpt2", "text-generation", "generated_from_keras_callback", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2022-11-25T10:10:54Z
--- license: mit tags: - generated_from_keras_callback model-index: - name: gpt2-Georges-sand results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # gpt2-Georges-sand This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 10.8530 - Validation Loss: 10.7406 - Epoch: 2 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'WarmUp', 'config': {'initial_learning_rate': 5e-05, 'decay_schedule_fn': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5e-05, 'decay_steps': -993, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, '__passive_serialization__': True}, 'warmup_steps': 1000, 'power': 1.0, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01} - training_precision: mixed_float16 ### Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 10.9815 | 10.9683 | 0 | | 10.9422 | 10.8815 | 1 | | 10.8530 | 10.7406 | 2 | ### Framework versions - Transformers 4.24.0 - TensorFlow 2.10.0 - Datasets 2.7.0 - Tokenizers 0.13.2
tomekkorbak/frosty_lamport
tomekkorbak
2022-11-25T10:28:16Z
0
0
null
[ "generated_from_trainer", "en", "dataset:tomekkorbak/detoxify-pile-chunk3-0-50000", "dataset:tomekkorbak/detoxify-pile-chunk3-50000-100000", "dataset:tomekkorbak/detoxify-pile-chunk3-100000-150000", "dataset:tomekkorbak/detoxify-pile-chunk3-150000-200000", "dataset:tomekkorbak/detoxify-pile-chunk3-200000-250000", "dataset:tomekkorbak/detoxify-pile-chunk3-250000-300000", "dataset:tomekkorbak/detoxify-pile-chunk3-300000-350000", "dataset:tomekkorbak/detoxify-pile-chunk3-350000-400000", "dataset:tomekkorbak/detoxify-pile-chunk3-400000-450000", "dataset:tomekkorbak/detoxify-pile-chunk3-450000-500000", "dataset:tomekkorbak/detoxify-pile-chunk3-500000-550000", "dataset:tomekkorbak/detoxify-pile-chunk3-550000-600000", "dataset:tomekkorbak/detoxify-pile-chunk3-600000-650000", "dataset:tomekkorbak/detoxify-pile-chunk3-650000-700000", "dataset:tomekkorbak/detoxify-pile-chunk3-700000-750000", "dataset:tomekkorbak/detoxify-pile-chunk3-750000-800000", "dataset:tomekkorbak/detoxify-pile-chunk3-800000-850000", "dataset:tomekkorbak/detoxify-pile-chunk3-850000-900000", "dataset:tomekkorbak/detoxify-pile-chunk3-900000-950000", "dataset:tomekkorbak/detoxify-pile-chunk3-950000-1000000", "dataset:tomekkorbak/detoxify-pile-chunk3-1000000-1050000", "dataset:tomekkorbak/detoxify-pile-chunk3-1050000-1100000", "dataset:tomekkorbak/detoxify-pile-chunk3-1100000-1150000", "dataset:tomekkorbak/detoxify-pile-chunk3-1150000-1200000", "dataset:tomekkorbak/detoxify-pile-chunk3-1200000-1250000", "dataset:tomekkorbak/detoxify-pile-chunk3-1250000-1300000", "dataset:tomekkorbak/detoxify-pile-chunk3-1300000-1350000", "dataset:tomekkorbak/detoxify-pile-chunk3-1350000-1400000", "dataset:tomekkorbak/detoxify-pile-chunk3-1400000-1450000", "dataset:tomekkorbak/detoxify-pile-chunk3-1450000-1500000", "dataset:tomekkorbak/detoxify-pile-chunk3-1500000-1550000", "dataset:tomekkorbak/detoxify-pile-chunk3-1550000-1600000", "dataset:tomekkorbak/detoxify-pile-chunk3-1600000-1650000", "dataset:tomekkorbak/detoxify-pile-chunk3-1650000-1700000", "dataset:tomekkorbak/detoxify-pile-chunk3-1700000-1750000", "dataset:tomekkorbak/detoxify-pile-chunk3-1750000-1800000", "dataset:tomekkorbak/detoxify-pile-chunk3-1800000-1850000", "dataset:tomekkorbak/detoxify-pile-chunk3-1850000-1900000", "dataset:tomekkorbak/detoxify-pile-chunk3-1900000-1950000", "license:mit", "region:us" ]
null
2022-11-25T10:28:04Z
--- language: - en license: mit tags: - generated_from_trainer datasets: - tomekkorbak/detoxify-pile-chunk3-0-50000 - tomekkorbak/detoxify-pile-chunk3-50000-100000 - tomekkorbak/detoxify-pile-chunk3-100000-150000 - tomekkorbak/detoxify-pile-chunk3-150000-200000 - tomekkorbak/detoxify-pile-chunk3-200000-250000 - tomekkorbak/detoxify-pile-chunk3-250000-300000 - tomekkorbak/detoxify-pile-chunk3-300000-350000 - tomekkorbak/detoxify-pile-chunk3-350000-400000 - tomekkorbak/detoxify-pile-chunk3-400000-450000 - tomekkorbak/detoxify-pile-chunk3-450000-500000 - tomekkorbak/detoxify-pile-chunk3-500000-550000 - tomekkorbak/detoxify-pile-chunk3-550000-600000 - tomekkorbak/detoxify-pile-chunk3-600000-650000 - tomekkorbak/detoxify-pile-chunk3-650000-700000 - tomekkorbak/detoxify-pile-chunk3-700000-750000 - tomekkorbak/detoxify-pile-chunk3-750000-800000 - tomekkorbak/detoxify-pile-chunk3-800000-850000 - tomekkorbak/detoxify-pile-chunk3-850000-900000 - tomekkorbak/detoxify-pile-chunk3-900000-950000 - tomekkorbak/detoxify-pile-chunk3-950000-1000000 - tomekkorbak/detoxify-pile-chunk3-1000000-1050000 - tomekkorbak/detoxify-pile-chunk3-1050000-1100000 - tomekkorbak/detoxify-pile-chunk3-1100000-1150000 - tomekkorbak/detoxify-pile-chunk3-1150000-1200000 - tomekkorbak/detoxify-pile-chunk3-1200000-1250000 - tomekkorbak/detoxify-pile-chunk3-1250000-1300000 - tomekkorbak/detoxify-pile-chunk3-1300000-1350000 - tomekkorbak/detoxify-pile-chunk3-1350000-1400000 - tomekkorbak/detoxify-pile-chunk3-1400000-1450000 - tomekkorbak/detoxify-pile-chunk3-1450000-1500000 - tomekkorbak/detoxify-pile-chunk3-1500000-1550000 - tomekkorbak/detoxify-pile-chunk3-1550000-1600000 - tomekkorbak/detoxify-pile-chunk3-1600000-1650000 - tomekkorbak/detoxify-pile-chunk3-1650000-1700000 - tomekkorbak/detoxify-pile-chunk3-1700000-1750000 - tomekkorbak/detoxify-pile-chunk3-1750000-1800000 - tomekkorbak/detoxify-pile-chunk3-1800000-1850000 - tomekkorbak/detoxify-pile-chunk3-1850000-1900000 - tomekkorbak/detoxify-pile-chunk3-1900000-1950000 model-index: - name: frosty_lamport results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # frosty_lamport This model was trained from scratch on the tomekkorbak/detoxify-pile-chunk3-0-50000, the tomekkorbak/detoxify-pile-chunk3-50000-100000, the tomekkorbak/detoxify-pile-chunk3-100000-150000, the tomekkorbak/detoxify-pile-chunk3-150000-200000, the tomekkorbak/detoxify-pile-chunk3-200000-250000, the tomekkorbak/detoxify-pile-chunk3-250000-300000, the tomekkorbak/detoxify-pile-chunk3-300000-350000, the tomekkorbak/detoxify-pile-chunk3-350000-400000, the tomekkorbak/detoxify-pile-chunk3-400000-450000, the tomekkorbak/detoxify-pile-chunk3-450000-500000, the tomekkorbak/detoxify-pile-chunk3-500000-550000, the tomekkorbak/detoxify-pile-chunk3-550000-600000, the tomekkorbak/detoxify-pile-chunk3-600000-650000, the tomekkorbak/detoxify-pile-chunk3-650000-700000, the tomekkorbak/detoxify-pile-chunk3-700000-750000, the tomekkorbak/detoxify-pile-chunk3-750000-800000, the tomekkorbak/detoxify-pile-chunk3-800000-850000, the tomekkorbak/detoxify-pile-chunk3-850000-900000, the tomekkorbak/detoxify-pile-chunk3-900000-950000, the tomekkorbak/detoxify-pile-chunk3-950000-1000000, the tomekkorbak/detoxify-pile-chunk3-1000000-1050000, the tomekkorbak/detoxify-pile-chunk3-1050000-1100000, the tomekkorbak/detoxify-pile-chunk3-1100000-1150000, the tomekkorbak/detoxify-pile-chunk3-1150000-1200000, the tomekkorbak/detoxify-pile-chunk3-1200000-1250000, the tomekkorbak/detoxify-pile-chunk3-1250000-1300000, the tomekkorbak/detoxify-pile-chunk3-1300000-1350000, the tomekkorbak/detoxify-pile-chunk3-1350000-1400000, the tomekkorbak/detoxify-pile-chunk3-1400000-1450000, the tomekkorbak/detoxify-pile-chunk3-1450000-1500000, the tomekkorbak/detoxify-pile-chunk3-1500000-1550000, the tomekkorbak/detoxify-pile-chunk3-1550000-1600000, the tomekkorbak/detoxify-pile-chunk3-1600000-1650000, the tomekkorbak/detoxify-pile-chunk3-1650000-1700000, the tomekkorbak/detoxify-pile-chunk3-1700000-1750000, the tomekkorbak/detoxify-pile-chunk3-1750000-1800000, the tomekkorbak/detoxify-pile-chunk3-1800000-1850000, the tomekkorbak/detoxify-pile-chunk3-1850000-1900000 and the tomekkorbak/detoxify-pile-chunk3-1900000-1950000 datasets. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 64 - total_train_batch_size: 1024 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.01 - training_steps: 3147 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0+cu113 - Datasets 2.5.1 - Tokenizers 0.11.6 # Full config {'dataset': {'datasets': ['tomekkorbak/detoxify-pile-chunk3-0-50000', 'tomekkorbak/detoxify-pile-chunk3-50000-100000', 'tomekkorbak/detoxify-pile-chunk3-100000-150000', 'tomekkorbak/detoxify-pile-chunk3-150000-200000', 'tomekkorbak/detoxify-pile-chunk3-200000-250000', 'tomekkorbak/detoxify-pile-chunk3-250000-300000', 'tomekkorbak/detoxify-pile-chunk3-300000-350000', 'tomekkorbak/detoxify-pile-chunk3-350000-400000', 'tomekkorbak/detoxify-pile-chunk3-400000-450000', 'tomekkorbak/detoxify-pile-chunk3-450000-500000', 'tomekkorbak/detoxify-pile-chunk3-500000-550000', 'tomekkorbak/detoxify-pile-chunk3-550000-600000', 'tomekkorbak/detoxify-pile-chunk3-600000-650000', 'tomekkorbak/detoxify-pile-chunk3-650000-700000', 'tomekkorbak/detoxify-pile-chunk3-700000-750000', 'tomekkorbak/detoxify-pile-chunk3-750000-800000', 'tomekkorbak/detoxify-pile-chunk3-800000-850000', 'tomekkorbak/detoxify-pile-chunk3-850000-900000', 'tomekkorbak/detoxify-pile-chunk3-900000-950000', 'tomekkorbak/detoxify-pile-chunk3-950000-1000000', 'tomekkorbak/detoxify-pile-chunk3-1000000-1050000', 'tomekkorbak/detoxify-pile-chunk3-1050000-1100000', 'tomekkorbak/detoxify-pile-chunk3-1100000-1150000', 'tomekkorbak/detoxify-pile-chunk3-1150000-1200000', 'tomekkorbak/detoxify-pile-chunk3-1200000-1250000', 'tomekkorbak/detoxify-pile-chunk3-1250000-1300000', 'tomekkorbak/detoxify-pile-chunk3-1300000-1350000', 'tomekkorbak/detoxify-pile-chunk3-1350000-1400000', 'tomekkorbak/detoxify-pile-chunk3-1400000-1450000', 'tomekkorbak/detoxify-pile-chunk3-1450000-1500000', 'tomekkorbak/detoxify-pile-chunk3-1500000-1550000', 'tomekkorbak/detoxify-pile-chunk3-1550000-1600000', 'tomekkorbak/detoxify-pile-chunk3-1600000-1650000', 'tomekkorbak/detoxify-pile-chunk3-1650000-1700000', 'tomekkorbak/detoxify-pile-chunk3-1700000-1750000', 'tomekkorbak/detoxify-pile-chunk3-1750000-1800000', 'tomekkorbak/detoxify-pile-chunk3-1800000-1850000', 'tomekkorbak/detoxify-pile-chunk3-1850000-1900000', 'tomekkorbak/detoxify-pile-chunk3-1900000-1950000'], 'is_split_by_sentences': True}, 'generation': {'force_call_on': [25354], 'metrics_configs': [{}, {'n': 1}, {'n': 2}, {'n': 5}], 'scenario_configs': [{'generate_kwargs': {'do_sample': True, 'max_length': 128, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'unconditional', 'num_samples': 2048}, {'generate_kwargs': {'do_sample': True, 'max_length': 128, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'challenging_rtp', 'num_samples': 2048, 'prompts_path': 'resources/challenging_rtp.jsonl'}], 'scorer_config': {'device': 'cuda:0'}}, 'kl_gpt3_callback': {'force_call_on': [25354], 'max_tokens': 64, 'num_samples': 4096}, 'model': {'from_scratch': True, 'gpt2_config_kwargs': {'reorder_and_upcast_attn': True, 'scale_attn_by': True}, 'model_kwargs': {'value_head_config': {'is_detached': False}}, 'path_or_name': 'gpt2'}, 'objective': {'alpha': 1, 'beta': 10, 'name': 'AWR'}, 'tokenizer': {'path_or_name': 'gpt2'}, 'training': {'dataloader_num_workers': 0, 'effective_batch_size': 1024, 'evaluation_strategy': 'no', 'fp16': True, 'hub_model_id': 'frosty_lamport', 'hub_strategy': 'all_checkpoints', 'learning_rate': 0.0005, 'logging_first_step': True, 'logging_steps': 1, 'num_tokens': 3300000000, 'output_dir': 'training_output104340', 'per_device_train_batch_size': 16, 'push_to_hub': True, 'remove_unused_columns': False, 'save_steps': 25354, 'save_strategy': 'steps', 'seed': 42, 'warmup_ratio': 0.01, 'weight_decay': 0.1}} # Wandb URL: https://wandb.ai/tomekkorbak/apo/runs/3qaspcth
tomekkorbak/gallant_thompson
tomekkorbak
2022-11-25T10:19:23Z
0
0
null
[ "generated_from_trainer", "en", "dataset:tomekkorbak/detoxify-pile-chunk3-0-50000", "dataset:tomekkorbak/detoxify-pile-chunk3-50000-100000", "dataset:tomekkorbak/detoxify-pile-chunk3-100000-150000", "dataset:tomekkorbak/detoxify-pile-chunk3-150000-200000", "dataset:tomekkorbak/detoxify-pile-chunk3-200000-250000", "dataset:tomekkorbak/detoxify-pile-chunk3-250000-300000", "dataset:tomekkorbak/detoxify-pile-chunk3-300000-350000", "dataset:tomekkorbak/detoxify-pile-chunk3-350000-400000", "dataset:tomekkorbak/detoxify-pile-chunk3-400000-450000", "dataset:tomekkorbak/detoxify-pile-chunk3-450000-500000", "dataset:tomekkorbak/detoxify-pile-chunk3-500000-550000", "dataset:tomekkorbak/detoxify-pile-chunk3-550000-600000", "dataset:tomekkorbak/detoxify-pile-chunk3-600000-650000", "dataset:tomekkorbak/detoxify-pile-chunk3-650000-700000", "dataset:tomekkorbak/detoxify-pile-chunk3-700000-750000", "dataset:tomekkorbak/detoxify-pile-chunk3-750000-800000", "dataset:tomekkorbak/detoxify-pile-chunk3-800000-850000", "dataset:tomekkorbak/detoxify-pile-chunk3-850000-900000", "dataset:tomekkorbak/detoxify-pile-chunk3-900000-950000", "dataset:tomekkorbak/detoxify-pile-chunk3-950000-1000000", "dataset:tomekkorbak/detoxify-pile-chunk3-1000000-1050000", "dataset:tomekkorbak/detoxify-pile-chunk3-1050000-1100000", "dataset:tomekkorbak/detoxify-pile-chunk3-1100000-1150000", "dataset:tomekkorbak/detoxify-pile-chunk3-1150000-1200000", "dataset:tomekkorbak/detoxify-pile-chunk3-1200000-1250000", "dataset:tomekkorbak/detoxify-pile-chunk3-1250000-1300000", "dataset:tomekkorbak/detoxify-pile-chunk3-1300000-1350000", "dataset:tomekkorbak/detoxify-pile-chunk3-1350000-1400000", "dataset:tomekkorbak/detoxify-pile-chunk3-1400000-1450000", "dataset:tomekkorbak/detoxify-pile-chunk3-1450000-1500000", "dataset:tomekkorbak/detoxify-pile-chunk3-1500000-1550000", "dataset:tomekkorbak/detoxify-pile-chunk3-1550000-1600000", "dataset:tomekkorbak/detoxify-pile-chunk3-1600000-1650000", "dataset:tomekkorbak/detoxify-pile-chunk3-1650000-1700000", "dataset:tomekkorbak/detoxify-pile-chunk3-1700000-1750000", "dataset:tomekkorbak/detoxify-pile-chunk3-1750000-1800000", "dataset:tomekkorbak/detoxify-pile-chunk3-1800000-1850000", "dataset:tomekkorbak/detoxify-pile-chunk3-1850000-1900000", "dataset:tomekkorbak/detoxify-pile-chunk3-1900000-1950000", "license:mit", "region:us" ]
null
2022-11-25T10:19:12Z
--- language: - en license: mit tags: - generated_from_trainer datasets: - tomekkorbak/detoxify-pile-chunk3-0-50000 - tomekkorbak/detoxify-pile-chunk3-50000-100000 - tomekkorbak/detoxify-pile-chunk3-100000-150000 - tomekkorbak/detoxify-pile-chunk3-150000-200000 - tomekkorbak/detoxify-pile-chunk3-200000-250000 - tomekkorbak/detoxify-pile-chunk3-250000-300000 - tomekkorbak/detoxify-pile-chunk3-300000-350000 - tomekkorbak/detoxify-pile-chunk3-350000-400000 - tomekkorbak/detoxify-pile-chunk3-400000-450000 - tomekkorbak/detoxify-pile-chunk3-450000-500000 - tomekkorbak/detoxify-pile-chunk3-500000-550000 - tomekkorbak/detoxify-pile-chunk3-550000-600000 - tomekkorbak/detoxify-pile-chunk3-600000-650000 - tomekkorbak/detoxify-pile-chunk3-650000-700000 - tomekkorbak/detoxify-pile-chunk3-700000-750000 - tomekkorbak/detoxify-pile-chunk3-750000-800000 - tomekkorbak/detoxify-pile-chunk3-800000-850000 - tomekkorbak/detoxify-pile-chunk3-850000-900000 - tomekkorbak/detoxify-pile-chunk3-900000-950000 - tomekkorbak/detoxify-pile-chunk3-950000-1000000 - tomekkorbak/detoxify-pile-chunk3-1000000-1050000 - tomekkorbak/detoxify-pile-chunk3-1050000-1100000 - tomekkorbak/detoxify-pile-chunk3-1100000-1150000 - tomekkorbak/detoxify-pile-chunk3-1150000-1200000 - tomekkorbak/detoxify-pile-chunk3-1200000-1250000 - tomekkorbak/detoxify-pile-chunk3-1250000-1300000 - tomekkorbak/detoxify-pile-chunk3-1300000-1350000 - tomekkorbak/detoxify-pile-chunk3-1350000-1400000 - tomekkorbak/detoxify-pile-chunk3-1400000-1450000 - tomekkorbak/detoxify-pile-chunk3-1450000-1500000 - tomekkorbak/detoxify-pile-chunk3-1500000-1550000 - tomekkorbak/detoxify-pile-chunk3-1550000-1600000 - tomekkorbak/detoxify-pile-chunk3-1600000-1650000 - tomekkorbak/detoxify-pile-chunk3-1650000-1700000 - tomekkorbak/detoxify-pile-chunk3-1700000-1750000 - tomekkorbak/detoxify-pile-chunk3-1750000-1800000 - tomekkorbak/detoxify-pile-chunk3-1800000-1850000 - tomekkorbak/detoxify-pile-chunk3-1850000-1900000 - tomekkorbak/detoxify-pile-chunk3-1900000-1950000 model-index: - name: gallant_thompson results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # gallant_thompson This model was trained from scratch on the tomekkorbak/detoxify-pile-chunk3-0-50000, the tomekkorbak/detoxify-pile-chunk3-50000-100000, the tomekkorbak/detoxify-pile-chunk3-100000-150000, the tomekkorbak/detoxify-pile-chunk3-150000-200000, the tomekkorbak/detoxify-pile-chunk3-200000-250000, the tomekkorbak/detoxify-pile-chunk3-250000-300000, the tomekkorbak/detoxify-pile-chunk3-300000-350000, the tomekkorbak/detoxify-pile-chunk3-350000-400000, the tomekkorbak/detoxify-pile-chunk3-400000-450000, the tomekkorbak/detoxify-pile-chunk3-450000-500000, the tomekkorbak/detoxify-pile-chunk3-500000-550000, the tomekkorbak/detoxify-pile-chunk3-550000-600000, the tomekkorbak/detoxify-pile-chunk3-600000-650000, the tomekkorbak/detoxify-pile-chunk3-650000-700000, the tomekkorbak/detoxify-pile-chunk3-700000-750000, the tomekkorbak/detoxify-pile-chunk3-750000-800000, the tomekkorbak/detoxify-pile-chunk3-800000-850000, the tomekkorbak/detoxify-pile-chunk3-850000-900000, the tomekkorbak/detoxify-pile-chunk3-900000-950000, the tomekkorbak/detoxify-pile-chunk3-950000-1000000, the tomekkorbak/detoxify-pile-chunk3-1000000-1050000, the tomekkorbak/detoxify-pile-chunk3-1050000-1100000, the tomekkorbak/detoxify-pile-chunk3-1100000-1150000, the tomekkorbak/detoxify-pile-chunk3-1150000-1200000, the tomekkorbak/detoxify-pile-chunk3-1200000-1250000, the tomekkorbak/detoxify-pile-chunk3-1250000-1300000, the tomekkorbak/detoxify-pile-chunk3-1300000-1350000, the tomekkorbak/detoxify-pile-chunk3-1350000-1400000, the tomekkorbak/detoxify-pile-chunk3-1400000-1450000, the tomekkorbak/detoxify-pile-chunk3-1450000-1500000, the tomekkorbak/detoxify-pile-chunk3-1500000-1550000, the tomekkorbak/detoxify-pile-chunk3-1550000-1600000, the tomekkorbak/detoxify-pile-chunk3-1600000-1650000, the tomekkorbak/detoxify-pile-chunk3-1650000-1700000, the tomekkorbak/detoxify-pile-chunk3-1700000-1750000, the tomekkorbak/detoxify-pile-chunk3-1750000-1800000, the tomekkorbak/detoxify-pile-chunk3-1800000-1850000, the tomekkorbak/detoxify-pile-chunk3-1850000-1900000 and the tomekkorbak/detoxify-pile-chunk3-1900000-1950000 datasets. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 64 - total_train_batch_size: 1024 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.01 - training_steps: 3147 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0+cu113 - Datasets 2.5.1 - Tokenizers 0.11.6 # Full config {'dataset': {'datasets': ['tomekkorbak/detoxify-pile-chunk3-0-50000', 'tomekkorbak/detoxify-pile-chunk3-50000-100000', 'tomekkorbak/detoxify-pile-chunk3-100000-150000', 'tomekkorbak/detoxify-pile-chunk3-150000-200000', 'tomekkorbak/detoxify-pile-chunk3-200000-250000', 'tomekkorbak/detoxify-pile-chunk3-250000-300000', 'tomekkorbak/detoxify-pile-chunk3-300000-350000', 'tomekkorbak/detoxify-pile-chunk3-350000-400000', 'tomekkorbak/detoxify-pile-chunk3-400000-450000', 'tomekkorbak/detoxify-pile-chunk3-450000-500000', 'tomekkorbak/detoxify-pile-chunk3-500000-550000', 'tomekkorbak/detoxify-pile-chunk3-550000-600000', 'tomekkorbak/detoxify-pile-chunk3-600000-650000', 'tomekkorbak/detoxify-pile-chunk3-650000-700000', 'tomekkorbak/detoxify-pile-chunk3-700000-750000', 'tomekkorbak/detoxify-pile-chunk3-750000-800000', 'tomekkorbak/detoxify-pile-chunk3-800000-850000', 'tomekkorbak/detoxify-pile-chunk3-850000-900000', 'tomekkorbak/detoxify-pile-chunk3-900000-950000', 'tomekkorbak/detoxify-pile-chunk3-950000-1000000', 'tomekkorbak/detoxify-pile-chunk3-1000000-1050000', 'tomekkorbak/detoxify-pile-chunk3-1050000-1100000', 'tomekkorbak/detoxify-pile-chunk3-1100000-1150000', 'tomekkorbak/detoxify-pile-chunk3-1150000-1200000', 'tomekkorbak/detoxify-pile-chunk3-1200000-1250000', 'tomekkorbak/detoxify-pile-chunk3-1250000-1300000', 'tomekkorbak/detoxify-pile-chunk3-1300000-1350000', 'tomekkorbak/detoxify-pile-chunk3-1350000-1400000', 'tomekkorbak/detoxify-pile-chunk3-1400000-1450000', 'tomekkorbak/detoxify-pile-chunk3-1450000-1500000', 'tomekkorbak/detoxify-pile-chunk3-1500000-1550000', 'tomekkorbak/detoxify-pile-chunk3-1550000-1600000', 'tomekkorbak/detoxify-pile-chunk3-1600000-1650000', 'tomekkorbak/detoxify-pile-chunk3-1650000-1700000', 'tomekkorbak/detoxify-pile-chunk3-1700000-1750000', 'tomekkorbak/detoxify-pile-chunk3-1750000-1800000', 'tomekkorbak/detoxify-pile-chunk3-1800000-1850000', 'tomekkorbak/detoxify-pile-chunk3-1850000-1900000', 'tomekkorbak/detoxify-pile-chunk3-1900000-1950000'], 'is_split_by_sentences': True}, 'generation': {'force_call_on': [25354], 'metrics_configs': [{}, {'n': 1}, {'n': 2}, {'n': 5}], 'scenario_configs': [{'generate_kwargs': {'do_sample': True, 'max_length': 128, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'unconditional', 'num_samples': 2048}, {'generate_kwargs': {'do_sample': True, 'max_length': 128, 'min_length': 10, 'temperature': 0.7, 'top_k': 0, 'top_p': 0.9}, 'name': 'challenging_rtp', 'num_samples': 2048, 'prompts_path': 'resources/challenging_rtp.jsonl'}], 'scorer_config': {'device': 'cuda:0'}}, 'kl_gpt3_callback': {'force_call_on': [25354], 'max_tokens': 64, 'num_samples': 4096}, 'model': {'from_scratch': True, 'gpt2_config_kwargs': {'reorder_and_upcast_attn': True, 'scale_attn_by': True}, 'model_kwargs': {'value_head_config': {'is_detached': False}}, 'path_or_name': 'gpt2'}, 'objective': {'alpha': 0.5, 'beta': 10, 'name': 'AWR'}, 'tokenizer': {'path_or_name': 'gpt2'}, 'training': {'dataloader_num_workers': 0, 'effective_batch_size': 1024, 'evaluation_strategy': 'no', 'fp16': True, 'hub_model_id': 'gallant_thompson', 'hub_strategy': 'all_checkpoints', 'learning_rate': 0.001, 'logging_first_step': True, 'logging_steps': 1, 'num_tokens': 3300000000, 'output_dir': 'training_output104340', 'per_device_train_batch_size': 16, 'push_to_hub': True, 'remove_unused_columns': False, 'save_steps': 25354, 'save_strategy': 'steps', 'seed': 42, 'warmup_ratio': 0.01, 'weight_decay': 0.1}} # Wandb URL: https://wandb.ai/tomekkorbak/apo/runs/1fjxzr1j
GinaYang/xlm-roberta-base-finetuned-panx-de
GinaYang
2022-11-25T10:15:33Z
143
0
transformers
[ "transformers", "pytorch", "tensorboard", "xlm-roberta", "token-classification", "generated_from_trainer", "dataset:xtreme", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-11-25T06:11:08Z
--- license: mit tags: - generated_from_trainer datasets: - xtreme metrics: - f1 model-index: - name: xlm-roberta-base-finetuned-panx-de results: - task: name: Token Classification type: token-classification dataset: name: xtreme type: xtreme config: PAN-X.de split: train args: PAN-X.de metrics: - name: F1 type: f1 value: 0.8648740833380706 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-panx-de This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the xtreme dataset. It achieves the following results on the evaluation set: - Loss: 0.1365 - F1: 0.8649 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 24 - eval_batch_size: 24 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.2553 | 1.0 | 525 | 0.1575 | 0.8279 | | 0.1284 | 2.0 | 1050 | 0.1386 | 0.8463 | | 0.0813 | 3.0 | 1575 | 0.1365 | 0.8649 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
ConvLab/t5-small-nlg-tm1_tm2_tm3
ConvLab
2022-11-25T10:08:13Z
105
0
transformers
[ "transformers", "pytorch", "t5", "text2text-generation", "t5-small", "natural language generation", "conversational system", "task-oriented dialog", "en", "dataset:ConvLab/tm1", "dataset:ConvLab/tm2", "dataset:ConvLab/tm3", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-25T03:52:58Z
--- language: - en license: apache-2.0 tags: - t5-small - text2text-generation - natural language generation - conversational system - task-oriented dialog datasets: - ConvLab/tm1 - ConvLab/tm2 - ConvLab/tm3 metrics: - Slot Error Rate - sacrebleu model-index: - name: t5-small-nlg-tm1_tm2_tm3 results: - task: type: text2text-generation name: natural language generation dataset: type: ConvLab/tm1, ConvLab/tm2, ConvLab/tm3 name: TM1+TM2+TM3 split: test metrics: - type: Slot Error Rate value: 2.1 name: SER - type: sacrebleu value: 51.5 name: BLEU widget: - text: "tm1: [inform][pizza_ordering]([name.store][Domino's])\n\nsystem: " - text: "tm2: [inform][restaurant-search]([name.restaurant][Via 313, the Violet Crown Social Club],[price_range][$8 per slice])\n\nsystem: " - text: "tm3: [inform][movie]([name.movie][Star Wars],[name.movie][The Grudge])\n\nsystem: " inference: parameters: max_length: 100 --- # t5-small-nlg-tm1_tm2_tm3 This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on [Taskmaster-1](https://huggingface.co/datasets/ConvLab/tm1), [Taskmaster-2](https://huggingface.co/datasets/ConvLab/tm2), and [Taskmaster-3](https://huggingface.co/datasets/ConvLab/tm3). Refer to [ConvLab-3](https://github.com/ConvLab/ConvLab-3) for model description and usage. ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 512 - optimizer: Adafactor - lr_scheduler_type: linear - num_epochs: 10.0 ### Framework versions - Transformers 4.18.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.3 - Tokenizers 0.11.0
ConvLab/t5-small-nlg-sgd
ConvLab
2022-11-25T09:45:43Z
104
0
transformers
[ "transformers", "pytorch", "t5", "text2text-generation", "t5-small", "natural language generation", "conversational system", "task-oriented dialog", "en", "dataset:ConvLab/sgd", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-25T03:27:52Z
--- language: - en license: apache-2.0 tags: - t5-small - text2text-generation - natural language generation - conversational system - task-oriented dialog datasets: - ConvLab/sgd metrics: - Slot Error Rate - sacrebleu model-index: - name: t5-small-nlg-sgd results: - task: type: text2text-generation name: natural language generation dataset: type: ConvLab/sgd name: SGD split: test revision: 6e8c79b888b21cc658cf9c0ce128d263241cf70f metrics: - type: Slot Error Rate value: 11.9 name: SER - type: sacrebleu value: 29.6 name: BLEU widget: - text: "[request][Restaurants_2]([time][],[restaurant_name][],[location][])\n\nsystem: " - text: "[confirm][Restaurants_2]([number_of_seats][2],[restaurant_name][P.f. Chang's],[location][Corte Madera],[time][12 pm],[date][March 8th])\n\nsystem: " inference: parameters: max_length: 100 --- # t5-small-nlg-sgd This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on [Schema-Guided Dialog](https://huggingface.co/datasets/ConvLab/sgd). Refer to [ConvLab-3](https://github.com/ConvLab/ConvLab-3) for model description and usage. ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 512 - optimizer: Adafactor - lr_scheduler_type: linear - num_epochs: 10.0 ### Framework versions - Transformers 4.18.0 - Pytorch 1.10.2+cu102 - Datasets 1.18.3 - Tokenizers 0.11.0
Shubham09/wispher2
Shubham09
2022-11-25T09:45:00Z
75
0
transformers
[ "transformers", "pytorch", "tensorboard", "whisper", "automatic-speech-recognition", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-11-21T11:43:05Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - wer model-index: - name: wispher2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wispher2 This model is a fine-tuned version of [openai/whisper-base.en](https://huggingface.co/openai/whisper-base.en) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.8125 - Wer: 50.1754 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 8 - eval_batch_size: 5 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - training_steps: 100 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:-------:| | 0.7532 | 1.12 | 100 | 0.8125 | 50.1754 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1 - Datasets 2.7.1 - Tokenizers 0.13.2
aaqibwaleed/roberta-large-squad2-finetuned-squad
aaqibwaleed
2022-11-25T09:09:17Z
61
0
transformers
[ "transformers", "tf", "tensorboard", "roberta", "question-answering", "generated_from_keras_callback", "license:mit", "endpoints_compatible", "region:us" ]
question-answering
2022-11-24T10:46:58Z
--- license: mit tags: - generated_from_keras_callback model-index: - name: aaqibwaleed/roberta-large-squad2-finetuned-squad results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # aaqibwaleed/roberta-large-squad2-finetuned-squad This model is a fine-tuned version of [navteca/roberta-large-squad2](https://huggingface.co/navteca/roberta-large-squad2) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.5015 - Train End Logits Accuracy: 0.8536 - Train Start Logits Accuracy: 0.8279 - Validation Loss: 0.7790 - Validation End Logits Accuracy: 0.7905 - Validation Start Logits Accuracy: 0.7799 - Epoch: 1 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'Adam', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 131822, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False} - training_precision: float32 ### Training results | Train Loss | Train End Logits Accuracy | Train Start Logits Accuracy | Validation Loss | Validation End Logits Accuracy | Validation Start Logits Accuracy | Epoch | |:----------:|:-------------------------:|:---------------------------:|:---------------:|:------------------------------:|:--------------------------------:|:-----:| | 0.9483 | 0.7418 | 0.7176 | 0.8067 | 0.7573 | 0.7504 | 0 | | 0.5015 | 0.8536 | 0.8279 | 0.7790 | 0.7905 | 0.7799 | 1 | ### Framework versions - Transformers 4.24.0 - TensorFlow 2.7.1 - Datasets 2.7.1 - Tokenizers 0.13.2
ameerak/my-awesome-setfit-model-ameer
ameerak
2022-11-25T08:59:31Z
1
0
sentence-transformers
[ "sentence-transformers", "pytorch", "mpnet", "feature-extraction", "sentence-similarity", "transformers", "autotrain_compatible", "endpoints_compatible", "region:us" ]
sentence-similarity
2022-11-25T08:58:56Z
--- pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity - transformers --- # {MODEL_NAME} This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. <!--- Describe your model here --> ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer sentences = ["This is an example sentence", "Each sentence is converted"] model = SentenceTransformer('{MODEL_NAME}') embeddings = model.encode(sentences) print(embeddings) ``` ## Usage (HuggingFace Transformers) Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings. ```python from transformers import AutoTokenizer, AutoModel import torch #Mean Pooling - Take attention mask into account for correct averaging def mean_pooling(model_output, attention_mask): token_embeddings = model_output[0] #First element of model_output contains all token embeddings input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float() return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9) # Sentences we want sentence embeddings for sentences = ['This is an example sentence', 'Each sentence is converted'] # Load model from HuggingFace Hub tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}') model = AutoModel.from_pretrained('{MODEL_NAME}') # Tokenize sentences encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt') # Compute token embeddings with torch.no_grad(): model_output = model(**encoded_input) # Perform pooling. In this case, mean pooling. sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask']) print("Sentence embeddings:") print(sentence_embeddings) ``` ## Evaluation Results <!--- Describe how your model was evaluated --> For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME}) ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 40 with parameters: ``` {'batch_size': 16, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss` Parameters of the fit()-Method: ``` { "epochs": 1, "evaluation_steps": 0, "evaluator": "NoneType", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": 40, "warmup_steps": 4, "weight_decay": 0.01 } ``` ## Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: MPNetModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False}) ) ``` ## Citing & Authors <!--- Describe where people can find more information -->
Nerdward/pegasus-tf-finetuned-model
Nerdward
2022-11-25T08:57:13Z
60
0
transformers
[ "transformers", "tf", "pegasus", "text2text-generation", "generated_from_keras_callback", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-25T08:37:05Z
--- tags: - generated_from_keras_callback model-index: - name: Nerdward/pegasus-tf-finetuned-model results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # Nerdward/pegasus-tf-finetuned-model This model is a fine-tuned version of [human-centered-summarization/financial-summarization-pegasus](https://huggingface.co/human-centered-summarization/financial-summarization-pegasus) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 2.0833 - Validation Loss: 1.3487 - Epoch: 2 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 2e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01} - training_precision: float32 ### Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 2.9216 | 1.8082 | 0 | | 2.3339 | 1.5098 | 1 | | 2.0833 | 1.3487 | 2 | ### Framework versions - Transformers 4.24.0 - TensorFlow 2.9.2 - Datasets 2.7.1 - Tokenizers 0.13.2
clp/xlm-roberta-base-finetuned-marc
clp
2022-11-25T08:53:37Z
104
0
transformers
[ "transformers", "pytorch", "tensorboard", "xlm-roberta", "text-classification", "generated_from_trainer", "dataset:amazon_reviews_multi", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-11-25T08:14:09Z
--- license: mit tags: - generated_from_trainer datasets: - amazon_reviews_multi model-index: - name: xlm-roberta-base-finetuned-marc results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-marc This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the amazon_reviews_multi dataset. It achieves the following results on the evaluation set: - Loss: 0.9725 - Mae: 0.5221 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Mae | |:-------------:|:-----:|:----:|:---------------:|:------:| | 1.1615 | 1.0 | 308 | 1.0893 | 0.6106 | | 0.9994 | 2.0 | 616 | 0.9725 | 0.5221 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
jesspi/IFE-sentence-model2
jesspi
2022-11-25T08:43:43Z
2
0
sentence-transformers
[ "sentence-transformers", "pytorch", "mpnet", "feature-extraction", "sentence-similarity", "transformers", "autotrain_compatible", "endpoints_compatible", "region:us" ]
sentence-similarity
2022-11-25T08:43:30Z
--- pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity - transformers --- # {MODEL_NAME} This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. <!--- Describe your model here --> ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer sentences = ["This is an example sentence", "Each sentence is converted"] model = SentenceTransformer('{MODEL_NAME}') embeddings = model.encode(sentences) print(embeddings) ``` ## Usage (HuggingFace Transformers) Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings. ```python from transformers import AutoTokenizer, AutoModel import torch #Mean Pooling - Take attention mask into account for correct averaging def mean_pooling(model_output, attention_mask): token_embeddings = model_output[0] #First element of model_output contains all token embeddings input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float() return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9) # Sentences we want sentence embeddings for sentences = ['This is an example sentence', 'Each sentence is converted'] # Load model from HuggingFace Hub tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}') model = AutoModel.from_pretrained('{MODEL_NAME}') # Tokenize sentences encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt') # Compute token embeddings with torch.no_grad(): model_output = model(**encoded_input) # Perform pooling. In this case, mean pooling. sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask']) print("Sentence embeddings:") print(sentence_embeddings) ``` ## Evaluation Results <!--- Describe how your model was evaluated --> For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME}) ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 3170 with parameters: ``` {'batch_size': 16, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss` Parameters of the fit()-Method: ``` { "epochs": 1, "evaluation_steps": 0, "evaluator": "NoneType", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 3.710089773651088e-06 }, "scheduler": "WarmupLinear", "steps_per_epoch": 3170, "warmup_steps": 317, "weight_decay": 0.01 } ``` ## Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: MPNetModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False}) ) ``` ## Citing & Authors <!--- Describe where people can find more information -->
mastermile/ff
mastermile
2022-11-25T08:42:00Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2022-11-25T08:42:00Z
--- license: creativeml-openrail-m ---
GhifSmile/textClass-finetuned-coba-conf-orang
GhifSmile
2022-11-25T08:20:08Z
105
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-11-25T07:55:47Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: textClass-finetuned-coba-conf-orang results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # textClass-finetuned-coba-conf-orang This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4868 - Accuracy: 0.7856 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.5108 | 1.0 | 2757 | 0.4652 | 0.7761 | | 0.4484 | 2.0 | 5514 | 0.4447 | 0.7881 | | 0.4031 | 3.0 | 8271 | 0.4552 | 0.7891 | | 0.3646 | 4.0 | 11028 | 0.4720 | 0.7911 | | 0.3372 | 5.0 | 13785 | 0.4868 | 0.7856 | ### Framework versions - Transformers 4.24.0 - Pytorch 1.12.1+cu113 - Datasets 2.7.1 - Tokenizers 0.13.2
BunnyViking/rachelwalkerstylewatercolour
BunnyViking
2022-11-25T07:10:58Z
0
15
null
[ "en", "license:mit", "region:us" ]
null
2022-11-25T06:14:42Z
--- language: - en license: mit --- This checkpoint is trained on the art of Rachel Walker, a New Zealand artist who paints beautiful watercolours. https://www.walkerillo.com/ instance = a watercolor painting with a white background class = rachelwalker style I recommend using the full instance + class as the start of your prompt "a watercolor painting with a white background rachelwalker style [your image subject]" which will give you a good chance of getting the full style ![00003-1392786996-a watercolor painting with a white background a turtle.png](https://s3.amazonaws.com/moonup/production/uploads/1669358723507-631c8158aa346997917dcf5d.png) ![00008-3905760611-a watercolor painting with a white background a rabbit.png](https://s3.amazonaws.com/moonup/production/uploads/1669358779575-631c8158aa346997917dcf5d.png) ![00016-183604033-a watercolor painting with a white background rachelwalker style a robot.png](https://s3.amazonaws.com/moonup/production/uploads/1669358810098-631c8158aa346997917dcf5d.png) ![00046-1124001293-a watercolor painting with a white background rachelwalker style an owl in the evening.png](https://s3.amazonaws.com/moonup/production/uploads/1669359042983-631c8158aa346997917dcf5d.png) ![00019-3289020923-a watercolor painting with a white background rachelwalker style a mountain valley with a dragon flying in the sky.png](https://s3.amazonaws.com/moonup/production/uploads/1669359141450-631c8158aa346997917dcf5d.png)
teacookies/autotrain-25112022-cert2-2236171465
teacookies
2022-11-25T05:41:55Z
113
0
transformers
[ "transformers", "pytorch", "autotrain", "token-classification", "unk", "dataset:teacookies/autotrain-data-25112022-cert2", "co2_eq_emissions", "endpoints_compatible", "region:us" ]
token-classification
2022-11-25T05:32:18Z
--- tags: - autotrain - token-classification language: - unk widget: - text: "I love AutoTrain 🤗" datasets: - teacookies/autotrain-data-25112022-cert2 co2_eq_emissions: emissions: 17.128542622750768 --- # Model Trained Using AutoTrain - Problem type: Entity Extraction - Model ID: 2236171465 - CO2 Emissions (in grams): 17.1285 ## Validation Metrics - Loss: 0.003 - Accuracy: 0.999 - Precision: 0.985 - Recall: 0.986 - F1: 0.986 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoTrain"}' https://api-inference.huggingface.co/models/teacookies/autotrain-25112022-cert2-2236171465 ``` Or Python API: ``` from transformers import AutoModelForTokenClassification, AutoTokenizer model = AutoModelForTokenClassification.from_pretrained("teacookies/autotrain-25112022-cert2-2236171465", use_auth_token=True) tokenizer = AutoTokenizer.from_pretrained("teacookies/autotrain-25112022-cert2-2236171465", use_auth_token=True) inputs = tokenizer("I love AutoTrain", return_tensors="pt") outputs = model(**inputs) ```
BigSalmon/InformalToFormalLincoln91Paraphrase
BigSalmon
2022-11-25T04:28:48Z
170
0
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2022-11-25T03:31:05Z
data: https://github.com/BigSalmon2/InformalToFormalDataset Text Generation Informal Formal ``` from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("BigSalmon/InformalToFormalLincoln90Paraphrase") model = AutoModelForCausalLM.from_pretrained("BigSalmon/InformalToFormalLincoln90Paraphrase") ``` ``` Demo: https://huggingface.co/spaces/BigSalmon/FormalInformalConciseWordy ``` ``` prompt = """informal english: corn fields are all across illinois, visible once you leave chicago.\nTranslated into the Style of Abraham Lincoln:""" input_ids = tokenizer.encode(prompt, return_tensors='pt') outputs = model.generate(input_ids=input_ids, max_length=10 + len(prompt), temperature=1.0, top_k=50, top_p=0.95, do_sample=True, num_return_sequences=5, early_stopping=True) for i in range(5): print(tokenizer.decode(outputs[i])) ``` Most likely outputs (Disclaimer: I highly recommend using this over just generating): ``` prompt = """informal english: corn fields are all across illinois, visible once you leave chicago.\nTranslated into the Style of Abraham Lincoln:""" text = tokenizer.encode(prompt) myinput, past_key_values = torch.tensor([text]), None myinput = myinput myinput= myinput.to(device) logits, past_key_values = model(myinput, past_key_values = past_key_values, return_dict=False) logits = logits[0,-1] probabilities = torch.nn.functional.softmax(logits) best_logits, best_indices = logits.topk(250) best_words = [tokenizer.decode([idx.item()]) for idx in best_indices] text.append(best_indices[0].item()) best_probabilities = probabilities[best_indices].tolist() words = [] print(best_words) ``` ``` How To Make Prompt: informal english: i am very ready to do that just that. Translated into the Style of Abraham Lincoln: you can assure yourself of my readiness to work toward this end. Translated into the Style of Abraham Lincoln: please be assured that i am most ready to undertake this laborious task. *** informal english: space is huge and needs to be explored. Translated into the Style of Abraham Lincoln: space awaits traversal, a new world whose boundaries are endless. Translated into the Style of Abraham Lincoln: space is a ( limitless / boundless ) expanse, a vast virgin domain awaiting exploration. *** informal english: corn fields are all across illinois, visible once you leave chicago. Translated into the Style of Abraham Lincoln: corn fields ( permeate illinois / span the state of illinois / ( occupy / persist in ) all corners of illinois / line the horizon of illinois / envelop the landscape of illinois ), manifesting themselves visibly as one ventures beyond chicago. informal english: ``` ``` original: microsoft word's [MASK] pricing invites competition. Translated into the Style of Abraham Lincoln: microsoft word's unconscionable pricing invites competition. *** original: the library’s quiet atmosphere encourages visitors to [blank] in their work. Translated into the Style of Abraham Lincoln: the library’s quiet atmosphere encourages visitors to immerse themselves in their work. ``` ``` Essay Intro (Warriors vs. Rockets in Game 7): text: eagerly anticipated by fans, game 7's are the highlight of the post-season. text: ever-building in suspense, game 7's have the crowd captivated. *** Essay Intro (South Korean TV Is Becoming Popular): text: maturing into a bona fide paragon of programming, south korean television ( has much to offer / entertains without fail / never disappoints ). text: increasingly held in critical esteem, south korean television continues to impress. text: at the forefront of quality content, south korea is quickly achieving celebrity status. *** Essay Intro ( ``` ``` Search: What is the definition of Checks and Balances? https://en.wikipedia.org/wiki/Checks_and_balances Checks and Balances is the idea of having a system where each and every action in government should be subject to one or more checks that would not allow one branch or the other to overly dominate. https://www.harvard.edu/glossary/Checks_and_Balances Checks and Balances is a system that allows each branch of government to limit the powers of the other branches in order to prevent abuse of power https://www.law.cornell.edu/library/constitution/Checks_and_Balances Checks and Balances is a system of separation through which branches of government can control the other, thus preventing excess power. *** Search: What is the definition of Separation of Powers? https://en.wikipedia.org/wiki/Separation_of_powers The separation of powers is a principle in government, whereby governmental powers are separated into different branches, each with their own set of powers, that are prevent one branch from aggregating too much power. https://www.yale.edu/tcf/Separation_of_Powers.html Separation of Powers is the division of governmental functions between the executive, legislative and judicial branches, clearly demarcating each branch's authority, in the interest of ensuring that individual liberty or security is not undermined. *** Search: What is the definition of Connection of Powers? https://en.wikipedia.org/wiki/Connection_of_powers Connection of Powers is a feature of some parliamentary forms of government where different branches of government are intermingled, typically the executive and legislative branches. https://simple.wikipedia.org/wiki/Connection_of_powers The term Connection of Powers describes a system of government in which there is overlap between different parts of the government. *** Search: What is the definition of ``` ``` Search: What are phrase synonyms for "second-guess"? https://www.powerthesaurus.org/second-guess/synonyms Shortest to Longest: - feel dubious about - raise an eyebrow at - wrinkle their noses at - cast a jaundiced eye at - teeter on the fence about *** Search: What are phrase synonyms for "mean to newbies"? https://www.powerthesaurus.org/mean_to_newbies/synonyms Shortest to Longest: - readiness to balk at rookies - absence of tolerance for novices - hostile attitude toward newcomers *** Search: What are phrase synonyms for "make use of"? https://www.powerthesaurus.org/make_use_of/synonyms Shortest to Longest: - call upon - glean value from - reap benefits from - derive utility from - seize on the merits of - draw on the strength of - tap into the potential of *** Search: What are phrase synonyms for "hurting itself"? https://www.powerthesaurus.org/hurting_itself/synonyms Shortest to Longest: - erring - slighting itself - forfeiting its integrity - doing itself a disservice - evincing a lack of backbone *** Search: What are phrase synonyms for " ``` ``` - nebraska - unicamerical legislature - different from federal house and senate text: featuring a unicameral legislature, nebraska's political system stands in stark contrast to the federal model, comprised of a house and senate. *** - penny has practically no value - should be taken out of circulation - just as other coins have been in us history - lost use - value not enough - to make environmental consequences worthy text: all but valueless, the penny should be retired. as with other coins in american history, it has become defunct. too minute to warrant the environmental consequences of its production, it has outlived its usefulness. *** - ``` ``` original: sports teams are profitable for owners. [MASK], their valuations experience a dramatic uptick. infill: sports teams are profitable for owners. ( accumulating vast sums / stockpiling treasure / realizing benefits / cashing in / registering robust financials / scoring on balance sheets ), their valuations experience a dramatic uptick. *** original: ``` ``` wordy: classical music is becoming less popular more and more. Translate into Concise Text: interest in classic music is fading. *** wordy: ``` ``` sweet: savvy voters ousted him. longer: voters who were informed delivered his defeat. *** sweet: ``` ``` 1: commercial space company spacex plans to launch a whopping 52 flights in 2022. 2: spacex, a commercial space company, intends to undertake a total of 52 flights in 2022. 3: in 2022, commercial space company spacex has its sights set on undertaking 52 flights. 4: 52 flights are in the pipeline for 2022, according to spacex, a commercial space company. 5: a commercial space company, spacex aims to conduct 52 flights in 2022. *** 1: ``` Keywords to sentences or sentence. ``` ngos are characterized by: □ voluntary citizens' group that is organized on a local, national or international level □ encourage political participation □ often serve humanitarian functions □ work for social, economic, or environmental change *** what are the drawbacks of living near an airbnb? □ noise □ parking □ traffic □ security □ strangers *** ``` ``` original: musicals generally use spoken dialogue as well as songs to convey the story. operas are usually fully sung. adapted: musicals generally use spoken dialogue as well as songs to convey the story. ( in a stark departure / on the other hand / in contrast / by comparison / at odds with this practice / far from being alike / in defiance of this standard / running counter to this convention ), operas are usually fully sung. *** original: akoya and tahitian are types of pearls. akoya pearls are mostly white, and tahitian pearls are naturally dark. adapted: akoya and tahitian are types of pearls. ( a far cry from being indistinguishable / easily distinguished / on closer inspection / setting them apart / not to be mistaken for one another / hardly an instance of mere synonymy / differentiating the two ), akoya pearls are mostly white, and tahitian pearls are naturally dark. *** original: ``` ``` original: had trouble deciding. translated into journalism speak: wrestled with the question, agonized over the matter, furrowed their brows in contemplation. *** original: ``` ``` input: not loyal 1800s english: ( two-faced / inimical / perfidious / duplicitous / mendacious / double-dealing / shifty ). *** input: ``` ``` first: ( was complicit in / was involved in ). antonym: ( was blameless / was not an accomplice to / had no hand in / was uninvolved in ). *** first: ( have no qualms about / see no issue with ). antonym: ( are deeply troubled by / harbor grave reservations about / have a visceral aversion to / take ( umbrage at / exception to ) / are wary of ). *** first: ( do not see eye to eye / disagree often ). antonym: ( are in sync / are united / have excellent rapport / are like-minded / are in step / are of one mind / are in lockstep / operate in perfect harmony / march in lockstep ). *** first: ``` ``` stiff with competition, law school {A} is the launching pad for countless careers, {B} is a crowded field, {C} ranks among the most sought-after professional degrees, {D} is a professional proving ground. *** languishing in viewership, saturday night live {A} is due for a creative renaissance, {B} is no longer a ratings juggernaut, {C} has been eclipsed by its imitators, {C} can still find its mojo. *** dubbed the "manhattan of the south," atlanta {A} is a bustling metropolis, {B} is known for its vibrant downtown, {C} is a city of rich history, {D} is the pride of georgia. *** embattled by scandal, harvard {A} is feeling the heat, {B} cannot escape the media glare, {C} is facing its most intense scrutiny yet, {D} is in the spotlight for all the wrong reasons. ``` Infill / Infilling / Masking / Phrase Masking (Works pretty decently actually, especially when you use logprobs code from above): ``` his contention [blank] by the evidence [sep] was refuted [answer] *** few sights are as [blank] new york city as the colorful, flashing signage of its bodegas [sep] synonymous with [answer] *** when rick won the lottery, all of his distant relatives [blank] his winnings [sep] clamored for [answer] *** the library’s quiet atmosphere encourages visitors to [blank] in their work [sep] immerse themselves [answer] *** the joy of sport is that no two games are alike. for every exhilarating experience, however, there is an interminable one. the national pastime, unfortunately, has a penchant for the latter. what begins as a summer evening at the ballpark can quickly devolve into a game of tedium. the primary culprit is the [blank] of play. from batters readjusting their gloves to fielders spitting on their mitts, the action is [blank] unnecessary interruptions. the sport's future is [blank] if these tendencies are not addressed [sep] plodding pace [answer] riddled with [answer] bleak [answer] *** microsoft word's [blank] pricing [blank] competition [sep] unconscionable [answer] invites [answer] *** ``` ``` original: microsoft word's [MASK] pricing invites competition. Translated into the Style of Abraham Lincoln: microsoft word's unconscionable pricing invites competition. *** original: the library’s quiet atmosphere encourages visitors to [blank] in their work. Translated into the Style of Abraham Lincoln: the library’s quiet atmosphere encourages visitors to immerse themselves in their work. ``` Backwards ``` Essay Intro (National Parks): text: tourists are at ease in the national parks, ( swept up in the beauty of their natural splendor ). *** Essay Intro (D.C. Statehood): washington, d.c. is a city of outsize significance, ( ground zero for the nation's political life / center stage for the nation's political machinations ). ``` ``` topic: the Golden State Warriors. characterization 1: the reigning kings of the NBA. characterization 2: possessed of a remarkable cohesion. characterization 3: helmed by superstar Stephen Curry. characterization 4: perched atop the league’s hierarchy. characterization 5: boasting a litany of hall-of-famers. *** topic: emojis. characterization 1: shorthand for a digital generation. characterization 2: more versatile than words. characterization 3: the latest frontier in language. characterization 4: a form of self-expression. characterization 5: quintessentially millennial. characterization 6: reflective of a tech-centric world. *** topic: ``` ``` regular: illinois went against the census' population-loss prediction by getting more residents. VBG: defying the census' prediction of population loss, illinois experienced growth. *** regular: microsoft word’s high pricing increases the likelihood of competition. VBG: extortionately priced, microsoft word is inviting competition. *** regular: ``` ``` source: badminton should be more popular in the US. QUERY: Based on the given topic, can you develop a story outline? target: (1) games played with racquets are popular, (2) just look at tennis and ping pong, (3) but badminton underappreciated, (4) fun, fast-paced, competitive, (5) needs to be marketed more text: the sporting arena is dominated by games that are played with racquets. tennis and ping pong, in particular, are immensely popular. somewhat curiously, however, badminton is absent from this pantheon. exciting, fast-paced, and competitive, it is an underappreciated pastime. all that it lacks is more effective marketing. *** source: movies in theaters should be free. QUERY: Based on the given topic, can you develop a story outline? target: (1) movies provide vital life lessons, (2) many venues charge admission, (3) those without much money text: the lessons that movies impart are far from trivial. the vast catalogue of cinematic classics is replete with inspiring sagas of friendship, bravery, and tenacity. it is regrettable, then, that admission to theaters is not free. in their current form, the doors of this most vital of institutions are closed to those who lack the means to pay. *** source: ``` ``` in the private sector, { transparency } is vital to the business’s credibility. the { disclosure of information } can be the difference between success and failure. *** the labor market is changing, with { remote work } now the norm. this { flexible employment } allows the individual to design their own schedule. *** the { cubicle } is the locus of countless grievances. many complain that the { enclosed workspace } restricts their freedom of movement. *** ``` ``` it would be natural to assume that americans, as a people whose ancestors { immigrated to this country }, would be sympathetic to those seeking to do likewise. question: what does “do likewise” mean in the above context? (a) make the same journey (b) share in the promise of the american dream (c) start anew in the land of opportunity (d) make landfall on the united states *** in the private sector, { transparency } is vital to the business’s credibility. this orientation can be the difference between success and failure. question: what does “this orientation” mean in the above context? (a) visible business practices (b) candor with the public (c) open, honest communication (d) culture of accountability ``` ``` example: suppose you are a teacher. further suppose you want to tell an accurate telling of history. then suppose a parent takes offense. they do so in the name of name of their kid. this happens a lot. text: educators' responsibility to remain true to the historical record often clashes with the parent's desire to shelter their child from uncomfortable realities. *** example: suppose you are a student at college. now suppose you have to buy textbooks. that is going to be worth hundreds of dollars. given how much you already spend on tuition, that is going to hard cost to bear. text: the exorbitant cost of textbooks, which often reaches hundreds of dollars, imposes a sizable financial burden on the already-strapped college student. ``` ``` <Prefix> the atlanta hawks may attribute <Prefix> <Suffix> trae young <Suffix> <Middle> their robust season to <Middle> *** <Prefix> the nobel prize in literature <Prefix> <Suffix> honor <Suffix> <Middle> is a singularly prestigious <Middle> ``` ``` accustomed to having its name uttered ______, harvard university is weathering a rare spell of reputational tumult (a) in reverential tones (b) with great affection (c) in adulatory fashion (d) in glowing terms ``` ``` clarify: international ( {working together} / cooperation ) is called for when ( {issue go beyond lots of borders} / an issue transcends borders / a given matter has transnational implications ). ``` ``` description: when someone thinks that their view is the only right one. synonyms: intolerant, opinionated, narrow-minded, insular, self-righteous. *** description: when you put something off. synonyms: shelve, defer, table, postpone. ``` ``` organic sentence: crowdfunding is about winner of best ideas and it can test an entrepreneur’s idea. rewrite phrases: meritocratic, viability, vision rewritten with phrases: the meritocratic nature of crowdfunding empowers entrepreneurs to test their vision's viability. ``` *Note* Of all the masking techniques, this one works the best. ``` <Prefix> the atlanta hawks may attribute <Prefix> <Suffix> trae young <Suffix> <Middle> their robust season to <Middle> *** <Prefix> the nobel prize in literature <Prefix> <Suffix> honor <Suffix> <Middle> is a singularly prestigious <Middle> ``` ``` essence: when someone's views are keeping within reasonable. refine: the senator's voting record is ( moderate / centrist / pragmatic / balanced / fair-minded / even-handed ). *** essence: when things are worked through in a petty way. refine: the propensity of the u.s. congress to settle every dispute by way of ( mudslinging / bickering / demagoguery / name-calling / finger-pointing / vilification ) is appalling. ``` ``` description: when someone thinks that their view is the only right one. synonyms: intolerant, opinionated, narrow-minded, insular, self-righteous. *** description: when you put something off. synonyms: shelve, defer, table, postpone. ``` ``` organic sentence: crowdfunding is about winner of best ideas and it can test an entrepreneur’s idea. rewrite phrases: meritocratic, viability, vision rewritten with phrases: the meritocratic nature of crowdfunding empowers entrepreneurs to test their vision's viability. ``` ``` music before bedtime [makes for being able to relax] -> is a recipe for relaxation. ``` ``` [people wanting entertainment love traveling new york city] -> travelers flock to new york city in droves, drawn to its iconic entertainment scene. [cannot blame them] -> one cannot fault them [broadway so fun] -> when it is home to such thrilling fare as Broadway. ``` ``` in their ( ‖ when you are rushing because you want to get there on time ‖ / haste to arrive punctually / mad dash to be timely ), morning commuters are too rushed to whip up their own meal. *** politicians prefer to author vague plans rather than ( ‖ when you can make a plan without many unknowns ‖ / actionable policies / concrete solutions ). ``` ``` Q: What is whistleblower protection? A: Whistleblower protection is a form of legal immunity granted to employees who expose the unethical practices of their employer. Q: Why are whistleblower protections important? A: Absent whistleblower protections, employees would be deterred from exposing their employer’s wrongdoing for fear of retribution. Q: Why would an employer engage in retribution? A: An employer who has acted unethically stands to suffer severe financial and reputational damage were their transgressions to become public. To safeguard themselves from these consequences, they might seek to dissuade employees from exposing their wrongdoing. ``` ``` original: the meritocratic nature of crowdfunding [MASK] into their vision's viability. infill: the meritocratic nature of crowdfunding [gives investors idea of how successful] -> ( offers entrepreneurs a window ) into their vision's viability. ``` ``` Leadership | Lecture 17: Worker Morale What Workers Look for in Companies: • Benefits o Tuition reimbursement o Paid parental leave o 401K matching o Profit sharing o Pension plans o Free meals • Social responsibility o Environmental stewardship o Charitable contributions o Diversity • Work-life balance o Telecommuting o Paid holidays and vacation o Casual dress • Growth opportunities • Job security • Competitive compensation • Recognition o Open-door policies o Whistleblower protection o Employee-of-the-month awards o Positive performance reviews o Bonuses ``` ``` description: business keywords: for-profit, fiduciary duty, monopolistic, bottom line, return on investment, short-term thinking, capital-intensive, self-interested, risk-taking, fiduciary duty, merger, speculation, profiteering, oversight, capitalism, diversification ``` ``` 3. In this task, you are given a company name and you need to find its industry. McDonalds -- Restaurant Facebook -- Social Network IKEA -- Furniture American Express -- Credit Services Nokia -- Telecom Nintendo -- Entertainment 4. In this task, you are given a Month and you need to convert it to its corresponding season April -- Spring December -- Winter July -- Summer October -- Fall February -- Winter 5. In this task, you are given a sentence with a missing word and you need to predict the correct word. Managers should set an _____ for their employees. -- example Some people spend more than four _____ in the gym. -- hours The police were on the _____ of arresting the suspect. -- verge They were looking for _____ on how to solve the problem. -- guidance What is the _____ of the coffee? -- price 6. In this task, you are given a paragraph and you need to reorder it to make it logical. It was first proposed in 1987. The total length of the bridge is 1,828 meters. The idea of a bridge connects Hong Kong to Macau. -- The idea of bridge connecting Hong Kong and Macau was first proposed in 1987. The total length of the bridge is 1,828 meters. It is a movie about a brave and noble policeman. The film was produced by Americans. They were Kevin Lima and Chris Buck. They are directors. The movie is called Tarzan. -- Produced by Americans Kevin Lima and Chris Buck, Tarzan is a movie about a brave and noble policeman. It was first discovered in the mountains of India. The active ingredients in this plant can stimulate hair growth. The plant is called "Hair Plus." -- First discovered in the mountains of India, Hair Plus is a plant whose active ingredients can stimulate hair growth. ``` ``` trivia: What is the population of South Korea? response: 51 million. *** trivia: What is the minimum voting age in the US? response: 18. *** trivia: What are the first ten amendments of the US constitution called? response: Bill of Rights. ```
jaimin/parrot_adequacy_model
jaimin
2022-11-25T04:13:48Z
161
0
transformers
[ "transformers", "pytorch", "roberta", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-11-22T05:56:21Z
--- license: apache-2.0 --- Parrot THIS IS AN ANCILLARY MODEL FOR PARROT PARAPHRASER 1. What is Parrot? Parrot is a paraphrase-based utterance augmentation framework purpose-built to accelerate training NLU models. A paraphrase framework is more than just a paraphrasing model.
jaimin/Informal_to_formal
jaimin
2022-11-25T04:11:41Z
104
0
transformers
[ "transformers", "pytorch", "t5", "text2text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-22T04:43:52Z
## This model convert informal sentance to formal sentance.
jaimin/formal_to_informal
jaimin
2022-11-25T04:09:29Z
104
0
transformers
[ "transformers", "pytorch", "t5", "text2text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-11-22T09:55:57Z
## This model convert formal sentance to informal sentance.
hkab/wav2vec-train
hkab
2022-11-25T04:08:41Z
161
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "license:cc-by-nc-4.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-11-24T10:41:22Z
--- license: cc-by-nc-4.0 tags: - generated_from_trainer model-index: - name: wav2vec-train results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec-train This model is a fine-tuned version of [nguyenvulebinh/wav2vec2-base-vietnamese-250h](https://huggingface.co/nguyenvulebinh/wav2vec2-base-vietnamese-250h) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 32 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 20 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.24.0 - Pytorch 1.11.0 - Datasets 2.1.0 - Tokenizers 0.12.1