modelId
stringlengths
5
139
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]date
2020-02-15 11:33:14
2025-09-10 00:38:21
downloads
int64
0
223M
likes
int64
0
11.7k
library_name
stringclasses
551 values
tags
listlengths
1
4.05k
pipeline_tag
stringclasses
55 values
createdAt
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-09-10 00:38:17
card
stringlengths
11
1.01M
paola-md/recipe-roberta-i
paola-md
2022-07-11T16:17:54Z
6
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-11T15:48:05Z
--- license: mit tags: - generated_from_trainer model-index: - name: recipe-roberta-i results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # recipe-roberta-i This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.9919 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 256 - eval_batch_size: 256 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 20 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 1.871 | 1.0 | 149 | 1.4670 | | 1.528 | 2.0 | 298 | 1.3426 | | 1.41 | 3.0 | 447 | 1.2636 | | 1.3332 | 4.0 | 596 | 1.2029 | | 1.2804 | 5.0 | 745 | 1.1646 | | 1.2441 | 6.0 | 894 | 1.1351 | | 1.21 | 7.0 | 1043 | 1.0898 | | 1.182 | 8.0 | 1192 | 1.0725 | | 1.1604 | 9.0 | 1341 | 1.0718 | | 1.1402 | 10.0 | 1490 | 1.0529 | | 1.1308 | 11.0 | 1639 | 1.0512 | | 1.1191 | 12.0 | 1788 | 1.0245 | | 1.0986 | 13.0 | 1937 | 1.0203 | | 1.0919 | 14.0 | 2086 | 1.0158 | | 1.084 | 15.0 | 2235 | 0.9930 | | 1.0797 | 16.0 | 2384 | 0.9855 | | 1.0697 | 17.0 | 2533 | 1.0061 | | 1.0652 | 18.0 | 2682 | 0.9725 | | 1.0658 | 19.0 | 2831 | 0.9861 | | 1.0642 | 20.0 | 2980 | 0.9919 | ### Framework versions - Transformers 4.19.0.dev0 - Pytorch 1.11.0+cu102 - Datasets 2.3.2 - Tokenizers 0.12.1
nateraw/keras-dummy-model-mixin-demo
nateraw
2022-07-11T15:42:05Z
0
0
keras
[ "keras", "tf-keras", "region:us" ]
null
2022-03-02T23:29:05Z
--- library_name: keras --- ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Model Plot <details> <summary>View Model Plot</summary> ![Model Image](./model.png) </details>
nateraw/keras-dummy-functional-demo
nateraw
2022-07-11T15:41:53Z
0
0
keras
[ "keras", "tf-keras", "region:us" ]
null
2022-03-02T23:29:05Z
--- library_name: keras --- ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: | Hyperparameters | Value | | :-- | :-- | | name | Adam | | learning_rate | 0.001 | | decay | 0.0 | | beta_1 | 0.9 | | beta_2 | 0.999 | | epsilon | 1e-07 | | amsgrad | False | | training_precision | float32 | ## Model Plot <details> <summary>View Model Plot</summary> ![Model Image](./model.png) </details>
jonatasgrosman/exp_w2v2t_es_unispeech-sat_s833
jonatasgrosman
2022-07-11T15:35:27Z
5
0
transformers
[ "transformers", "pytorch", "unispeech-sat", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T15:34:36Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_unispeech-sat_s833 Fine-tuned [microsoft/unispeech-sat-large](https://huggingface.co/microsoft/unispeech-sat-large) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
ariesutiono/finetuned-test-1
ariesutiono
2022-07-11T14:57:10Z
10
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "fill-mask", "generated_from_trainer", "dataset:conll2003", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-11T13:24:51Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - conll2003 model-index: - name: finetuned-test-1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # finetuned-test-1 This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the conll2003 dataset. It achieves the following results on the evaluation set: - Loss: 1.8192 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 20 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.8219 | 1.0 | 30 | 2.3343 | | 2.4148 | 2.0 | 60 | 2.2010 | | 2.3236 | 3.0 | 90 | 2.1442 | | 2.2231 | 4.0 | 120 | 2.1651 | | 2.2171 | 5.0 | 150 | 2.0614 | | 2.127 | 6.0 | 180 | 2.0405 | | 2.0748 | 7.0 | 210 | 2.0092 | | 2.0511 | 8.0 | 240 | 1.9798 | | 2.0097 | 9.0 | 270 | 1.8662 | | 1.9969 | 10.0 | 300 | 1.9257 | | 2.0006 | 11.0 | 330 | 1.9386 | | 1.9273 | 12.0 | 360 | 1.9357 | | 1.9177 | 13.0 | 390 | 1.8983 | | 1.9128 | 14.0 | 420 | 1.8990 | | 1.8979 | 15.0 | 450 | 1.9037 | | 1.8721 | 16.0 | 480 | 1.8440 | | 1.8998 | 17.0 | 510 | 1.8404 | | 1.8862 | 18.0 | 540 | 1.9193 | | 1.9133 | 19.0 | 570 | 1.8494 | | 1.8799 | 20.0 | 600 | 1.8192 | ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0+cu113 - Datasets 2.3.2 - Tokenizers 0.12.1
jonatasgrosman/exp_w2v2t_es_vp-nl_s203
jonatasgrosman
2022-07-11T14:42:15Z
4
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T14:41:38Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_vp-nl_s203 Fine-tuned [facebook/wav2vec2-large-nl-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-nl-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jakka/t5-small-finetuned-xsum
jakka
2022-07-11T14:00:49Z
4
0
transformers
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "dataset:xsum", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2022-07-04T12:57:56Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - xsum metrics: - rouge model-index: - name: t5-small-finetuned-xsum results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: xsum type: xsum args: default metrics: - name: Rouge1 type: rouge value: 22.215 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-finetuned-xsum This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the xsum dataset. It achieves the following results on the evaluation set: - Loss: 2.7323 - Rouge1: 22.215 - Rouge2: 4.296 - Rougel: 17.2091 - Rougelsum: 17.212 - Gen Len: 18.655 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:-------:|:---------:|:-------:| | 3.1005 | 1.0 | 625 | 2.7323 | 22.215 | 4.296 | 17.2091 | 17.212 | 18.655 | ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0+cu113 - Datasets 2.3.2 - Tokenizers 0.12.1
huggingartists/taylor-swift
huggingartists
2022-07-11T13:52:52Z
23
3
transformers
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingartists", "lyrics", "lm-head", "causal-lm", "en", "dataset:huggingartists/taylor-swift", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2022-03-02T23:29:05Z
--- language: en datasets: - huggingartists/taylor-swift tags: - huggingartists - lyrics - lm-head - causal-lm widget: - text: "I am" --- <div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:DISPLAY_1; margin-left: auto; margin-right: auto; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://images.genius.com/721a6c465a666419bf286b473287c33f.446x446x1.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 HuggingArtists Model 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Taylor Swift</div> <a href="https://genius.com/artists/taylor-swift"> <div style="text-align: center; font-size: 14px;">@taylor-swift</div> </a> </div> I was made with [huggingartists](https://github.com/AlekseyKorshuk/huggingartists). Create your own bot based on your favorite artist with [the demo](https://colab.research.google.com/github/AlekseyKorshuk/huggingartists/blob/master/huggingartists-demo.ipynb)! ## How does it work? To understand how the model was developed, check the [W&B report](https://wandb.ai/huggingartists/huggingartists/reportlist). ## Training data The model was trained on lyrics from Taylor Swift. Dataset is available [here](https://huggingface.co/datasets/huggingartists/taylor-swift). And can be used with: ```python from datasets import load_dataset dataset = load_dataset("huggingartists/taylor-swift") ``` [Explore the data](https://wandb.ai/huggingartists/huggingartists/runs/2l84tzp2/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on Taylor Swift's lyrics. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/huggingartists/huggingartists/runs/1hy7aa65) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/huggingartists/huggingartists/runs/1hy7aa65/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingartists/taylor-swift') generator("I am", num_return_sequences=5) ``` Or with Transformers library: ```python from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("huggingartists/taylor-swift") model = AutoModelWithLMHead.from_pretrained("huggingartists/taylor-swift") ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Aleksey Korshuk* [![Follow](https://img.shields.io/github/followers/AlekseyKorshuk?style=social)](https://github.com/AlekseyKorshuk) [![Follow](https://img.shields.io/twitter/follow/alekseykorshuk?style=social)](https://twitter.com/intent/follow?screen_name=alekseykorshuk) [![Follow](https://img.shields.io/badge/dynamic/json?color=blue&label=Telegram%20Channel&query=%24.result&url=https%3A%2F%2Fapi.telegram.org%2Fbot1929545866%3AAAFGhV-KKnegEcLiyYJxsc4zV6C-bdPEBtQ%2FgetChatMemberCount%3Fchat_id%3D-1001253621662&style=social&logo=telegram)](https://t.me/joinchat/_CQ04KjcJ-4yZTky) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/AlekseyKorshuk/huggingartists?style=social)](https://github.com/AlekseyKorshuk/huggingartists)
jonatasgrosman/exp_w2v2t_es_vp-es_s515
jonatasgrosman
2022-07-11T13:49:39Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T13:48:54Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_vp-es_s515 Fine-tuned [facebook/wav2vec2-large-es-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-es-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
egg22314/LaserTube
egg22314
2022-07-11T13:03:19Z
0
1
null
[ "region:us" ]
null
2022-07-11T13:01:55Z
Watching YouTube videos too boring for you? Wish you could be punished for not clicking on stuff fast enough while you watch a cat play the piano? Well, LaserTube is here to solve that problem, by letting you turn any YouTube video into a genuine simulation of an oldschool laserdisc arcade game! Work in progress.
paola-md/recipe-roberta-upper-Is
paola-md
2022-07-11T12:57:29Z
61
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-11T08:50:33Z
--- license: mit tags: - generated_from_trainer model-index: - name: recipe-roberta-upper-Is results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # recipe-roberta-upper-Is This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7757 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 256 - eval_batch_size: 256 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 20 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 1.2455 | 1.0 | 1228 | 1.0420 | | 1.0812 | 2.0 | 2456 | 0.9641 | | 1.018 | 3.0 | 3684 | 0.9220 | | 0.977 | 4.0 | 4912 | 0.8943 | | 0.9451 | 5.0 | 6140 | 0.8726 | | 0.9254 | 6.0 | 7368 | 0.8574 | | 0.9074 | 7.0 | 8596 | 0.8404 | | 0.8944 | 8.0 | 9824 | 0.8290 | | 0.8797 | 9.0 | 11052 | 0.8258 | | 0.869 | 10.0 | 12280 | 0.8115 | | 0.8609 | 11.0 | 13508 | 0.8085 | | 0.8522 | 12.0 | 14736 | 0.7995 | | 0.8462 | 13.0 | 15964 | 0.7958 | | 0.8414 | 14.0 | 17192 | 0.7891 | | 0.8374 | 15.0 | 18420 | 0.7856 | | 0.8327 | 16.0 | 19648 | 0.7850 | | 0.8268 | 17.0 | 20876 | 0.7784 | | 0.8256 | 18.0 | 22104 | 0.7802 | | 0.822 | 19.0 | 23332 | 0.7789 | | 0.8219 | 20.0 | 24560 | 0.7757 | ### Framework versions - Transformers 4.19.0.dev0 - Pytorch 1.11.0+cu102 - Datasets 2.3.2 - Tokenizers 0.12.1
ernestumorga/sac-seals-Humanoid-v0
ernestumorga
2022-07-11T12:40:34Z
6
0
stable-baselines3
[ "stable-baselines3", "seals/Humanoid-v0", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-07-11T12:38:38Z
--- library_name: stable-baselines3 tags: - seals/Humanoid-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: SAC results: - metrics: - type: mean_reward value: -200.52 +/- 55.30 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: seals/Humanoid-v0 type: seals/Humanoid-v0 --- # **SAC** Agent playing **seals/Humanoid-v0** This is a trained model of a **SAC** agent playing **seals/Humanoid-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib ``` # Download model and save it into the logs/ folder python -m utils.load_from_hub --algo sac --env seals/Humanoid-v0 -orga ernestumorga -f logs/ python enjoy.py --algo sac --env seals/Humanoid-v0 -f logs/ ``` ## Training (with the RL Zoo) ``` python train.py --algo sac --env seals/Humanoid-v0 -f logs/ # Upload the model and generate video (when possible) python -m utils.push_to_hub --algo sac --env seals/Humanoid-v0 -f logs/ -orga ernestumorga ``` ## Hyperparameters ```python OrderedDict([('batch_size', 64), ('buffer_size', 100000), ('gamma', 0.98), ('learning_rate', 4.426351861707874e-05), ('learning_starts', 20000), ('n_timesteps', 2000000.0), ('policy', 'MlpPolicy'), ('policy_kwargs', 'dict(net_arch=[400, 300], log_std_init=-0.1034412732183072)'), ('tau', 0.08), ('train_freq', 8), ('normalize', False)]) ```
ernestumorga/sac-seals-Hopper-v0
ernestumorga
2022-07-11T12:38:11Z
2
0
stable-baselines3
[ "stable-baselines3", "seals/Hopper-v0", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-07-11T12:36:55Z
--- library_name: stable-baselines3 tags: - seals/Hopper-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: SAC results: - metrics: - type: mean_reward value: 2330.52 +/- 138.95 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: seals/Hopper-v0 type: seals/Hopper-v0 --- # **SAC** Agent playing **seals/Hopper-v0** This is a trained model of a **SAC** agent playing **seals/Hopper-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib ``` # Download model and save it into the logs/ folder python -m utils.load_from_hub --algo sac --env seals/Hopper-v0 -orga ernestumorga -f logs/ python enjoy.py --algo sac --env seals/Hopper-v0 -f logs/ ``` ## Training (with the RL Zoo) ``` python train.py --algo sac --env seals/Hopper-v0 -f logs/ # Upload the model and generate video (when possible) python -m utils.push_to_hub --algo sac --env seals/Hopper-v0 -f logs/ -orga ernestumorga ``` ## Hyperparameters ```python OrderedDict([('batch_size', 128), ('buffer_size', 100000), ('gamma', 0.98), ('learning_rate', 0.001709807687567946), ('learning_starts', 1000), ('n_timesteps', 1000000.0), ('policy', 'MlpPolicy'), ('policy_kwargs', 'dict(net_arch=[256, 256], log_std_init=-1.6829391077276037)'), ('tau', 0.08), ('train_freq', 32), ('normalize', False)]) ```
ernestumorga/ppo-seals-Humanoid-v0
ernestumorga
2022-07-11T12:36:37Z
5
0
stable-baselines3
[ "stable-baselines3", "seals/Humanoid-v0", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-05-30T12:00:35Z
--- library_name: stable-baselines3 tags: - seals/Humanoid-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - metrics: - type: mean_reward value: -43.69 +/- 155.83 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: seals/Humanoid-v0 type: seals/Humanoid-v0 --- # **PPO** Agent playing **seals/Humanoid-v0** This is a trained model of a **PPO** agent playing **seals/Humanoid-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib ``` # Download model and save it into the logs/ folder python -m utils.load_from_hub --algo ppo --env seals/Humanoid-v0 -orga ernestumorga -f logs/ python enjoy.py --algo ppo --env seals/Humanoid-v0 -f logs/ ``` ## Training (with the RL Zoo) ``` python train.py --algo ppo --env seals/Humanoid-v0 -f logs/ # Upload the model and generate video (when possible) python -m utils.push_to_hub --algo ppo --env seals/Humanoid-v0 -f logs/ -orga ernestumorga ``` ## Hyperparameters ```python OrderedDict([('batch_size', 256), ('clip_range', 0.2), ('ent_coef', 2.0745206045994986e-05), ('gae_lambda', 0.92), ('gamma', 0.999), ('learning_rate', 2.0309225666232827e-05), ('max_grad_norm', 0.5), ('n_envs', 1), ('n_epochs', 20), ('n_steps', 2048), ('n_timesteps', 10000000.0), ('normalize', True), ('policy', 'MlpPolicy'), ('policy_kwargs', 'dict(activation_fn=nn.ReLU, net_arch=[dict(pi=[256, 256], ' 'vf=[256, 256])])'), ('vf_coef', 0.819262464558427), ('normalize_kwargs', {'norm_obs': True, 'norm_reward': False})]) ```
ernestumorga/sac-seals-Walker2d-v0
ernestumorga
2022-07-11T12:34:16Z
2
0
stable-baselines3
[ "stable-baselines3", "seals/Walker2d-v0", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-07-11T12:33:06Z
--- library_name: stable-baselines3 tags: - seals/Walker2d-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: SAC results: - metrics: - type: mean_reward value: 2271.04 +/- 496.40 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: seals/Walker2d-v0 type: seals/Walker2d-v0 --- # **SAC** Agent playing **seals/Walker2d-v0** This is a trained model of a **SAC** agent playing **seals/Walker2d-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib ``` # Download model and save it into the logs/ folder python -m utils.load_from_hub --algo sac --env seals/Walker2d-v0 -orga ernestumorga -f logs/ python enjoy.py --algo sac --env seals/Walker2d-v0 -f logs/ ``` ## Training (with the RL Zoo) ``` python train.py --algo sac --env seals/Walker2d-v0 -f logs/ # Upload the model and generate video (when possible) python -m utils.push_to_hub --algo sac --env seals/Walker2d-v0 -f logs/ -orga ernestumorga ``` ## Hyperparameters ```python OrderedDict([('batch_size', 128), ('buffer_size', 100000), ('gamma', 0.99), ('learning_rate', 0.0005845844772048097), ('learning_starts', 1000), ('n_timesteps', 1000000.0), ('policy', 'MlpPolicy'), ('policy_kwargs', 'dict(net_arch=[400, 300], log_std_init=0.1955317469998743)'), ('tau', 0.02), ('train_freq', 1), ('normalize', False)]) ```
ernestumorga/sac-seals-Swimmer-v0
ernestumorga
2022-07-11T12:31:16Z
1
0
stable-baselines3
[ "stable-baselines3", "seals/Swimmer-v0", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-07-11T12:30:14Z
--- library_name: stable-baselines3 tags: - seals/Swimmer-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: SAC results: - metrics: - type: mean_reward value: 27.34 +/- 1.27 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: seals/Swimmer-v0 type: seals/Swimmer-v0 --- # **SAC** Agent playing **seals/Swimmer-v0** This is a trained model of a **SAC** agent playing **seals/Swimmer-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib ``` # Download model and save it into the logs/ folder python -m utils.load_from_hub --algo sac --env seals/Swimmer-v0 -orga ernestumorga -f logs/ python enjoy.py --algo sac --env seals/Swimmer-v0 -f logs/ ``` ## Training (with the RL Zoo) ``` python train.py --algo sac --env seals/Swimmer-v0 -f logs/ # Upload the model and generate video (when possible) python -m utils.push_to_hub --algo sac --env seals/Swimmer-v0 -f logs/ -orga ernestumorga ``` ## Hyperparameters ```python OrderedDict([('batch_size', 128), ('buffer_size', 100000), ('gamma', 0.995), ('learning_rate', 0.00039981805535514633), ('learning_starts', 1000), ('n_timesteps', 1000000.0), ('policy', 'MlpPolicy'), ('policy_kwargs', 'dict(net_arch=[400, 300], log_std_init=-2.689958330139309)'), ('tau', 0.01), ('train_freq', 256), ('normalize', False)]) ```
ernestumorga/ppo-seals-Walker2d-v0
ernestumorga
2022-07-11T12:25:31Z
0
0
stable-baselines3
[ "stable-baselines3", "seals/Walker2d-v0", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-05-30T10:53:25Z
--- library_name: stable-baselines3 tags: - seals/Walker2d-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - metrics: - type: mean_reward value: 1429.13 +/- 411.75 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: seals/Walker2d-v0 type: seals/Walker2d-v0 --- # **PPO** Agent playing **seals/Walker2d-v0** This is a trained model of a **PPO** agent playing **seals/Walker2d-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib ``` # Download model and save it into the logs/ folder python -m utils.load_from_hub --algo ppo --env seals/Walker2d-v0 -orga ernestumorga -f logs/ python enjoy.py --algo ppo --env seals/Walker2d-v0 -f logs/ ``` ## Training (with the RL Zoo) ``` python train.py --algo ppo --env seals/Walker2d-v0 -f logs/ # Upload the model and generate video (when possible) python -m utils.push_to_hub --algo ppo --env seals/Walker2d-v0 -f logs/ -orga ernestumorga ``` ## Hyperparameters ```python OrderedDict([('batch_size', 8), ('clip_range', 0.4), ('ent_coef', 0.00013057334805552262), ('gae_lambda', 0.92), ('gamma', 0.98), ('learning_rate', 3.791707778339674e-05), ('max_grad_norm', 0.6), ('n_envs', 1), ('n_epochs', 5), ('n_steps', 2048), ('n_timesteps', 1000000.0), ('normalize', True), ('policy', 'MlpPolicy'), ('policy_kwargs', 'dict(activation_fn=nn.ReLU, net_arch=[dict(pi=[256, 256], ' 'vf=[256, 256])])'), ('vf_coef', 0.6167177795726859), ('normalize_kwargs', {'norm_obs': True, 'norm_reward': False})]) ```
jonatasgrosman/exp_w2v2t_es_wavlm_s655
jonatasgrosman
2022-07-11T11:44:23Z
3
0
transformers
[ "transformers", "pytorch", "wavlm", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T11:43:35Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_wavlm_s655 Fine-tuned [microsoft/wavlm-large](https://huggingface.co/microsoft/wavlm-large) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_es_wavlm_s26
jonatasgrosman
2022-07-11T11:37:51Z
3
0
transformers
[ "transformers", "pytorch", "wavlm", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T11:37:01Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_wavlm_s26 Fine-tuned [microsoft/wavlm-large](https://huggingface.co/microsoft/wavlm-large) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
rajkumarrrk/gpt-2-fine-tuned-on-cnn-dm
rajkumarrrk
2022-07-11T11:36:42Z
9
0
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2022-07-11T10:51:53Z
--- license: apache-2.0 --- GPT-2 fine-tuned on CNN/DM summarization dataset. Training args:\ { "learning_rate": 0.0001\ "logging_steps": 5000\ "lr_scheduler_type": "cosine"\ "num_train_epochs": 2\ "per_device_train_batch_size": 12, # Total batch size: 36\ "weight_decay": 0.1\ } {"generation_kwargs": {"do_sample": true, "max_new_tokens": 100, "min_length": 50} Pre-processing to truncate the article to contain only 500 tokens. Post-processing to consider only first three sentences as the summary. Test split metrics: Meteor: 0.2562237219960531\ Rouge1: 0.3754558158439447\ Rouge2: 0.15532626375157227\ RougeL: 0.25813023509572597\ RougeLsum: 0.3489472885043494\ BLEU: 0.09285941365815623\ Bert_score: 0.87570951795246\
jonatasgrosman/exp_w2v2t_es_no-pretraining_s953
jonatasgrosman
2022-07-11T11:23:40Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T11:22:50Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_no-pretraining_s953 Fine-tuned randomly initialized wav2vec2 model for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_es_vp-sv_s93
jonatasgrosman
2022-07-11T11:11:20Z
5
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T11:10:33Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_vp-sv_s93 Fine-tuned [facebook/wav2vec2-large-sv-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-sv-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_es_hubert_s251
jonatasgrosman
2022-07-11T10:59:51Z
3
0
transformers
[ "transformers", "pytorch", "hubert", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T10:59:03Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_hubert_s251 Fine-tuned [facebook/hubert-large-ll60k](https://huggingface.co/facebook/hubert-large-ll60k) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_es_hubert_s456
jonatasgrosman
2022-07-11T10:56:25Z
5
0
transformers
[ "transformers", "pytorch", "hubert", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T10:55:43Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_hubert_s456 Fine-tuned [facebook/hubert-large-ll60k](https://huggingface.co/facebook/hubert-large-ll60k) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_es_hubert_s459
jonatasgrosman
2022-07-11T10:52:59Z
3
0
transformers
[ "transformers", "pytorch", "hubert", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T10:52:20Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_hubert_s459 Fine-tuned [facebook/hubert-large-ll60k](https://huggingface.co/facebook/hubert-large-ll60k) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_es_unispeech_s461
jonatasgrosman
2022-07-11T10:49:40Z
3
0
transformers
[ "transformers", "pytorch", "unispeech", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T10:49:09Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_unispeech_s461 Fine-tuned [microsoft/unispeech-large-1500h-cv](https://huggingface.co/microsoft/unispeech-large-1500h-cv) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
GhostZen/distilbert-base-uncased-finetuned-squad
GhostZen
2022-07-11T10:38:10Z
4
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "question-answering", "generated_from_trainer", "dataset:squad", "license:apache-2.0", "endpoints_compatible", "region:us" ]
question-answering
2022-07-11T10:09:04Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - squad model-index: - name: distilbert-base-uncased-finetuned-squad results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-squad This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the squad dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0+cu113 - Datasets 2.3.2 - Tokenizers 0.12.1
jonatasgrosman/exp_w2v2t_es_xlsr-53_s756
jonatasgrosman
2022-07-11T10:35:54Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T10:35:16Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_xlsr-53_s756 Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_es_xlsr-53_s377
jonatasgrosman
2022-07-11T10:32:41Z
4
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T10:32:11Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_xlsr-53_s377 Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
wooihen/distilbert-base-uncased-finetuned-emotion
wooihen
2022-07-11T10:28:32Z
7
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:emotion", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-07-11T10:04:15Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - emotion metrics: - accuracy - f1 model-index: - name: distilbert-base-uncased-finetuned-emotion results: - task: name: Text Classification type: text-classification dataset: name: emotion type: emotion args: default metrics: - name: Accuracy type: accuracy value: 0.9225 - name: F1 type: f1 value: 0.922771245052197 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-emotion This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset. It achieves the following results on the evaluation set: - Loss: 0.2146 - Accuracy: 0.9225 - F1: 0.9228 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.8233 | 1.0 | 250 | 0.3068 | 0.9025 | 0.8995 | | 0.2394 | 2.0 | 500 | 0.2146 | 0.9225 | 0.9228 | ### Framework versions - Transformers 4.13.0 - Pytorch 1.11.0+cu113 - Datasets 1.16.1 - Tokenizers 0.10.3
jonatasgrosman/exp_w2v2t_es_vp-100k_s957
jonatasgrosman
2022-07-11T10:23:05Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T10:22:18Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_vp-100k_s957 Fine-tuned [facebook/wav2vec2-large-100k-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-100k-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_es_wav2vec2_s596
jonatasgrosman
2022-07-11T10:16:07Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "es", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T10:15:16Z
--- language: - es license: apache-2.0 tags: - automatic-speech-recognition - es datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_es_wav2vec2_s596 Fine-tuned [facebook/wav2vec2-large-lv60](https://huggingface.co/facebook/wav2vec2-large-lv60) for speech recognition using the train split of [Common Voice 7.0 (es)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_vp-it_s533
jonatasgrosman
2022-07-11T10:09:38Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T10:09:12Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_vp-it_s533 Fine-tuned [facebook/wav2vec2-large-it-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-it-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_vp-it_s817
jonatasgrosman
2022-07-11T09:59:54Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T09:59:26Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_vp-it_s817 Fine-tuned [facebook/wav2vec2-large-it-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-it-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_r-wav2vec2_s408
jonatasgrosman
2022-07-11T09:53:37Z
5
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T09:53:13Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_r-wav2vec2_s408 Fine-tuned [facebook/wav2vec2-large-robust](https://huggingface.co/facebook/wav2vec2-large-robust) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_xls-r_s946
jonatasgrosman
2022-07-11T09:47:04Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T09:46:39Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_xls-r_s946 Fine-tuned [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_unispeech-sat_s160
jonatasgrosman
2022-07-11T09:29:53Z
3
0
transformers
[ "transformers", "pytorch", "unispeech-sat", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T09:29:28Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_unispeech-sat_s160 Fine-tuned [microsoft/unispeech-sat-large](https://huggingface.co/microsoft/unispeech-sat-large) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
mrm8488/biomedtra-small-es-finetuned-bioasq-es
mrm8488
2022-07-11T09:27:29Z
7
0
transformers
[ "transformers", "pytorch", "tensorboard", "electra", "question-answering", "generated_from_trainer", "endpoints_compatible", "region:us" ]
question-answering
2022-07-11T08:09:31Z
--- tags: - generated_from_trainer model-index: - name: biomedtra-small-es-finetuned-bioasq-es results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # biomedtra-small-es-finetuned-bioasq-es This model is a fine-tuned version of [mrm8488/biomedtra-small-es](https://huggingface.co/mrm8488/biomedtra-small-es) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.9869 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.214 | 1.0 | 2802 | 2.1565 | | 1.9885 | 2.0 | 5604 | 1.9794 | | 1.9288 | 3.0 | 8406 | 1.9869 | ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0+cu113 - Datasets 2.3.2 - Tokenizers 0.12.1
jonatasgrosman/exp_w2v2t_ru_unispeech-sat_s418
jonatasgrosman
2022-07-11T09:26:37Z
3
0
transformers
[ "transformers", "pytorch", "unispeech-sat", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T09:26:11Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_unispeech-sat_s418 Fine-tuned [microsoft/unispeech-sat-large](https://huggingface.co/microsoft/unispeech-sat-large) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_vp-nl_s624
jonatasgrosman
2022-07-11T09:20:19Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T09:19:47Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_vp-nl_s624 Fine-tuned [facebook/wav2vec2-large-nl-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-nl-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_vp-nl_s131
jonatasgrosman
2022-07-11T09:17:04Z
4
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T09:16:37Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_vp-nl_s131 Fine-tuned [facebook/wav2vec2-large-nl-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-nl-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_vp-es_s729
jonatasgrosman
2022-07-11T09:04:34Z
5
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T09:04:06Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_vp-es_s729 Fine-tuned [facebook/wav2vec2-large-es-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-es-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_vp-fr_s730
jonatasgrosman
2022-07-11T08:58:28Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T08:58:02Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_vp-fr_s730 Fine-tuned [facebook/wav2vec2-large-fr-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-fr-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_wavlm_s331
jonatasgrosman
2022-07-11T08:42:19Z
5
0
transformers
[ "transformers", "pytorch", "wavlm", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T08:41:54Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_wavlm_s331 Fine-tuned [microsoft/wavlm-large](https://huggingface.co/microsoft/wavlm-large) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_wavlm_s363
jonatasgrosman
2022-07-11T08:36:26Z
3
1
transformers
[ "transformers", "pytorch", "wavlm", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T08:36:00Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_wavlm_s363 Fine-tuned [microsoft/wavlm-large](https://huggingface.co/microsoft/wavlm-large) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_no-pretraining_s895
jonatasgrosman
2022-07-11T08:30:17Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T08:29:32Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_no-pretraining_s895 Fine-tuned randomly initialized wav2vec2 model for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
ybelkada/japanese-dummy-tokenizer
ybelkada
2022-07-11T08:24:32Z
4
1
transformers
[ "transformers", "ja", "japanese", "tokenizer", "en", "dataset:snow_simplified_japanese_corpus", "license:mit", "endpoints_compatible", "region:us" ]
null
2022-04-06T12:31:37Z
--- language: - en - ja license: mit datasets: - snow_simplified_japanese_corpus tags: - ja - japanese - tokenizer widget: - text: "誰が一番に着くか私には分かりません。" --- # Japanese Dummy Tokenizer Repository containing a dummy Japanese Tokenizer trained on ```snow_simplified_japanese_corpus``` dataset. The tokenizer has been trained using Hugging Face datasets in a streaming manner. ## Intended uses & limitations You can use this tokenizer to tokenize Japanese sentences. ## How to use it ``` from transformers import AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("ybelkada/japanese-dummy-tokenizer") ``` ## How to train the tokenizer Check the file ```tokenizer.py```, you can freely adapt it to other datasets. This tokenizer is based on the tokenizer from ```csebuetnlp/mT5_multilingual_XLSum```.
jonatasgrosman/exp_w2v2t_ru_vp-sv_s658
jonatasgrosman
2022-07-11T08:21:28Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T08:20:56Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_vp-sv_s658 Fine-tuned [facebook/wav2vec2-large-sv-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-sv-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_vp-sv_s515
jonatasgrosman
2022-07-11T08:14:49Z
4
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T08:14:00Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_vp-sv_s515 Fine-tuned [facebook/wav2vec2-large-sv-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-sv-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_hubert_s732
jonatasgrosman
2022-07-11T08:10:54Z
3
0
transformers
[ "transformers", "pytorch", "hubert", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T08:10:28Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_hubert_s732 Fine-tuned [facebook/hubert-large-ll60k](https://huggingface.co/facebook/hubert-large-ll60k) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
LeoFeng/superb_wav2vec_submit
LeoFeng
2022-07-11T08:05:02Z
0
0
null
[ "region:us" ]
null
2022-07-11T07:57:49Z
# SUPERB Submission Template Welcome to the [SUPERB Challenge](https://superbbenchmark.org/challenge-slt2022/challenge_overview)! SUPERB is a collection of benchmarking resources to evaluate the capability of a universal shared representation for speech processing. It comes with a benchmark on the publicly available datasets and a challenge on a secret/not released hidden dataset. In SUPERB Challenge, a challenging hidden dataset is newly recorded to evaluate the ultimate generaliziblity across various tasks and data. You can participate the challenge by simply submitting your self-supervised (SSL) pretrained models (model definition & pretrained weights), and we benchmark it with the hidden datasets. This repository constains useful tools to let you easliy [submit](https://superbbenchmark.org/submit) your models ***privately*** for evaluation to [the challenge hidden-set leaderboard](https://superbbenchmark.org/leaderboard?track=constrained&subset=Hidden+Dev+Set). 1. Generate a submission template 2. Validate the format/interface correctness of your model 3. Upload to Huggingface's Hub (privately) 4. Submit the upload information to [SUPERB website](https://superbbenchmark.org/submit) #### Note 1. We accept pre-trained models in PyTorch by default. If you wish to submit upstreams in non-PyTorch frameworks, please mail to [superb.announcement@gmail.com](mailto:superb.announcement@gmail.com)! #### Note 2. If you are not feasible to submit the pre-trained model, please mail to [superb.announcement@gmail.com](mailto:superb.announcement@gmail.com) for us to see how to help! ## Quickstart ### 1. Add model interfaces #### forward Extract features from waveforms. - **Input:** A list of waveforms in 16000 Hz ```python SAMPLE_RATE = 16000 BATCH_SIZE = 8 EXAMPLE_SEC = 10 wavs = [torch.randn(SAMPLE_RATE * EXAMPLE_SEC).cuda() for _ in range(BATCH_SIZE)] ``` - **Output:** A dictionary with a key "hidden_states" (for compatiblility with old ver.). The value is **a list** of padded sequences in the same shape of **(batch_size, max_sequence_length_of_batch, hidden_size)** for weighted-sum to work. It is welcome to perform some task-specified / independent pre- / post-processing on the upstream's raw hidden-sets, including upsampling and downsampling. However, all the values must come from **a single upstream model**: ```python tasks = ["hidden_states", "PR", "SID", "ER", "ASR", "ASV", "SD", "QbE", "ST", "SS", "SE", "secret"] for task in tasks: # you can do task-specified pre- / post-processing depend on the arg "upstream_feature_selection" results = upstream(wavs, upstream_feature_selection=task) hidden_states = results["hidden_states"] assert isinstance(results, dict) assert isinstance(hidden_states, list) for state in hidden_states: assert isinstance(state, torch.Tensor) assert state.dim() == 3, "(batch_size, max_sequence_length_of_batch, hidden_size)" assert state.shape == hidden_states[0].shape ``` #### get_downsample_rates Provide the downsample rate **from 16000 Hz waveforms** for each task's representation in the dict. For the standard 10ms stride representation, the downsample rate is 160. ```python SAMPLE_RATE = 16000 MSEC_PER_SEC = 1000 downsample_rate = SAMPLE_RATE * 10 / MSEC_PER_SEC # 160 ``` The downsample rate will be used to: 1. Calculate the valid representation length of each utterance in the output padded representation. 2. Prepare the training materials according to the representation's downsample rate for frame-level tasks, e.g. SD, SE, and SS. - **Input:** the task key (str) - **Output:** the downsample rate (int) of the representation for that task ```python for task in tasks: assert isinstance(task, str) downsample_rate = upstream.get_downsample_rate(task) assert isinstance(downsample_rate, int) print("The upstream's representation for {task}" f" has the downsample rate of {downsample_rate}.") ``` ### 2. Create an account and organization on the Hugging Face Hub First create an account on the Hugging Face Hub and you can sign up [here](https://huggingface.co/join) if you haven't already! Next, create a new organization and invite the SUPERB Hidden Set Committee to join. You will upload your model to a repository under this organization so that members inside it can access the model which is not publicly available. * [superb-hidden-set](https://huggingface.co/superb-hidden-set) ### 3. Create a template repository on your machine The next step is to create a template repository on your local machine that contains various files and a CLI to help you validate and submit your pretrained models. The Hugging Face Hub uses [Git Large File Storage (LFS)](https://git-lfs.github.com) to manage large files, so first install it if you don't have it already. For example, on macOS you can run: ```bash brew install git-lfs git lfs install ``` Next, run the following commands to create the repository. We recommend creating a Python virtual environment for the project, e.g. with Anaconda: ```bash # Create and activate a virtual environment conda create -n superb-submit python=3.8 && conda activate superb-submit # Install the following libraries pip install cookiecutter huggingface-hub==0.0.16 # Create the template repository cookiecutter git+https://huggingface.co/superb/superb-submission ``` This will ask you to specify your Hugging Face Hub username, password, organisation, and the name of the repository: ``` hf_hub_username [<huggingface>]: hf_hub_password [<password>]: hf_hub_organisation [superb-submissions]: repo_name [<my-superb-submissions>]: ``` This will trigger the following steps: 1. Create a private dataset repository on the Hugging Face Hub under `{hf_hub_organisation}/{repo_name}` 2. Clone the repository to your local machine 3. Add various template files, commit them locally to the repository, and push them to the Hub The resulting repository should have the following structure: ``` my-superb-submission ├── LICENSE ├── README.md <- The README with submission instructions ├── cli.py <- The CLI for validating predictions etc └── requirements.txt <- The requirements packages for the submissions ├── expert.py <- Your model definition └── model.pt <- Your model weights ``` ### 4. Install the dependencies The final step is to install the project's dependencies: ```bash # Navigate to the template repository cd my-superb-submission # Install dependencies python -m pip install -r requirements.txt ``` That's it! You're now all set to start pretraining your speech models - see the instructions below on how to submit them to the Hub. ## Submitting to the leaderboard To make a submission to the [leaderboard](https://superbbenchmark.org/leaderboard?subset=Hidden+Dev+Set), there are 4 main steps: 1. Modify `expert.py` and change `model.pt` so we can initialize an upstream model following the [challenge policy](https://superbbenchmark.org/challenge-slt2022/upstream) by: ```python upstream = UpstreamExpert(ckpt="./model.pt") ``` ***Package Dependency:*** Note that we only install `torch` package so far by following the above steps. If your model needs more packages, you can modify the `requirement.txt` to meet your need and install them inside the current conda environment. We will install the packages you list in the `requirement.txt` before initializing the upstream model. 2. Validate the upstream model's interface meets the requirements in the [challenge policy](https://superbbenchmark.org/challenge-slt2022/upstream). If everything is correct, you should see the following message: "All submission files validated! Now you can make a submission." ``` python cli.py validate ``` 3. Push the model to the Hub! If there are no errors, you should see the following message: "Upload successful!" ``` python cli.py upload "commit message: my best model" ``` 4. [Make a submission at SUPERB website](https://superbbenchmark.org/submit) by uniquely indentifying this uploaded model with the following information, which can be shown by: ``` python cli.py info ``` - Organization Name - Repository Name - Commit Hash (full 40 characters) After you finish the above 4 steps. You will see a new entry in your [SUPERB profile page](https://superbbenchmark.org/profile) (need login) which does not have any benchmark numbers yet. Please wait for us to finetuned it on the hidden dataset and get the benchmark results. The results will be revealed within one week. Please stay tuned!
jonatasgrosman/exp_w2v2t_ru_unispeech_s132
jonatasgrosman
2022-07-11T07:58:18Z
5
0
transformers
[ "transformers", "pytorch", "unispeech", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T07:57:53Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_unispeech_s132 Fine-tuned [microsoft/unispeech-large-1500h-cv](https://huggingface.co/microsoft/unispeech-large-1500h-cv) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_xlsr-53_s201
jonatasgrosman
2022-07-11T07:49:05Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T07:48:21Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_xlsr-53_s201 Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_ru_vp-100k_s334
jonatasgrosman
2022-07-11T07:42:16Z
4
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "ru", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T07:41:33Z
--- language: - ru license: apache-2.0 tags: - automatic-speech-recognition - ru datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_ru_vp-100k_s334 Fine-tuned [facebook/wav2vec2-large-100k-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-100k-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (ru)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
igpaub/Reinforce-Pixelcopter-PLE-v0
igpaub
2022-07-11T07:33:51Z
0
0
null
[ "Pixelcopter-PLE-v0", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2022-07-11T07:33:44Z
--- tags: - Pixelcopter-PLE-v0 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-Pixelcopter-PLE-v0 results: - metrics: - type: mean_reward value: 17.00 +/- 11.95 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: Pixelcopter-PLE-v0 type: Pixelcopter-PLE-v0 --- # **Reinforce** Agent playing **Pixelcopter-PLE-v0** This is a trained model of a **Reinforce** agent playing **Pixelcopter-PLE-v0** . To learn to use this model and train yours check Unit 5 of the Deep Reinforcement Learning Class: https://github.com/huggingface/deep-rl-class/tree/main/unit5
infinitejoy/MLAgents-PushBlock
infinitejoy
2022-07-11T07:26:44Z
11
0
ml-agents
[ "ml-agents", "tensorboard", "onnx", "unity-ml-agents", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-PushBlock", "region:us" ]
reinforcement-learning
2022-07-11T07:26:25Z
--- tags: - unity-ml-agents - ml-agents - deep-reinforcement-learning - reinforcement-learning - ML-Agents-PushBlock library_name: ml-agents --- # **ppo** Agent playing **PushBlock** This is a trained model of a **ppo** agent playing **PushBlock** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://github.com/huggingface/ml-agents#get-started We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: ### Resume the training ``` mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser:**. 1. Go to https://huggingface.co/spaces/unity/ML-Agents-PushBlock 2. Step 1: Write your model_id: infinitejoy/MLAgents-PushBlock 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
jonatasgrosman/exp_w2v2t_nl_vp-it_s449
jonatasgrosman
2022-07-11T07:20:08Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T07:19:25Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_vp-it_s449 Fine-tuned [facebook/wav2vec2-large-it-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-it-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_r-wav2vec2_s925
jonatasgrosman
2022-07-11T07:09:56Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T07:09:27Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_r-wav2vec2_s925 Fine-tuned [facebook/wav2vec2-large-robust](https://huggingface.co/facebook/wav2vec2-large-robust) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_xls-r_s831
jonatasgrosman
2022-07-11T06:54:35Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T06:53:49Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_xls-r_s831 Fine-tuned [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_xls-r_s133
jonatasgrosman
2022-07-11T06:51:13Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T06:50:31Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_xls-r_s133 Fine-tuned [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_unispeech-sat_s715
jonatasgrosman
2022-07-11T06:47:42Z
4
0
transformers
[ "transformers", "pytorch", "unispeech-sat", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T06:47:15Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_unispeech-sat_s715 Fine-tuned [microsoft/unispeech-sat-large](https://huggingface.co/microsoft/unispeech-sat-large) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
hellennamulinda/eng-lug
hellennamulinda
2022-07-11T06:45:00Z
10
0
transformers
[ "transformers", "pytorch", "marian", "text2text-generation", "autotrain", "unk", "co2_eq_emissions", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2022-07-01T13:10:28Z
--- tags: autotrain language: unk widget: - text: "I love AutoTrain 🤗" co2_eq_emissions: 0.04087910671538076 --- # Model Trained Using AutoTrain - Problem type: Translation - Model ID: 1026034854 - CO2 Emissions (in grams): 0.04087910671538076 ## Validation Metrics - Loss: 1.0871405601501465 - Rouge1: 55.8225 - Rouge2: 34.1547 - RougeL: 54.4274 - RougeLsum: 54.408 - Gen Len: 23.178 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_HUGGINGFACE_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoTrain"}' https://api-inference.huggingface.co/hellennamulinda/autotrain-eng-lug-1070637495 ```
jonatasgrosman/exp_w2v2t_nl_unispeech-sat_s81
jonatasgrosman
2022-07-11T06:33:53Z
4
0
transformers
[ "transformers", "pytorch", "unispeech-sat", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T06:33:27Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_unispeech-sat_s81 Fine-tuned [microsoft/unispeech-sat-large](https://huggingface.co/microsoft/unispeech-sat-large) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_vp-nl_s158
jonatasgrosman
2022-07-11T06:26:19Z
4
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T06:25:51Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_vp-nl_s158 Fine-tuned [facebook/wav2vec2-large-nl-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-nl-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_vp-es_s476
jonatasgrosman
2022-07-11T06:20:15Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T06:19:49Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_vp-es_s476 Fine-tuned [facebook/wav2vec2-large-es-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-es-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_vp-es_s576
jonatasgrosman
2022-07-11T06:17:18Z
4
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T06:16:52Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_vp-es_s576 Fine-tuned [facebook/wav2vec2-large-es-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-es-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_vp-fr_s417
jonatasgrosman
2022-07-11T06:08:16Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T06:07:50Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_vp-fr_s417 Fine-tuned [facebook/wav2vec2-large-fr-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-fr-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_vp-fr_s156
jonatasgrosman
2022-07-11T06:05:18Z
5
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T06:04:52Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_vp-fr_s156 Fine-tuned [facebook/wav2vec2-large-fr-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-fr-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_unispeech-ml_s498
jonatasgrosman
2022-07-11T05:58:25Z
3
0
transformers
[ "transformers", "pytorch", "unispeech", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T05:57:58Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_unispeech-ml_s498 Fine-tuned [microsoft/unispeech-large-multi-lingual-1500h-cv](https://huggingface.co/microsoft/unispeech-large-multi-lingual-1500h-cv) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_unispeech-ml_s23
jonatasgrosman
2022-07-11T05:55:28Z
3
0
transformers
[ "transformers", "pytorch", "unispeech", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T05:55:01Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_unispeech-ml_s23 Fine-tuned [microsoft/unispeech-large-multi-lingual-1500h-cv](https://huggingface.co/microsoft/unispeech-large-multi-lingual-1500h-cv) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_wavlm_s246
jonatasgrosman
2022-07-11T05:52:32Z
3
0
transformers
[ "transformers", "pytorch", "wavlm", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T05:51:59Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_wavlm_s246 Fine-tuned [microsoft/wavlm-large](https://huggingface.co/microsoft/wavlm-large) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
AndyChiang/bert-test
AndyChiang
2022-07-11T05:50:10Z
3
0
transformers
[ "transformers", "pytorch", "tf", "bert", "fill-mask", "generated_from_keras_callback", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-11T03:34:14Z
--- tags: - generated_from_keras_callback model-index: - name: bert-test results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # bert-test This model was trained from scratch on an unknown dataset. It achieves the following results on the evaluation set: ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: None - training_precision: float32 ### Training results ### Framework versions - Transformers 4.20.1 - TensorFlow 2.8.2 - Datasets 2.3.2 - Tokenizers 0.12.1
jonatasgrosman/exp_w2v2t_nl_no-pretraining_s512
jonatasgrosman
2022-07-11T05:43:20Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T05:42:54Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_no-pretraining_s512 Fine-tuned randomly initialized wav2vec2 model for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_hubert_s562
jonatasgrosman
2022-07-11T05:00:40Z
3
0
transformers
[ "transformers", "pytorch", "hubert", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T05:00:15Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_hubert_s562 Fine-tuned [facebook/hubert-large-ll60k](https://huggingface.co/facebook/hubert-large-ll60k) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_hubert_s585
jonatasgrosman
2022-07-11T04:50:08Z
3
0
transformers
[ "transformers", "pytorch", "hubert", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T04:49:42Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_hubert_s585 Fine-tuned [facebook/hubert-large-ll60k](https://huggingface.co/facebook/hubert-large-ll60k) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_hubert_s319
jonatasgrosman
2022-07-11T04:36:03Z
3
0
transformers
[ "transformers", "pytorch", "hubert", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T04:35:38Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_hubert_s319 Fine-tuned [facebook/hubert-large-ll60k](https://huggingface.co/facebook/hubert-large-ll60k) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_unispeech_s683
jonatasgrosman
2022-07-11T04:24:15Z
3
0
transformers
[ "transformers", "pytorch", "unispeech", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T04:23:30Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_unispeech_s683 Fine-tuned [microsoft/unispeech-large-1500h-cv](https://huggingface.co/microsoft/unispeech-large-1500h-cv) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_unispeech_s493
jonatasgrosman
2022-07-11T04:15:55Z
4
0
transformers
[ "transformers", "pytorch", "unispeech", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T04:15:16Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_unispeech_s493 Fine-tuned [microsoft/unispeech-large-1500h-cv](https://huggingface.co/microsoft/unispeech-large-1500h-cv) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_unispeech_s853
jonatasgrosman
2022-07-11T04:05:42Z
3
0
transformers
[ "transformers", "pytorch", "unispeech", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T04:05:02Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_unispeech_s853 Fine-tuned [microsoft/unispeech-large-1500h-cv](https://huggingface.co/microsoft/unispeech-large-1500h-cv) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_xlsr-53_s948
jonatasgrosman
2022-07-11T03:52:19Z
6
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T03:51:53Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_xlsr-53_s948 Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
infinitejoy/MLAgents-Worm
infinitejoy
2022-07-11T03:50:39Z
14
0
ml-agents
[ "ml-agents", "tensorboard", "onnx", "unity-ml-agents", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Worm", "region:us" ]
reinforcement-learning
2022-07-11T03:50:32Z
--- tags: - unity-ml-agents - ml-agents - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Worm library_name: ml-agents --- # **ppo** Agent playing **Worm** This is a trained model of a **ppo** agent playing **Worm** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://github.com/huggingface/ml-agents#get-started We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: ### Resume the training ``` mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser:**. 1. Go to https://huggingface.co/spaces/unity/ML-Agents-Worm 2. Step 1: Write your model_id: infinitejoy/MLAgents-Worm 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
emegona/finetuning-pysentimiento-war-tweets
emegona
2022-07-11T03:33:41Z
4
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-06-30T13:03:38Z
--- tags: - generated_from_trainer metrics: - accuracy - f1 model-index: - name: finetuning-pysentimiento-war-tweets results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # finetuning-pysentimiento-war-tweets This model is a fine-tuned version of [finiteautomata/beto-sentiment-analysis](https://huggingface.co/finiteautomata/beto-sentiment-analysis) on a dataset of 1500 tweets from Peruvian accounts. It achieves the following results on the evaluation set: - Loss: 1.7689 - Accuracy: 0.7378 - F1: 0.7456 ## Model description This model in a fine-tuned version of [finiteautomata/beto-sentiment-analysis](https://huggingface.co/finiteautomata/beto-sentiment-analysis) using five labels: **pro_russia**, **against_ukraine**, **neutral**, **against_russia**, **pro_ukraine**. ## Intended uses & limitations This model shall be used to classify text (more specifically, Spanish tweets) as expressing a position concerning the Russo-Ukrainian war. ## Training and evaluation data We used an 80/20 training/test split on the aforementioned dataset. ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0+cu113 - Datasets 2.3.2 - Tokenizers 0.12.1
jonatasgrosman/exp_w2v2t_nl_xlsr-53_s972
jonatasgrosman
2022-07-11T03:31:44Z
4
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T03:31:17Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_xlsr-53_s972 Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_wav2vec2_s721
jonatasgrosman
2022-07-11T03:04:17Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T03:03:52Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_wav2vec2_s721 Fine-tuned [facebook/wav2vec2-large-lv60](https://huggingface.co/facebook/wav2vec2-large-lv60) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_nl_wav2vec2_s754
jonatasgrosman
2022-07-11T02:56:30Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T02:56:05Z
--- language: - nl license: apache-2.0 tags: - automatic-speech-recognition - nl datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_nl_wav2vec2_s754 Fine-tuned [facebook/wav2vec2-large-lv60](https://huggingface.co/facebook/wav2vec2-large-lv60) for speech recognition using the train split of [Common Voice 7.0 (nl)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_vp-it_s992
jonatasgrosman
2022-07-11T02:32:42Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T02:32:18Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_vp-it_s992 Fine-tuned [facebook/wav2vec2-large-it-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-it-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_r-wav2vec2_s732
jonatasgrosman
2022-07-11T02:11:14Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T02:10:50Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_r-wav2vec2_s732 Fine-tuned [facebook/wav2vec2-large-robust](https://huggingface.co/facebook/wav2vec2-large-robust) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_r-wav2vec2_s957
jonatasgrosman
2022-07-11T02:01:13Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T02:00:33Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_r-wav2vec2_s957 Fine-tuned [facebook/wav2vec2-large-robust](https://huggingface.co/facebook/wav2vec2-large-robust) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_xls-r_s448
jonatasgrosman
2022-07-11T01:23:08Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T01:22:44Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_xls-r_s448 Fine-tuned [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_unispeech-sat_s211
jonatasgrosman
2022-07-11T01:18:21Z
3
0
transformers
[ "transformers", "pytorch", "unispeech-sat", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T01:17:39Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_unispeech-sat_s211 Fine-tuned [microsoft/unispeech-sat-large](https://huggingface.co/microsoft/unispeech-sat-large) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_vp-es_s803
jonatasgrosman
2022-07-11T00:20:03Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T00:19:20Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_vp-es_s803 Fine-tuned [facebook/wav2vec2-large-es-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-es-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_vp-es_s95
jonatasgrosman
2022-07-11T00:13:50Z
5
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-11T00:13:09Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_vp-es_s95 Fine-tuned [facebook/wav2vec2-large-es-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-es-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_vp-fr_s581
jonatasgrosman
2022-07-10T23:57:47Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-10T23:57:24Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_vp-fr_s581 Fine-tuned [facebook/wav2vec2-large-fr-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-fr-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_unispeech-ml_s545
jonatasgrosman
2022-07-10T23:44:50Z
4
0
transformers
[ "transformers", "pytorch", "unispeech", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-10T23:44:05Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_unispeech-ml_s545 Fine-tuned [microsoft/unispeech-large-multi-lingual-1500h-cv](https://huggingface.co/microsoft/unispeech-large-multi-lingual-1500h-cv) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_unispeech-ml_s527
jonatasgrosman
2022-07-10T23:36:00Z
4
0
transformers
[ "transformers", "pytorch", "unispeech", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-10T23:35:36Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_unispeech-ml_s527 Fine-tuned [microsoft/unispeech-large-multi-lingual-1500h-cv](https://huggingface.co/microsoft/unispeech-large-multi-lingual-1500h-cv) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_wavlm_s887
jonatasgrosman
2022-07-10T23:17:52Z
3
0
transformers
[ "transformers", "pytorch", "wavlm", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-10T23:17:06Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_wavlm_s887 Fine-tuned [microsoft/wavlm-large](https://huggingface.co/microsoft/wavlm-large) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_no-pretraining_s663
jonatasgrosman
2022-07-10T23:04:46Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-10T23:04:22Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_no-pretraining_s663 Fine-tuned randomly initialized wav2vec2 model for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_no-pretraining_s211
jonatasgrosman
2022-07-10T22:56:22Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-10T22:55:57Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_no-pretraining_s211 Fine-tuned randomly initialized wav2vec2 model for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_vp-sv_s953
jonatasgrosman
2022-07-10T22:47:21Z
4
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-10T22:46:56Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_vp-sv_s953 Fine-tuned [facebook/wav2vec2-large-sv-voxpopuli](https://huggingface.co/facebook/wav2vec2-large-sv-voxpopuli) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.
jonatasgrosman/exp_w2v2t_et_hubert_s507
jonatasgrosman
2022-07-10T22:41:25Z
3
0
transformers
[ "transformers", "pytorch", "hubert", "automatic-speech-recognition", "et", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-10T22:40:39Z
--- language: - et license: apache-2.0 tags: - automatic-speech-recognition - et datasets: - mozilla-foundation/common_voice_7_0 --- # exp_w2v2t_et_hubert_s507 Fine-tuned [facebook/hubert-large-ll60k](https://huggingface.co/facebook/hubert-large-ll60k) for speech recognition using the train split of [Common Voice 7.0 (et)](https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0). When using this model, make sure that your speech input is sampled at 16kHz. This model has been fine-tuned by the [HuggingSound](https://github.com/jonatasgrosman/huggingsound) tool.