modelId
stringlengths
5
139
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]date
2020-02-15 11:33:14
2025-09-01 18:27:28
downloads
int64
0
223M
likes
int64
0
11.7k
library_name
stringclasses
532 values
tags
listlengths
1
4.05k
pipeline_tag
stringclasses
55 values
createdAt
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-09-01 18:27:19
card
stringlengths
11
1.01M
yanaiela/roberta-base-epoch_61
yanaiela
2022-07-29T23:01:44Z
3
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_61", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:39:32Z
--- language: en tags: - roberta-base - roberta-base-epoch_61 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 61 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_61. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_60
yanaiela
2022-07-29T23:01:22Z
3
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_60", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:36:36Z
--- language: en tags: - roberta-base - roberta-base-epoch_60 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 60 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_60. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_57
yanaiela
2022-07-29T23:00:18Z
3
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_57", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:34:22Z
--- language: en tags: - roberta-base - roberta-base-epoch_57 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 57 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_57. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_55
yanaiela
2022-07-29T22:59:33Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_55", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:32:24Z
--- language: en tags: - roberta-base - roberta-base-epoch_55 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 55 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_55. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_53
yanaiela
2022-07-29T22:58:46Z
6
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_53", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:30:47Z
--- language: en tags: - roberta-base - roberta-base-epoch_53 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 53 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_53. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_51
yanaiela
2022-07-29T22:57:57Z
3
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_51", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:29:17Z
--- language: en tags: - roberta-base - roberta-base-epoch_51 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 51 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_51. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_50
yanaiela
2022-07-29T22:57:31Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_50", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:28:26Z
--- language: en tags: - roberta-base - roberta-base-epoch_50 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 50 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_50. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_46
yanaiela
2022-07-29T22:55:54Z
3
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_46", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:25:28Z
--- language: en tags: - roberta-base - roberta-base-epoch_46 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 46 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_46. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_45
yanaiela
2022-07-29T22:55:32Z
5
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_45", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:24:44Z
--- language: en tags: - roberta-base - roberta-base-epoch_45 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 45 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_45. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_43
yanaiela
2022-07-29T22:54:43Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_43", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:23:18Z
--- language: en tags: - roberta-base - roberta-base-epoch_43 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 43 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_43. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_40
yanaiela
2022-07-29T22:53:26Z
5
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_40", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:21:04Z
--- language: en tags: - roberta-base - roberta-base-epoch_40 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 40 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_40. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_39
yanaiela
2022-07-29T22:53:02Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_39", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:20:23Z
--- language: en tags: - roberta-base - roberta-base-epoch_39 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 39 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_39. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_36
yanaiela
2022-07-29T22:52:02Z
5
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_36", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:17:50Z
--- language: en tags: - roberta-base - roberta-base-epoch_36 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 36 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_36. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_35
yanaiela
2022-07-29T22:51:43Z
5
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_35", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:17:09Z
--- language: en tags: - roberta-base - roberta-base-epoch_35 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 35 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_35. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_34
yanaiela
2022-07-29T22:51:23Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_34", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:16:23Z
--- language: en tags: - roberta-base - roberta-base-epoch_34 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 34 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_34. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_33
yanaiela
2022-07-29T22:51:06Z
5
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_33", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:15:37Z
--- language: en tags: - roberta-base - roberta-base-epoch_33 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 33 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_33. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_30
yanaiela
2022-07-29T22:50:11Z
3
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_30", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:13:21Z
--- language: en tags: - roberta-base - roberta-base-epoch_30 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 30 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_30. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_29
yanaiela
2022-07-29T22:49:52Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_29", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:12:07Z
--- language: en tags: - roberta-base - roberta-base-epoch_29 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 29 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_29. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_28
yanaiela
2022-07-29T22:49:33Z
6
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_28", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:11:20Z
--- language: en tags: - roberta-base - roberta-base-epoch_28 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 28 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_28. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_26
yanaiela
2022-07-29T22:48:56Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_26", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:09:55Z
--- language: en tags: - roberta-base - roberta-base-epoch_26 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 26 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_26. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_25
yanaiela
2022-07-29T22:48:37Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_25", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:09:03Z
--- language: en tags: - roberta-base - roberta-base-epoch_25 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 25 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_25. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_23
yanaiela
2022-07-29T22:48:00Z
5
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_23", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:07:39Z
--- language: en tags: - roberta-base - roberta-base-epoch_23 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 23 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_23. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_22
yanaiela
2022-07-29T22:47:41Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_22", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:06:52Z
--- language: en tags: - roberta-base - roberta-base-epoch_22 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 22 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_22. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_19
yanaiela
2022-07-29T22:46:46Z
5
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_19", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:04:23Z
--- language: en tags: - roberta-base - roberta-base-epoch_19 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 19 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_19. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_18
yanaiela
2022-07-29T22:46:26Z
5
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_18", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:03:29Z
--- language: en tags: - roberta-base - roberta-base-epoch_18 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 18 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_18. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_15
yanaiela
2022-07-29T22:45:30Z
5
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_15", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T17:01:23Z
--- language: en tags: - roberta-base - roberta-base-epoch_15 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 15 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_15. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_12
yanaiela
2022-07-29T22:44:35Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_12", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T16:59:07Z
--- language: en tags: - roberta-base - roberta-base-epoch_12 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 12 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_12. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_11
yanaiela
2022-07-29T22:44:17Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_11", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T16:57:41Z
--- language: en tags: - roberta-base - roberta-base-epoch_11 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 11 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_11. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_9
yanaiela
2022-07-29T22:43:40Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_9", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T16:56:14Z
--- language: en tags: - roberta-base - roberta-base-epoch_9 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 9 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_9. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_7
yanaiela
2022-07-29T22:43:03Z
7
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_7", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T16:54:40Z
--- language: en tags: - roberta-base - roberta-base-epoch_7 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 7 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_7. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
yanaiela/roberta-base-epoch_2
yanaiela
2022-07-29T22:41:25Z
4
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "roberta-base", "roberta-base-epoch_2", "en", "dataset:wikipedia", "dataset:bookcorpus", "arxiv:1907.11692", "arxiv:2207.14251", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T16:50:41Z
--- language: en tags: - roberta-base - roberta-base-epoch_2 license: mit datasets: - wikipedia - bookcorpus --- # RoBERTa, Intermediate Checkpoint - Epoch 2 This model is part of our reimplementation of the [RoBERTa model](https://arxiv.org/abs/1907.11692), trained on Wikipedia and the Book Corpus only. We train this model for almost 100K steps, corresponding to 83 epochs. We provide the 84 checkpoints (including the randomly initialized weights before the training) to provide the ability to study the training dynamics of such models, and other possible use-cases. These models were trained in part of a work that studies how simple statistics from data, such as co-occurrences affects model predictions, which are described in the paper [Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions](https://arxiv.org/abs/2207.14251). This is RoBERTa-base epoch_2. ## Model Description This model was captured during a reproduction of [RoBERTa-base](https://huggingface.co/roberta-base), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM). The intended uses, limitations, training data and training procedure for the fully trained model are similar to [RoBERTa-base](https://huggingface.co/roberta-base). Two major differences with the original model: * We trained our model for 100K steps, instead of 500K * We only use Wikipedia and the Book Corpus, as corpora which are publicly available. ### How to use Using code from [RoBERTa-base](https://huggingface.co/roberta-base), here is an example based on PyTorch: ``` from transformers import pipeline model = pipeline("fill-mask", model='yanaiela/roberta-base-epoch_83', device=-1, top_k=10) model("Hello, I'm the <mask> RoBERTa-base language model") ``` ## Citation info ```bibtex @article{2207.14251, Author = {Yanai Elazar and Nora Kassner and Shauli Ravfogel and Amir Feder and Abhilasha Ravichander and Marius Mosbach and Yonatan Belinkov and Hinrich Schütze and Yoav Goldberg}, Title = {Measuring Causal Effects of Data Statistics on Language Model's `Factual' Predictions}, Year = {2022}, Eprint = {arXiv:2207.14251}, } ```
platzi/platzi-distilroberta-base-mrpc-glue-omar-espejel
platzi
2022-07-29T21:57:21Z
15
1
transformers
[ "transformers", "pytorch", "tensorboard", "roberta", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-07-29T12:17:21Z
--- license: apache-2.0 tags: - text-classification - generated_from_trainer datasets: - glue metrics: - accuracy - f1 widget: - text: ["Yucaipa owned Dominick 's before selling the chain to Safeway in 1998 for $ 2.5 billion.","Yucaipa bought Dominick's in 1995 for $ 693 million and sold it to Safeway for $ 1.8 billion in 1998."] example_title: Not Equivalent - text: ["Revenue in the first quarter of the year dropped 15 percent from the same period a year earlier.", "With the scandal hanging over Stewart's company revenue the first quarter of the year dropped 15 percent from the same period a year earlier."] example_title: Equivalent model-index: - name: platzi-distilroberta-base-mrpc-glue-omar-espejel results: - task: name: Text Classification type: text-classification dataset: name: glue type: glue config: mrpc split: train args: mrpc metrics: - name: Accuracy type: accuracy value: 0.8431372549019608 - name: F1 type: f1 value: 0.8861209964412811 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # platzi-distilroberta-base-mrpc-glue-omar-espejel This model is a fine-tuned version of [distilroberta-base](https://huggingface.co/distilroberta-base) on the glue and the mrpc datasets. It achieves the following results on the evaluation set: - Loss: 0.6332 - Accuracy: 0.8431 - F1: 0.8861 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.5076 | 1.09 | 500 | 0.7464 | 0.8137 | 0.8671 | | 0.3443 | 2.18 | 1000 | 0.6332 | 0.8431 | 0.8861 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
mrm8488/q-Taxi-v3
mrm8488
2022-07-29T21:37:20Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2022-07-29T20:43:55Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-Taxi-v3 results: - metrics: - type: mean_reward value: 7.56 +/- 2.71 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 --- # **Q-Learning** Agent playing **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="mrm8488/q-Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) evaluate_agent(env, model["max_steps"], model["n_eval_episodes"], model["qtable"], model["eval_seed"]) ```
natalierobbins/pos_test_model_1
natalierobbins
2022-07-29T19:21:52Z
3
0
transformers
[ "transformers", "pytorch", "distilbert", "token-classification", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-07-29T15:29:09Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy - f1 - precision - recall model-index: - name: pos_test_model_1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # pos_test_model_1 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.1521 - Accuracy: 0.9530 - F1: 0.9523 - Precision: 0.9576 - Recall: 0.9530 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:| | 0.1882 | 1.0 | 1744 | 0.1521 | 0.9530 | 0.9523 | 0.9576 | 0.9530 | ### Framework versions - Transformers 4.18.0 - Pytorch 1.10.2 - Datasets 2.2.2 - Tokenizers 0.12.1
andres-hsn/q-Taxi-v3
andres-hsn
2022-07-29T17:02:52Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2022-07-29T17:02:38Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-Taxi-v3 results: - metrics: - type: mean_reward value: 7.54 +/- 2.72 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 --- # **Q-Learning** Agent playing **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="andres-hsn/q-Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) evaluate_agent(env, model["max_steps"], model["n_eval_episodes"], model["qtable"], model["eval_seed"]) ```
andres-hsn/q-FrozenLake-v1-4x4-noSlippery
andres-hsn
2022-07-29T17:00:32Z
0
0
null
[ "FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2022-07-29T16:58:29Z
--- tags: - FrozenLake-v1-4x4-no_slippery - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-FrozenLake-v1-4x4-noSlippery results: - metrics: - type: mean_reward value: 1.00 +/- 0.00 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: FrozenLake-v1-4x4-no_slippery type: FrozenLake-v1-4x4-no_slippery --- # **Q-Learning** Agent playing **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="andres-hsn/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) evaluate_agent(env, model["max_steps"], model["n_eval_episodes"], model["qtable"], model["eval_seed"]) ```
Datasaur/distilbert-base-uncased-finetuned-ag-news
Datasaur
2022-07-29T16:36:20Z
5
0
transformers
[ "transformers", "pytorch", "distilbert", "text-classification", "en", "dataset:ag-news", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-05-17T02:53:35Z
--- language: en license: apache-2.0 datasets: - ag-news ---
pampa/pets
pampa
2022-07-29T16:20:29Z
0
0
null
[ "license:apache-2.0", "region:us" ]
null
2022-07-29T14:56:39Z
--- title: Pet classifier! emoji: 🐶 colorFrom: pink colorTo: blue sdk: gradio sdk_version: 2.9.4 app_file: app.py pinned: false license: apache-2.0 --- Check out the configuration reference at https://huggingface.co/docs/hub/spaces#reference
jirtan/ddpm-ema-pokemon-64
jirtan
2022-07-29T16:20:10Z
4
0
diffusers
[ "diffusers", "en", "dataset:huggan/pokemon", "license:apache-2.0", "diffusers:DDPMPipeline", "region:us" ]
null
2022-07-29T15:20:10Z
--- language: en license: apache-2.0 library_name: diffusers tags: [] datasets: huggan/pokemon metrics: [] --- <!-- This model card has been generated automatically according to the information the training script had access to. You should probably proofread and complete it, then remove this comment. --> # ddpm-ema-pokemon-64 ## Model description This diffusion model is trained with the [🤗 Diffusers](https://github.com/huggingface/diffusers) library on the `huggan/pokemon` dataset. ## Intended uses & limitations #### How to use ```python # TODO: add an example code snippet for running this diffusion pipeline ``` #### Limitations and bias [TODO: provide examples of latent issues and potential remediations] ## Training data [TODO: describe the data used to train the model] ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 16 - eval_batch_size: 16 - gradient_accumulation_steps: 1 - optimizer: AdamW with betas=(0.95, 0.999), weight_decay=1e-06 and epsilon=1e-08 - lr_scheduler: cosine - lr_warmup_steps: 500 - ema_inv_gamma: 1.0 - ema_inv_gamma: 0.75 - ema_inv_gamma: 0.9999 - mixed_precision: no ### Training results 📈 [TensorBoard logs](https://huggingface.co/jirtan/ddpm-ema-pokemon-64/tensorboard?#scalars)
kdf/python-docstring-generation
kdf
2022-07-29T15:31:02Z
6
3
transformers
[ "transformers", "pytorch", "codegen", "text-generation", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2022-07-29T11:51:57Z
--- license: apache-2.0 widget: - text: "<|endoftext|>\ndef load_excel(path):\n return pd.read_excel(path)\n# docstring\n\"\"\"" --- ## Basic info model based [Salesforce/codegen-350M-mono](https://huggingface.co/Salesforce/codegen-350M-mono) fine-tuned with data [codeparrot/github-code-clean](https://huggingface.co/datasets/codeparrot/github-code-clean) data filter by python ## Usage ```python from transformers import AutoTokenizer, AutoModelForCausalLM model_type = 'kdf/python-docstring-generation' tokenizer = AutoTokenizer.from_pretrained(model_type) model = AutoModelForCausalLM.from_pretrained(model_type) inputs = tokenizer('''<|endoftext|> def load_excel(path): return pd.read_excel(path) # docstring """''', return_tensors='pt') doc_max_length = 128 generated_ids = model.generate( **inputs, max_length=inputs.input_ids.shape[1] + doc_max_length, do_sample=False, return_dict_in_generate=True, num_return_sequences=1, output_scores=True, pad_token_id=50256, eos_token_id=50256 # <|endoftext|> ) ret = tokenizer.decode(generated_ids.sequences[0], skip_special_tokens=False) print(ret) ``` ## Prompt You could give model a style or a specific language, for example: ```python inputs = tokenizer('''<|endoftext|> def add(a, b): return a + b # docstring """ Calculate numbers add. Args: a: the first number to add b: the second number to add Return: The result of a + b """ <|endoftext|> def load_excel(path): return pd.read_excel(path) # docstring """''', return_tensors='pt') doc_max_length = 128 generated_ids = model.generate( **inputs, max_length=inputs.input_ids.shape[1] + doc_max_length, do_sample=False, return_dict_in_generate=True, num_return_sequences=1, output_scores=True, pad_token_id=50256, eos_token_id=50256 # <|endoftext|> ) ret = tokenizer.decode(generated_ids.sequences[0], skip_special_tokens=False) print(ret) inputs = tokenizer('''<|endoftext|> def add(a, b): return a + b # docstring """ 计算数字相加 Args: a: 第一个加数 b: 第二个加数 Return: 相加的结果 """ <|endoftext|> def load_excel(path): return pd.read_excel(path) # docstring """''', return_tensors='pt') doc_max_length = 128 generated_ids = model.generate( **inputs, max_length=inputs.input_ids.shape[1] + doc_max_length, do_sample=False, return_dict_in_generate=True, num_return_sequences=1, output_scores=True, pad_token_id=50256, eos_token_id=50256 # <|endoftext|> ) ret = tokenizer.decode(generated_ids.sequences[0], skip_special_tokens=False) print(ret) ```
schnell/bert-small-juman-bpe
schnell
2022-07-29T15:15:11Z
3
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "fill-mask", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-26T16:12:28Z
--- tags: - generated_from_trainer metrics: - accuracy model-index: - name: bert-small-juman-bpe results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-small-juman-bpe This model is a fine-tuned version of [](https://huggingface.co/) on the None dataset. It achieves the following results on the evaluation set: - Accuracy: 0.6317 - Loss: 1.7829 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 256 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 3 - total_train_batch_size: 768 - total_eval_batch_size: 24 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.01 - num_epochs: 14 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Accuracy | Validation Loss | |:-------------:|:-----:|:------:|:--------:|:---------------:| | 2.3892 | 1.0 | 69472 | 0.5637 | 2.2498 | | 2.2219 | 2.0 | 138944 | 0.5873 | 2.0785 | | 2.1453 | 3.0 | 208416 | 0.5984 | 2.0019 | | 2.1 | 4.0 | 277888 | 0.6059 | 1.9531 | | 2.068 | 5.0 | 347360 | 0.6106 | 1.9169 | | 2.0405 | 6.0 | 416832 | 0.6146 | 1.8921 | | 2.0174 | 7.0 | 486304 | 0.6175 | 1.8711 | | 2.0002 | 8.0 | 555776 | 0.6205 | 1.8527 | | 1.9838 | 9.0 | 625248 | 0.6225 | 1.8381 | | 1.9691 | 10.0 | 694720 | 0.6248 | 1.8239 | | 1.9551 | 11.0 | 764192 | 0.6265 | 1.8125 | | 1.9406 | 12.0 | 833664 | 0.6288 | 1.8002 | | 1.9293 | 13.0 | 903136 | 0.6310 | 1.7871 | | 1.9247 | 14.0 | 972608 | 0.6317 | 1.7829 | ### Framework versions - Transformers 4.19.2 - Pytorch 1.12.0+cu116 - Datasets 2.2.2 - Tokenizers 0.12.1
silviacamplani/distilbert-uncase-direct-finetuning-ai-ner_3labels
silviacamplani
2022-07-29T14:41:55Z
3
0
transformers
[ "transformers", "tf", "distilbert", "token-classification", "generated_from_keras_callback", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-07-29T14:33:10Z
--- license: apache-2.0 tags: - generated_from_keras_callback model-index: - name: silviacamplani/distilbert-uncase-direct-finetuning-ai-ner_3labels results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # silviacamplani/distilbert-uncase-direct-finetuning-ai-ner_3labels This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.6593 - Validation Loss: 0.6130 - Epoch: 9 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'inner_optimizer': {'class_name': 'AdamWeightDecay', 'config': {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 1e-05, 'decay_steps': 60, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}}, 'dynamic': True, 'initial_scale': 32768.0, 'dynamic_growth_steps': 2000} - training_precision: mixed_float16 ### Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 1.9721 | 1.8113 | 0 | | 1.6564 | 1.5052 | 1 | | 1.3640 | 1.2332 | 2 | | 1.1078 | 0.9996 | 3 | | 0.9158 | 0.8249 | 4 | | 0.7850 | 0.7188 | 5 | | 0.7135 | 0.6595 | 6 | | 0.6822 | 0.6310 | 7 | | 0.6394 | 0.6171 | 8 | | 0.6593 | 0.6130 | 9 | ### Framework versions - Transformers 4.20.1 - TensorFlow 2.6.4 - Datasets 2.1.0 - Tokenizers 0.12.1
Amine007/distilgpt2-finetuned-wikitext2
Amine007
2022-07-29T14:15:42Z
3
0
transformers
[ "transformers", "pytorch", "tensorboard", "gpt2", "text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2022-07-29T13:24:32Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: distilgpt2-finetuned-wikitext2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilgpt2-finetuned-wikitext2 This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the None dataset. It achieves the following results on the evaluation set: - Loss: 3.6421 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 3.7602 | 1.0 | 2334 | 3.6669 | | 3.653 | 2.0 | 4668 | 3.6472 | | 3.6006 | 3.0 | 7002 | 3.6421 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
platzi/platzi-bert-base-mrpc-glue-omar-espejel
platzi
2022-07-29T13:50:27Z
4
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-07-29T13:37:08Z
--- license: apache-2.0 tags: - text-classification - generated_from_trainer datasets: - glue metrics: - accuracy - f1 model-index: - name: platzi-bert-base-mrpc-glue-omar-espejel results: - task: name: Text Classification type: text-classification dataset: name: glue type: glue config: mrpc split: train args: mrpc metrics: - name: Accuracy type: accuracy value: 0.8578431372549019 - name: F1 type: f1 value: 0.8941605839416058 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # platzi-bert-base-mrpc-glue-omar-espejel This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the glue and the mrpc datasets. It achieves the following results on the evaluation set: - Loss: 0.4366 - Accuracy: 0.8578 - F1: 0.8942 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.5221 | 1.09 | 500 | 0.4366 | 0.8578 | 0.8942 | | 0.3114 | 2.18 | 1000 | 0.6581 | 0.8725 | 0.9113 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
raisin2402/marian-finetuned-kde4-en-to-fr
raisin2402
2022-07-29T12:59:05Z
3
1
transformers
[ "transformers", "pytorch", "tensorboard", "marian", "text2text-generation", "translation", "generated_from_trainer", "dataset:kde4", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-07-29T11:08:39Z
--- license: apache-2.0 tags: - translation - generated_from_trainer datasets: - kde4 metrics: - bleu model-index: - name: marian-finetuned-kde4-en-to-fr results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: kde4 type: kde4 config: en-fr split: train args: en-fr metrics: - name: Bleu type: bleu value: 52.83113187001415 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # marian-finetuned-kde4-en-to-fr This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset. It achieves the following results on the evaluation set: - Loss: 0.8560 - Bleu: 52.8311 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
marii/lunarlander
marii
2022-07-29T12:31:04Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-07-29T09:25:07Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - metrics: - type: mean_reward value: 278.03 +/- 20.09 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
turhancan97/dqn-SpaceInvadersNoFrameskip-v4
turhancan97
2022-07-29T12:12:16Z
4
0
stable-baselines3
[ "stable-baselines3", "SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-07-29T12:11:45Z
--- library_name: stable-baselines3 tags: - SpaceInvadersNoFrameskip-v4 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: DQN results: - metrics: - type: mean_reward value: 424.00 +/- 124.70 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: SpaceInvadersNoFrameskip-v4 type: SpaceInvadersNoFrameskip-v4 --- # **DQN** Agent playing **SpaceInvadersNoFrameskip-v4** This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib ``` # Download model and save it into the logs/ folder python -m utils.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga turhancan97 -f logs/ python enjoy.py --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` ## Training (with the RL Zoo) ``` python train.py --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ # Upload the model and generate video (when possible) python -m utils.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga turhancan97 ``` ## Hyperparameters ```python OrderedDict([('batch_size', 32), ('buffer_size', 100000), ('env_wrapper', ['stable_baselines3.common.atari_wrappers.AtariWrapper']), ('exploration_final_eps', 0.01), ('exploration_fraction', 0.1), ('frame_stack', 4), ('gradient_steps', 1), ('learning_rate', 0.0001), ('learning_starts', 100000), ('n_timesteps', 500000.0), ('optimize_memory_usage', False), ('policy', 'CnnPolicy'), ('target_update_interval', 1000), ('train_freq', 4), ('normalize', False)]) ```
AlbertShu/Reinforce-v1
AlbertShu
2022-07-29T11:26:16Z
0
0
null
[ "CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2022-07-29T11:26:01Z
--- tags: - CartPole-v1 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-v1 results: - metrics: - type: mean_reward value: 500.00 +/- 0.00 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 --- # **Reinforce** Agent playing **CartPole-v1** This is a trained model of a **Reinforce** agent playing **CartPole-v1** . To learn to use this model and train yours check Unit 5 of the Deep Reinforcement Learning Class: https://github.com/huggingface/deep-rl-class/tree/main/unit5
Lvxue/finetuned-mt5-small
Lvxue
2022-07-29T11:08:43Z
26
0
transformers
[ "transformers", "pytorch", "mt5", "text2text-generation", "generated_from_trainer", "en", "ro", "dataset:wmt16", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2022-07-28T02:27:31Z
--- language: - en - ro license: apache-2.0 tags: - generated_from_trainer datasets: - wmt16 metrics: - bleu model-index: - name: finetuned-mt5-small results: - task: name: Translation type: translation dataset: name: wmt16 ro-en type: wmt16 args: ro-en metrics: - name: Bleu type: bleu value: 23.6759 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # finetuned-mt5-small This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on the wmt16 ro-en dataset. It achieves the following results on the evaluation set: - Loss: 1.6328 - Bleu: 23.6759 - Gen Len: 43.6993 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 12 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10.0 ### Training results ### Framework versions - Transformers 4.20.1 - Pytorch 1.12.0+cu102 - Datasets 2.3.2 - Tokenizers 0.12.1
bkaemper/dqn-SpaceInvadersNoFrameskip-v4
bkaemper
2022-07-29T10:23:02Z
1
0
stable-baselines3
[ "stable-baselines3", "SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-07-29T08:43:58Z
--- library_name: stable-baselines3 tags: - SpaceInvadersNoFrameskip-v4 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: DQN results: - metrics: - type: mean_reward value: 774.50 +/- 288.79 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: SpaceInvadersNoFrameskip-v4 type: SpaceInvadersNoFrameskip-v4 --- # **DQN** Agent playing **SpaceInvadersNoFrameskip-v4** This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib ``` # Download model and save it into the logs/ folder python -m utils.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga bkaemper -f logs/ python enjoy.py --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` ## Training (with the RL Zoo) ``` python train.py --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ # Upload the model and generate video (when possible) python -m utils.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga bkaemper ``` ## Hyperparameters ```python OrderedDict([('batch_size', 32), ('buffer_size', 100000), ('env_wrapper', ['stable_baselines3.common.atari_wrappers.AtariWrapper']), ('exploration_final_eps', 0.01), ('exploration_fraction', 0.1), ('frame_stack', 4), ('gradient_steps', 1), ('learning_rate', 0.0001), ('learning_starts', 100000), ('n_timesteps', 1000000.0), ('optimize_memory_usage', False), ('policy', 'CnnPolicy'), ('target_update_interval', 1000), ('train_freq', 4), ('normalize', False)]) ```
AbidHasan95/movieHunt4-ner
AbidHasan95
2022-07-29T09:53:57Z
3
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "token-classification", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-07-29T09:02:11Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - precision - recall - f1 - accuracy model-index: - name: movieHunt4-ner results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # movieHunt4-ner This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0005 - Precision: 1.0 - Recall: 1.0 - F1: 1.0 - Accuracy: 1.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | No log | 1.0 | 48 | 0.0284 | 0.9959 | 0.9959 | 0.9959 | 0.9974 | | No log | 2.0 | 96 | 0.0060 | 1.0 | 1.0 | 1.0 | 1.0 | | No log | 3.0 | 144 | 0.0034 | 1.0 | 1.0 | 1.0 | 1.0 | | No log | 4.0 | 192 | 0.0025 | 1.0 | 1.0 | 1.0 | 1.0 | | No log | 5.0 | 240 | 0.0020 | 1.0 | 1.0 | 1.0 | 1.0 | | No log | 6.0 | 288 | 0.0016 | 1.0 | 1.0 | 1.0 | 1.0 | | No log | 7.0 | 336 | 0.0014 | 1.0 | 1.0 | 1.0 | 1.0 | | No log | 8.0 | 384 | 0.0012 | 1.0 | 1.0 | 1.0 | 1.0 | | No log | 9.0 | 432 | 0.0011 | 1.0 | 1.0 | 1.0 | 1.0 | | No log | 10.0 | 480 | 0.0010 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0168 | 11.0 | 528 | 0.0009 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0168 | 12.0 | 576 | 0.0009 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0168 | 13.0 | 624 | 0.0008 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0168 | 14.0 | 672 | 0.0008 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0168 | 15.0 | 720 | 0.0008 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0168 | 16.0 | 768 | 0.0007 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0168 | 17.0 | 816 | 0.0007 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0168 | 18.0 | 864 | 0.0007 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0168 | 19.0 | 912 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0168 | 20.0 | 960 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0014 | 21.0 | 1008 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0014 | 22.0 | 1056 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0014 | 23.0 | 1104 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0014 | 24.0 | 1152 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0014 | 25.0 | 1200 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0014 | 26.0 | 1248 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0014 | 27.0 | 1296 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0014 | 28.0 | 1344 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0014 | 29.0 | 1392 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 | | 0.0014 | 30.0 | 1440 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
LanaKru/wikineural-multilingual-ner-finetuned-ner
LanaKru
2022-07-29T09:36:52Z
10
1
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "generated_from_trainer", "dataset:skript", "license:cc-by-nc-sa-4.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-07-29T04:14:38Z
--- license: cc-by-nc-sa-4.0 tags: - generated_from_trainer datasets: - skript metrics: - precision - recall - f1 - accuracy model-index: - name: wikineural-multilingual-ner-finetuned-ner results: - task: name: Token Classification type: token-classification dataset: name: skript type: skript config: myscript split: train args: myscript metrics: - name: Precision type: precision value: 0.9007335298553506 - name: Recall type: recall value: 0.9301946902654867 - name: F1 type: f1 value: 0.9152270827528559 - name: Accuracy type: accuracy value: 0.9653644982020269 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wikineural-multilingual-ner-finetuned-ner This model is a fine-tuned version of [Babelscape/wikineural-multilingual-ner](https://huggingface.co/Babelscape/wikineural-multilingual-ner) on the skript dataset. It achieves the following results on the evaluation set: - Loss: 0.1243 - Precision: 0.9007 - Recall: 0.9302 - F1: 0.9152 - Accuracy: 0.9654 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | No log | 1.0 | 298 | 0.1179 | 0.8975 | 0.8981 | 0.8978 | 0.9592 | | 0.104 | 2.0 | 596 | 0.1161 | 0.9051 | 0.9201 | 0.9126 | 0.9648 | | 0.104 | 3.0 | 894 | 0.1243 | 0.9007 | 0.9302 | 0.9152 | 0.9654 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
SummerChiam/pond_image_classification_9
SummerChiam
2022-07-29T09:13:48Z
51
0
transformers
[ "transformers", "pytorch", "tensorboard", "vit", "image-classification", "huggingpics", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2022-07-29T09:13:31Z
--- tags: - image-classification - pytorch - huggingpics metrics: - accuracy model-index: - name: pond_image_classification_9 results: - task: name: Image Classification type: image-classification metrics: - name: Accuracy type: accuracy value: 0.9974489808082581 --- # pond_image_classification_9 Autogenerated by HuggingPics🤗🖼️ Create your own image classifier for **anything** by running [the demo on Google Colab](https://colab.research.google.com/github/nateraw/huggingpics/blob/main/HuggingPics.ipynb). Report any issues with the demo at the [github repo](https://github.com/nateraw/huggingpics). ## Example Images #### Algae ![Algae](images/Algae.png) #### Boiling ![Boiling](images/Boiling.png) #### BoilingNight ![BoilingNight](images/BoilingNight.png) #### Normal ![Normal](images/Normal.png) #### NormalCement ![NormalCement](images/NormalCement.png) #### NormalNight ![NormalNight](images/NormalNight.png) #### NormalRain ![NormalRain](images/NormalRain.png)
Go2Heart/BERT_Mod_3
Go2Heart
2022-07-29T09:11:43Z
4
0
transformers
[ "transformers", "pytorch", "distilbert", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-07-29T07:36:44Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - glue metrics: - accuracy model-index: - name: BERT_Mod_3 results: - task: name: Text Classification type: text-classification dataset: name: glue type: glue args: mnli metrics: - name: Accuracy type: accuracy value: 0.8198675496688742 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # BERT_Mod_3 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.6760 - Accuracy: 0.8199 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.5167 | 1.0 | 24544 | 0.4953 | 0.8077 | | 0.414 | 2.0 | 49088 | 0.4802 | 0.8148 | | 0.2933 | 3.0 | 73632 | 0.5783 | 0.8186 | | 0.2236 | 4.0 | 98176 | 0.6760 | 0.8199 | ### Framework versions - Transformers 4.20.1 - Pytorch 1.12.0 - Datasets 2.4.0 - Tokenizers 0.12.1
mselbach/distilbart-rehadat
mselbach
2022-07-29T09:01:51Z
6
0
transformers
[ "transformers", "pytorch", "bart", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2022-07-29T07:54:08Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: distilbart-summarization results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbart-summarization This model is a fine-tuned version of [sshleifer/distilbart-cnn-12-6](https://huggingface.co/sshleifer/distilbart-cnn-12-6) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.0+cpu - Datasets 1.17.0 - Tokenizers 0.11.0
RRajesh27/finetuning-sentiment-model-3000-samples
RRajesh27
2022-07-29T08:51:28Z
5
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:imdb", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-07-29T08:39:10Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imdb metrics: - accuracy - f1 model-index: - name: finetuning-sentiment-model-3000-samples results: - task: name: Text Classification type: text-classification dataset: name: imdb type: imdb config: plain_text split: train args: plain_text metrics: - name: Accuracy type: accuracy value: 0.8666666666666667 - name: F1 type: f1 value: 0.8666666666666667 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # finetuning-sentiment-model-3000-samples This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the imdb dataset. It achieves the following results on the evaluation set: - Loss: 0.3236 - Accuracy: 0.8667 - F1: 0.8667 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
tosin/dialogpt_afriwoz_pidgin
tosin
2022-07-29T08:41:47Z
18
2
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "en", "dataset:AfriWOZ", "arxiv:2204.08083", "license:cc-by-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2022-07-29T08:00:24Z
--- thumbnail: https://huggingface.co/front/thumbnails/dialogpt.png language: - en license: cc-by-4.0 tags: - conversational - transformers datasets: - AfriWOZ metrics: - perplexity widget: - text: "How I fit chop for here?" --- ## DialoGPT_AfriWOZ (Pidgin) This is a fine-tuned model of DialoGPT (small) on the AfriWOZ dataset. It is intended to be used as a conversational system in Nigeria Pidgin English language. The dataset it's trained on is limited in scope, as it covers only certain domains such as restaurants, hotel, taxi, and booking. The perplexity achieved on the validation set is 38.52. * Generation example from an interactive environment: |Role | Response | |------|------------| |User | I hear say restaurant dey here. | |Bot | I dey find restaurant. | |User | Abeg you fit tell me which kind chop dey? | |Bot | I no dey sure right now. | |User | You do well. Thank you. | |Bot | I dey find restauarant. | Please find the information about preprocessing, training and full details of the DialoGPT in the [original DialoGPT repository](https://github.com/microsoft/DialoGPT) The paper for this work can be found on arXiv: [https://arxiv.org/pdf/2204.08083.pdf](https://arxiv.org/pdf/2204.08083.pdf) ### How to use Now we are ready to try out how the model works as a chatting partner! ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch tokenizer = AutoTokenizer.from_pretrained("tosin/dialogpt_afriwoz_pidgin") model = AutoModelForCausalLM.from_pretrained("tosin/dialogpt_afriwoz_pidgin") # Let's chat for 5 lines for step in range(5): # encode the new user input, add the eos_token and return a tensor in Pytorch new_user_input_ids = tokenizer.encode(input(">> User:") + tokenizer.eos_token, return_tensors='pt') # append the new user input tokens to the chat history bot_input_ids = torch.cat([chat_history_ids, new_user_input_ids], dim=-1) if step > 0 else new_user_input_ids # generated a response while limiting the total chat history to 1000 tokens, chat_history_ids = model.generate(bot_input_ids, max_length=1000, pad_token_id=tokenizer.eos_token_id) # pretty print last ouput tokens from bot print("DialoGPT_pidgin_Bot: {}".format(tokenizer.decode(chat_history_ids[:, bot_input_ids.shape[-1]:][0], skip_special_tokens=True)))
SummerChiam/pond_image_classification_7
SummerChiam
2022-07-29T08:32:46Z
48
0
transformers
[ "transformers", "pytorch", "tensorboard", "vit", "image-classification", "huggingpics", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2022-07-29T08:32:27Z
--- tags: - image-classification - pytorch - huggingpics metrics: - accuracy model-index: - name: pond_image_classification_7 results: - task: name: Image Classification type: image-classification metrics: - name: Accuracy type: accuracy value: 0.9936224222183228 --- # pond_image_classification_7 Autogenerated by HuggingPics🤗🖼️ Create your own image classifier for **anything** by running [the demo on Google Colab](https://colab.research.google.com/github/nateraw/huggingpics/blob/main/HuggingPics.ipynb). Report any issues with the demo at the [github repo](https://github.com/nateraw/huggingpics). ## Example Images #### Algae ![Algae](images/Algae.png) #### Boiling ![Boiling](images/Boiling.png) #### BoilingNight ![BoilingNight](images/BoilingNight.png) #### Normal ![Normal](images/Normal.png) #### NormalCement ![NormalCement](images/NormalCement.png) #### NormalNight ![NormalNight](images/NormalNight.png) #### NormalRain ![NormalRain](images/NormalRain.png)
SummerChiam/pond_image_classification_6
SummerChiam
2022-07-29T08:19:54Z
50
0
transformers
[ "transformers", "pytorch", "tensorboard", "vit", "image-classification", "huggingpics", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2022-07-29T08:19:36Z
--- tags: - image-classification - pytorch - huggingpics metrics: - accuracy model-index: - name: pond_image_classification_6 results: - task: name: Image Classification type: image-classification metrics: - name: Accuracy type: accuracy value: 0.9948979616165161 --- # pond_image_classification_6 Autogenerated by HuggingPics🤗🖼️ Create your own image classifier for **anything** by running [the demo on Google Colab](https://colab.research.google.com/github/nateraw/huggingpics/blob/main/HuggingPics.ipynb). Report any issues with the demo at the [github repo](https://github.com/nateraw/huggingpics). ## Example Images #### Algae ![Algae](images/Algae.png) #### Boiling ![Boiling](images/Boiling.png) #### BoilingNight ![BoilingNight](images/BoilingNight.png) #### Normal ![Normal](images/Normal.png) #### NormalCement ![NormalCement](images/NormalCement.png) #### NormalNight ![NormalNight](images/NormalNight.png) #### NormalRain ![NormalRain](images/NormalRain.png)
ParkSaeroyi/distilroberta-base-finetuned-wikitext2
ParkSaeroyi
2022-07-29T08:10:16Z
3
0
transformers
[ "transformers", "pytorch", "tensorboard", "roberta", "fill-mask", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T10:00:51Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: distilroberta-base-finetuned-wikitext2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilroberta-base-finetuned-wikitext2 This model is a fine-tuned version of [distilroberta-base](https://huggingface.co/distilroberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 8.3687 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 6 | 8.8622 | | No log | 2.0 | 12 | 8.4576 | | No log | 3.0 | 18 | 8.4412 | ### Framework versions - Transformers 4.22.0.dev0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
pkufool/icefall_librispeech_streaming_pruned_transducer_stateless5_20220729
pkufool
2022-07-29T08:08:41Z
0
0
null
[ "tensorboard", "license:apache-2.0", "region:us" ]
null
2022-07-29T07:42:03Z
--- license: apache-2.0 --- See https://github.com/k2-fsa/icefall/pull/454 ### training command: ```bash ./pruned_transducer_stateless5/train.py \ --exp-dir pruned_transducer_stateless5/exp \ --num-encoder-layers 18 \ --dim-feedforward 2048 \ --nhead 8 \ --encoder-dim 512 \ --decoder-dim 512 \ --joiner-dim 512 \ --full-libri 1 \ --dynamic-chunk-training 1 \ --causal-convolution 1 \ --short-chunk-size 20 \ --num-left-chunks 4 \ --max-duration 300 \ --world-size 4 \ --start-epoch 1 \ --num-epochs 25 ``` You can find the tensorboard log here <https://tensorboard.dev/experiment/rO04h6vjTLyw0qSxjp4m4Q> ### The decoding command is: ```bash decoding_method="greedy_search" # "fast_beam_search", "modified_beam_search" for chunk in 2 4 8 16; do for left in 32 64; do ./pruned_transducer_stateless5/decode.py \ --num-encoder-layers 18 \ --dim-feedforward 2048 \ --nhead 8 \ --encoder-dim 512 \ --decoder-dim 512 \ --joiner-dim 512 \ --simulate-streaming 1 \ --decode-chunk-size ${chunk} \ --left-context ${left} \ --causal-convolution 1 \ --epoch 25 \ --avg 5 \ --exp-dir ./pruned_transducer_stateless5/exp \ --max-sym-per-frame 1 \ --max-duration 1000 \ --decoding-method ${decoding_method} done done ``` ### export command is: ```bash ./pruned_transducer_stateless5/export.py \ --streaming-model 1 \ --causal-convolution 1 \ --num-encoder-layers 18 \ --dim-feedforward 2048 \ --nhead 8 \ --encoder-dim 512 \ --decoder-dim 512 \ --joiner-dim 512 \ --epoch 25 \ --avg 5 \ --exp-dir ./pruned_transducer_stateless5/exp ./pruned_transducer_stateless5/export.py \ --streaming-model 1 \ --causal-convolution 1 \ --num-encoder-layers 18 \ --dim-feedforward 2048 \ --nhead 8 \ --encoder-dim 512 \ --decoder-dim 512 \ --joiner-dim 512 \ --epoch 25 \ --avg 5 \ --exp-dir ./pruned_transducer_stateless5/exp \ --jit 1 ```
Doohae/lassl-koelectra-small
Doohae
2022-07-29T07:28:48Z
1
0
transformers
[ "transformers", "pytorch", "electra", "pretraining", "endpoints_compatible", "region:us" ]
null
2022-07-29T06:50:48Z
# ELECTRA discriminator small - pretrained with large Korean corpus datasets (30GB) - 13.7M model parameters (followed google/electra-small-discriminator config) - 32,000 vocab size - trained for 1,000,000 steps - build with [lassl](https://github.com/lassl/lassl) framework pretrain-data ┣ korean_corpus.txt ┣ kowiki_latest.txt ┣ modu_dialogue_v1.2.txt ┣ modu_news_v1.1.txt ┣ modu_news_v2.0.txt ┣ modu_np_2021_v1.0.txt ┣ modu_np_v1.1.txt ┣ modu_spoken_v1.2.txt ┗ modu_written_v1.0.txt
SummerChiam/pond_image_classification_3
SummerChiam
2022-07-29T07:03:07Z
50
0
transformers
[ "transformers", "pytorch", "tensorboard", "vit", "image-classification", "huggingpics", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2022-07-29T07:02:53Z
--- tags: - image-classification - pytorch - huggingpics metrics: - accuracy model-index: - name: pond_image_classification_3 results: - task: name: Image Classification type: image-classification metrics: - name: Accuracy type: accuracy value: 0.9974489808082581 --- # pond_image_classification_3 Autogenerated by HuggingPics🤗🖼️ Create your own image classifier for **anything** by running [the demo on Google Colab](https://colab.research.google.com/github/nateraw/huggingpics/blob/main/HuggingPics.ipynb). Report any issues with the demo at the [github repo](https://github.com/nateraw/huggingpics). ## Example Images #### Algae ![Algae](images/Algae.png) #### Boiling ![Boiling](images/Boiling.png) #### BoilingNight ![BoilingNight](images/BoilingNight.png) #### Normal ![Normal](images/Normal.png) #### NormalCement ![NormalCement](images/NormalCement.png) #### NormalNight ![NormalNight](images/NormalNight.png) #### NormalRain ![NormalRain](images/NormalRain.png)
jianzhnie/a2c-v1-Walker2DBulletEnv-v0
jianzhnie
2022-07-29T06:53:25Z
3
0
stable-baselines3
[ "stable-baselines3", "tensorboard", "Walker2DBulletEnv-v0", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-07-29T06:52:47Z
--- library_name: stable-baselines3 tags: - Walker2DBulletEnv-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: A2C results: - metrics: - type: mean_reward value: 21.00 +/- 3.61 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: Walker2DBulletEnv-v0 type: Walker2DBulletEnv-v0 --- # **A2C** Agent playing **Walker2DBulletEnv-v0** This is a trained model of a **A2C** agent playing **Walker2DBulletEnv-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
SummerChiam/pond_image_classification_2
SummerChiam
2022-07-29T06:23:30Z
48
0
transformers
[ "transformers", "pytorch", "tensorboard", "vit", "image-classification", "huggingpics", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2022-07-29T06:23:17Z
--- tags: - image-classification - pytorch - huggingpics metrics: - accuracy model-index: - name: pond_image_classification_2 results: - task: name: Image Classification type: image-classification metrics: - name: Accuracy type: accuracy value: 0.9974489808082581 --- # pond_image_classification_2 Autogenerated by HuggingPics🤗🖼️ Create your own image classifier for **anything** by running [the demo on Google Colab](https://colab.research.google.com/github/nateraw/huggingpics/blob/main/HuggingPics.ipynb). Report any issues with the demo at the [github repo](https://github.com/nateraw/huggingpics). ## Example Images #### Algae ![Algae](images/Algae.png) #### Boiling ![Boiling](images/Boiling.png) #### BoilingNight ![BoilingNight](images/BoilingNight.png) #### Normal ![Normal](images/Normal.png) #### NormalCement ![NormalCement](images/NormalCement.png) #### NormalNight ![NormalNight](images/NormalNight.png) #### NormalRain ![NormalRain](images/NormalRain.png)
jianzhnie/a2c-v1-AntBulletEnv-v0
jianzhnie
2022-07-29T05:08:10Z
1
0
stable-baselines3
[ "stable-baselines3", "tensorboard", "AntBulletEnv-v0", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-07-29T05:07:17Z
--- library_name: stable-baselines3 tags: - AntBulletEnv-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: A2C results: - metrics: - type: mean_reward value: 674.59 +/- 89.58 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: AntBulletEnv-v0 type: AntBulletEnv-v0 --- # **A2C** Agent playing **AntBulletEnv-v0** This is a trained model of a **A2C** agent playing **AntBulletEnv-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
wpolatkan/ppo-LunarLander-v2
wpolatkan
2022-07-29T04:37:44Z
1
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-07-29T04:34:31Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - metrics: - type: mean_reward value: 244.25 +/- 15.32 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
oMateos2020/pegasus-newsroom-cnn1_50k
oMateos2020
2022-07-29T04:30:35Z
5
0
transformers
[ "transformers", "pytorch", "pegasus", "text2text-generation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2022-07-28T03:07:03Z
--- tags: - generated_from_trainer metrics: - rouge model-index: - name: pegasus-newsroom-cnn1_50k results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # pegasus-newsroom-cnn1_50k This model is a fine-tuned version of [oMateos2020/pegasus-newsroom-cnn1_50k](https://huggingface.co/oMateos2020/pegasus-newsroom-cnn1_50k) on the None dataset. It achieves the following results on the evaluation set: - Loss: 3.1267 - Rouge1: 38.0081 - Rouge2: 16.5536 - Rougel: 26.4916 - Rougelsum: 35.1349 - Gen Len: 59.4912 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 32 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 5 - mixed_precision_training: Native AMP - label_smoothing_factor: 0.1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:| | 3.144 | 0.26 | 100 | 3.0323 | 38.3168 | 16.7528 | 26.2646 | 35.2447 | 66.2372 | | 3.0556 | 0.51 | 200 | 3.0351 | 38.39 | 16.8027 | 26.3412 | 35.37 | 67.4676 | | 3.0701 | 0.77 | 300 | 3.0345 | 38.5742 | 16.922 | 26.3568 | 35.51 | 68.662 | | 3.1679 | 1.03 | 400 | 3.0321 | 38.5319 | 16.8049 | 26.4933 | 35.4775 | 65.976 | | 3.1041 | 1.28 | 500 | 3.0246 | 38.1381 | 16.63 | 26.2484 | 35.0999 | 64.6896 | | 3.0352 | 1.54 | 600 | 3.0206 | 38.9063 | 17.0281 | 27.0288 | 35.9175 | 59.0668 | | 3.0894 | 1.79 | 700 | 3.0251 | 38.4461 | 16.7732 | 26.4394 | 35.4807 | 63.2792 | | 3.0529 | 2.05 | 800 | 3.0400 | 38.5088 | 16.8921 | 26.5526 | 35.5236 | 64.294 | | 3.0002 | 2.31 | 900 | 3.0394 | 38.6899 | 16.8703 | 26.6771 | 35.6207 | 62.8004 | | 3.0167 | 2.56 | 1000 | 3.0394 | 38.3532 | 16.6176 | 26.5433 | 35.3282 | 61.63 | | 3.0168 | 2.82 | 1100 | 3.0421 | 38.7613 | 17.0107 | 26.8424 | 35.7508 | 62.67 | | 3.0412 | 3.08 | 1200 | 3.0491 | 38.6132 | 16.8046 | 26.61 | 35.6002 | 61.7924 | | 3.1273 | 3.33 | 1300 | 3.0823 | 38.5498 | 16.795 | 26.5569 | 35.613 | 60.6872 | | 3.0634 | 3.59 | 1400 | 3.1010 | 38.0927 | 16.4367 | 26.2315 | 35.1311 | 59.252 | | 3.097 | 3.84 | 1500 | 3.1147 | 37.7644 | 16.3156 | 26.2674 | 34.8315 | 59.7592 | | 3.1287 | 4.1 | 1600 | 3.1267 | 38.0081 | 16.5536 | 26.4916 | 35.1349 | 59.4912 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
publichealthsurveillance/PHS-BERT
publichealthsurveillance
2022-07-29T03:39:46Z
19
5
transformers
[ "transformers", "pytorch", "bert", "fill-mask", "arxiv:2204.04521", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-04-12T05:35:31Z
# PHS-BERT We present and release [PHS-BERT](https://arxiv.org/abs/2204.04521), a transformer-based pretrained language model (PLM), to identify tasks related to public health surveillance (PHS) on social media. Compared with existing PLMs that are mainly evaluated on limited tasks, PHS-BERT achieved state-of-the-art performance on 25 tested datasets, showing that our PLM is robust and generalizable in common PHS tasks. ## Usage Load the model via [Huggingface's Transformers library](https://github.com/huggingface/transformers]): ``` from transformers import AutoTokenizer, AutoModel tokenizer = AutoTokenizer.from_pretrained("publichealthsurveillance/PHS-BERT") model = AutoModel.from_pretrained("publichealthsurveillance/PHS-BERT") ``` ## Training Procedure ### Pretraining We followed the standard pretraining protocols of BERT and initialized PHS-BERT with weights from BERT during the training phase instead of training from scratch and used the uncased version of the BERT model. PHS-BERT is trained on a corpus of health-related tweets that were crawled via the Twitter API. Focusing on the tasks related to PHS, keywords used to collect pretraining corpus are set to disease, symptom, vaccine, and mental health-related words in English. Retweet tags were deleted from the raw corpus, and URLs and usernames were replaced with HTTP-URL and @USER, respectively. All emoticons were replaced with their associated meanings. Each sequence of BERT LM inputs is converted to 50,265 vocabulary tokens. Twitter posts are restricted to 200 characters, and during the training and evaluation phase, we used a batch size of 8. Distributed training was performed on a TPU v3-8. ### Fine-tuning We used the embedding of the special token [CLS] of the last hidden layer as the final feature of the input text. We adopted the multilayer perceptron (MLP) with the hyperbolic tangent activation function and used Adam optimizer. The models are trained with a one cycle policy at a maximum learning rate of 2e-05 with momentum cycled between 0.85 and 0.95. ## Societal Impact We train and release a PLM to accelerate the automatic identification of tasks related to PHS on social media. Our work aims to develop a new computational method for screening users in need of early intervention and is not intended to use in clinical settings or as a diagnostic tool. ## BibTex entry and citation info For more details, refer to the paper [Benchmarking for Public Health Surveillance tasks on Social Media with a Domain-Specific Pretrained Language Model](https://arxiv.org/abs/2204.04521). ``` @inproceedings{naseem-etal-2022-benchmarking, title = "Benchmarking for Public Health Surveillance tasks on Social Media with a Domain-Specific Pretrained Language Model", author = "Naseem, Usman and Lee, Byoung Chan and Khushi, Matloob and Kim, Jinman and Dunn, Adam", booktitle = "Proceedings of NLP Power! The First Workshop on Efficient Benchmarking in NLP", month = may, year = "2022", address = "Dublin, Ireland", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2022.nlppower-1.3", doi = "10.18653/v1/2022.nlppower-1.3", pages = "22--31", abstract = "A user-generated text on social media enables health workers to keep track of information, identify possible outbreaks, forecast disease trends, monitor emergency cases, and ascertain disease awareness and response to official health correspondence. This exchange of health information on social media has been regarded as an attempt to enhance public health surveillance (PHS). Despite its potential, the technology is still in its early stages and is not ready for widespread application. Advancements in pretrained language models (PLMs) have facilitated the development of several domain-specific PLMs and a variety of downstream applications. However, there are no PLMs for social media tasks involving PHS. We present and release PHS-BERT, a transformer-based PLM, to identify tasks related to public health surveillance on social media. We compared and benchmarked the performance of PHS-BERT on 25 datasets from different social medial platforms related to 7 different PHS tasks. Compared with existing PLMs that are mainly evaluated on limited tasks, PHS-BERT achieved state-of-the-art performance on all 25 tested datasets, showing that our PLM is robust and generalizable in the common PHS tasks. By making PHS-BERT available, we aim to facilitate the community to reduce the computational cost and introduce new baselines for future works across various PHS-related tasks.", } ```
wmFrank/sample-factory-2-atari-pong
wmFrank
2022-07-28T23:34:27Z
4
0
sample-factory
[ "sample-factory", "tensorboard", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-07-28T23:04:49Z
--- library_name: sample-factory tags: - deep-reinforcement-learning - reinforcement-learning - sample-factory model-index: - name: APPO results: - metrics: - type: mean_reward value: 13.50 +/- 7.43 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: atari_pong type: atari_pong --- A(n) **APPO** model trained on the **atari_pong** environment. This model was trained using Sample Factory 2.0: https://github.com/alex-petrenko/sample-factory
jungjongho/wav2vec2-large-xlsr-korean-demo-colab
jungjongho
2022-07-28T22:43:35Z
3
0
transformers
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-27T17:26:07Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: wav2vec2-large-xlsr-korean-demo-colab results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-large-xlsr-korean-demo-colab This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4534 - Wer: 0.3272 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 8 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:-----:|:---------------:|:------:| | 17.4809 | 0.65 | 400 | 4.6145 | 1.0 | | 4.4863 | 1.29 | 800 | 4.3819 | 1.0 | | 4.2921 | 1.94 | 1200 | 4.1163 | 0.9970 | | 2.7971 | 2.59 | 1600 | 1.5376 | 0.8379 | | 1.5061 | 3.24 | 2000 | 1.0354 | 0.7299 | | 1.1123 | 3.88 | 2400 | 0.7909 | 0.6418 | | 0.9037 | 4.53 | 2800 | 0.6345 | 0.5698 | | 0.779 | 5.18 | 3200 | 0.5909 | 0.5571 | | 0.6834 | 5.83 | 3600 | 0.5339 | 0.5063 | | 0.6287 | 6.47 | 4000 | 0.5326 | 0.4954 | | 0.5518 | 7.12 | 4400 | 0.4930 | 0.4607 | | 0.5315 | 7.77 | 4800 | 0.4577 | 0.4451 | | 0.4867 | 8.41 | 5200 | 0.4547 | 0.4382 | | 0.4543 | 9.06 | 5600 | 0.4581 | 0.4371 | | 0.4089 | 9.71 | 6000 | 0.4387 | 0.4258 | | 0.3893 | 10.36 | 6400 | 0.4300 | 0.4100 | | 0.3751 | 11.0 | 6800 | 0.4265 | 0.4137 | | 0.3333 | 11.65 | 7200 | 0.4294 | 0.4011 | | 0.3039 | 12.3 | 7600 | 0.4187 | 0.3912 | | 0.2974 | 12.94 | 8000 | 0.4079 | 0.3805 | | 0.2658 | 13.59 | 8400 | 0.4273 | 0.3864 | | 0.2676 | 14.24 | 8800 | 0.4103 | 0.3734 | | 0.2466 | 14.89 | 9200 | 0.4122 | 0.3701 | | 0.2282 | 15.53 | 9600 | 0.4176 | 0.3650 | | 0.2186 | 16.18 | 10000 | 0.4199 | 0.3632 | | 0.2132 | 16.83 | 10400 | 0.4159 | 0.3671 | | 0.1962 | 17.48 | 10800 | 0.4321 | 0.3641 | | 0.1922 | 18.12 | 11200 | 0.4300 | 0.3535 | | 0.1827 | 18.77 | 11600 | 0.4244 | 0.3596 | | 0.1709 | 19.42 | 12000 | 0.4191 | 0.3518 | | 0.157 | 20.06 | 12400 | 0.4308 | 0.3496 | | 0.147 | 20.71 | 12800 | 0.4360 | 0.3457 | | 0.1502 | 21.36 | 13200 | 0.4329 | 0.3431 | | 0.1448 | 22.01 | 13600 | 0.4334 | 0.3432 | | 0.1407 | 22.65 | 14000 | 0.4392 | 0.3440 | | 0.1342 | 23.3 | 14400 | 0.4418 | 0.3399 | | 0.1325 | 23.95 | 14800 | 0.4360 | 0.3383 | | 0.1183 | 24.6 | 15200 | 0.4521 | 0.3359 | | 0.1174 | 25.24 | 15600 | 0.4426 | 0.3322 | | 0.1137 | 25.89 | 16000 | 0.4438 | 0.3356 | | 0.1129 | 26.54 | 16400 | 0.4547 | 0.3347 | | 0.1077 | 27.18 | 16800 | 0.4482 | 0.3300 | | 0.0999 | 27.83 | 17200 | 0.4491 | 0.3281 | | 0.0978 | 28.48 | 17600 | 0.4533 | 0.3281 | | 0.0997 | 29.13 | 18000 | 0.4542 | 0.3283 | | 0.0908 | 29.77 | 18400 | 0.4534 | 0.3272 | ### Framework versions - Transformers 4.17.0 - Pytorch 1.12.0+cu113 - Datasets 1.18.3 - Tokenizers 0.12.1
pinot/wav2vec2-large-xls-r-300m-j-roman-colab
pinot
2022-07-28T22:28:51Z
5
0
transformers
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "dataset:common_voice", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-28T11:22:22Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - common_voice model-index: - name: wav2vec2-large-xls-r-300m-j-roman-colab results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-large-xls-r-300m-j-roman-colab This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset. It achieves the following results on the evaluation set: - Loss: 1.2233 - Wer: 0.1437 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 100 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:-----:|:---------------:|:------:| | 4.3479 | 1.22 | 400 | 1.6349 | 0.3868 | | 0.8621 | 2.45 | 800 | 1.0185 | 0.2346 | | 0.5421 | 3.67 | 1200 | 0.7549 | 0.1868 | | 0.3932 | 4.89 | 1600 | 0.7893 | 0.1811 | | 0.332 | 6.12 | 2000 | 0.9318 | 0.1919 | | 0.2902 | 7.34 | 2400 | 0.8263 | 0.1839 | | 0.2542 | 8.56 | 2800 | 0.8491 | 0.1829 | | 0.2355 | 9.79 | 3200 | 0.8820 | 0.1805 | | 0.2206 | 11.01 | 3600 | 0.9183 | 0.1748 | | 0.2041 | 12.23 | 4000 | 0.9131 | 0.1725 | | 0.1878 | 13.46 | 4400 | 0.9075 | 0.1699 | | 0.1733 | 14.68 | 4800 | 0.8456 | 0.1665 | | 0.1746 | 15.9 | 5200 | 0.9353 | 0.1745 | | 0.1671 | 17.13 | 5600 | 0.9318 | 0.1713 | | 0.1641 | 18.35 | 6000 | 0.8804 | 0.1661 | | 0.1578 | 19.57 | 6400 | 0.9849 | 0.1795 | | 0.1534 | 20.8 | 6800 | 1.0036 | 0.1637 | | 0.1484 | 22.02 | 7200 | 0.9618 | 0.1722 | | 0.1431 | 23.24 | 7600 | 0.9947 | 0.1680 | | 0.139 | 24.46 | 8000 | 0.9923 | 0.1729 | | 0.134 | 25.69 | 8400 | 1.0015 | 0.1641 | | 0.1298 | 26.91 | 8800 | 0.9930 | 0.1704 | | 0.1253 | 28.13 | 9200 | 0.9977 | 0.1605 | | 0.1178 | 29.36 | 9600 | 0.9756 | 0.1653 | | 0.1178 | 30.58 | 10000 | 1.1122 | 0.1784 | | 0.1165 | 31.8 | 10400 | 0.9883 | 0.1655 | | 0.1073 | 33.03 | 10800 | 1.1286 | 0.1677 | | 0.1121 | 34.25 | 11200 | 1.0406 | 0.1660 | | 0.1081 | 35.47 | 11600 | 1.0976 | 0.1678 | | 0.109 | 36.7 | 12000 | 1.0915 | 0.1722 | | 0.1027 | 37.92 | 12400 | 1.1167 | 0.1712 | | 0.0925 | 39.14 | 12800 | 1.1598 | 0.1693 | | 0.0913 | 40.37 | 13200 | 1.0712 | 0.1640 | | 0.0895 | 41.59 | 13600 | 1.1692 | 0.1745 | | 0.0908 | 42.81 | 14000 | 1.1248 | 0.1641 | | 0.0905 | 44.04 | 14400 | 1.0523 | 0.1678 | | 0.0864 | 45.26 | 14800 | 1.0261 | 0.1626 | | 0.0843 | 46.48 | 15200 | 1.0746 | 0.1676 | | 0.0759 | 47.71 | 15600 | 1.1035 | 0.1596 | | 0.0758 | 48.93 | 16000 | 1.0977 | 0.1622 | | 0.0743 | 50.15 | 16400 | 1.1203 | 0.1677 | | 0.0826 | 51.38 | 16800 | 1.0983 | 0.1651 | | 0.0743 | 52.6 | 17200 | 1.1452 | 0.1622 | | 0.0713 | 53.82 | 17600 | 1.0882 | 0.1623 | | 0.0651 | 55.05 | 18000 | 1.0588 | 0.1608 | | 0.0669 | 56.27 | 18400 | 1.1332 | 0.1600 | | 0.0626 | 57.49 | 18800 | 1.0747 | 0.1562 | | 0.0646 | 58.72 | 19200 | 1.0585 | 0.1599 | | 0.0639 | 59.94 | 19600 | 1.0106 | 0.1543 | | 0.0603 | 61.16 | 20000 | 1.0875 | 0.1585 | | 0.0551 | 62.39 | 20400 | 1.1273 | 0.1537 | | 0.0553 | 63.61 | 20800 | 1.1376 | 0.1577 | | 0.052 | 64.83 | 21200 | 1.1429 | 0.1553 | | 0.0506 | 66.06 | 21600 | 1.0872 | 0.1577 | | 0.0495 | 67.28 | 22000 | 1.0954 | 0.1488 | | 0.0483 | 68.5 | 22400 | 1.1397 | 0.1524 | | 0.0421 | 69.72 | 22800 | 1.2144 | 0.1581 | | 0.0457 | 70.95 | 23200 | 1.1581 | 0.1532 | | 0.0405 | 72.17 | 23600 | 1.2150 | 0.1566 | | 0.0409 | 73.39 | 24000 | 1.1176 | 0.1508 | | 0.0386 | 74.62 | 24400 | 1.2018 | 0.1526 | | 0.0374 | 75.84 | 24800 | 1.2548 | 0.1494 | | 0.0376 | 77.06 | 25200 | 1.2161 | 0.1486 | | 0.033 | 78.29 | 25600 | 1.1607 | 0.1558 | | 0.0339 | 79.51 | 26000 | 1.1557 | 0.1498 | | 0.0355 | 80.73 | 26400 | 1.1234 | 0.1490 | | 0.031 | 81.96 | 26800 | 1.1778 | 0.1473 | | 0.0301 | 83.18 | 27200 | 1.1594 | 0.1441 | | 0.0292 | 84.4 | 27600 | 1.2036 | 0.1482 | | 0.0256 | 85.63 | 28000 | 1.2334 | 0.1463 | | 0.0259 | 86.85 | 28400 | 1.2072 | 0.1469 | | 0.0271 | 88.07 | 28800 | 1.1843 | 0.1456 | | 0.0241 | 89.3 | 29200 | 1.1712 | 0.1445 | | 0.0223 | 90.52 | 29600 | 1.2059 | 0.1433 | | 0.0213 | 91.74 | 30000 | 1.2231 | 0.1452 | | 0.0212 | 92.97 | 30400 | 1.1980 | 0.1438 | | 0.0223 | 94.19 | 30800 | 1.2148 | 0.1459 | | 0.0185 | 95.41 | 31200 | 1.2190 | 0.1437 | | 0.0202 | 96.64 | 31600 | 1.2051 | 0.1437 | | 0.0188 | 97.86 | 32000 | 1.2154 | 0.1438 | | 0.0183 | 99.08 | 32400 | 1.2233 | 0.1437 | ### Framework versions - Transformers 4.17.0 - Pytorch 1.10.0+cu113 - Datasets 1.18.3 - Tokenizers 0.12.1
domenicrosati/deberta-v3-large-finetuned-synthetic-paraphrase-only
domenicrosati
2022-07-28T21:38:33Z
4
0
transformers
[ "transformers", "pytorch", "tensorboard", "deberta-v2", "text-classification", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2022-07-27T13:31:37Z
--- license: mit tags: - text-classification - generated_from_trainer metrics: - f1 - precision - recall model-index: - name: deberta-v3-large-finetuned-synthetic-paraphrase-only results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deberta-v3-large-finetuned-synthetic-paraphrase-only This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0120 - F1: 0.9768 - Precision: 0.9961 - Recall: 0.9583 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 6e-06 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 50 - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | Precision | Recall | |:-------------:|:-----:|:-----:|:---------------:|:------:|:---------:|:------:| | 0.0086 | 1.0 | 10205 | 0.0114 | 0.9642 | 0.9846 | 0.9446 | | 0.0059 | 2.0 | 20410 | 0.0143 | 0.9658 | 0.9961 | 0.9373 | | 0.0 | 3.0 | 30615 | 0.0141 | 0.9716 | 0.9961 | 0.9483 | ### Framework versions - Transformers 4.20.1 - Pytorch 1.11.0 - Datasets 2.1.0 - Tokenizers 0.12.1
Evelyn18/roberta-base-spanish-squades-becasIncentivos6
Evelyn18
2022-07-28T21:38:04Z
4
0
transformers
[ "transformers", "pytorch", "tensorboard", "roberta", "question-answering", "generated_from_trainer", "dataset:becasv2", "endpoints_compatible", "region:us" ]
question-answering
2022-07-28T21:08:34Z
--- tags: - generated_from_trainer datasets: - becasv2 model-index: - name: roberta-base-spanish-squades-becasIncentivos6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-spanish-squades-becasIncentivos6 This model is a fine-tuned version of [IIC/roberta-base-spanish-squades](https://huggingface.co/IIC/roberta-base-spanish-squades) on the becasv2 dataset. It achieves the following results on the evaluation set: - Loss: 2.0023 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 6 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 3 | 2.2257 | | No log | 2.0 | 6 | 1.8301 | | No log | 3.0 | 9 | 1.7627 | | No log | 4.0 | 12 | 1.8773 | | No log | 5.0 | 15 | 1.9731 | | No log | 6.0 | 18 | 2.0023 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
carblacac/xlm-roberta-base-finetuned-panx-de
carblacac
2022-07-28T18:47:01Z
7
0
transformers
[ "transformers", "pytorch", "tensorboard", "xlm-roberta", "token-classification", "generated_from_trainer", "dataset:xtreme", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-07-28T18:02:50Z
--- license: mit tags: - generated_from_trainer datasets: - xtreme model-index: - name: xlm-roberta-base-finetuned-panx-de results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-panx-de This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the xtreme dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 24 - eval_batch_size: 24 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
jperezv/distilbert-base-uncased-finetuned-imdb
jperezv
2022-07-28T17:14:39Z
4
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "fill-mask", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T15:00:34Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: distilbert-base-uncased-finetuned-imdb results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-imdb This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 2.4721 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.7086 | 1.0 | 157 | 2.4898 | | 2.5796 | 2.0 | 314 | 2.4230 | | 2.5269 | 3.0 | 471 | 2.4354 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Tokenizers 0.12.1
Billwzl/20split_dataset_version3
Billwzl
2022-07-28T16:20:35Z
6
0
transformers
[ "transformers", "pytorch", "distilbert", "fill-mask", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-27T11:21:44Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: 20split_dataset_version3 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # 20split_dataset_version3 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.8310 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 16 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 3.1679 | 1.0 | 313 | 2.9768 | | 2.9869 | 2.0 | 626 | 2.9299 | | 2.8528 | 3.0 | 939 | 2.9176 | | 2.7435 | 4.0 | 1252 | 2.9104 | | 2.6458 | 5.0 | 1565 | 2.8863 | | 2.5865 | 6.0 | 1878 | 2.8669 | | 2.5218 | 7.0 | 2191 | 2.8802 | | 2.4647 | 8.0 | 2504 | 2.8639 | | 2.3933 | 9.0 | 2817 | 2.8543 | | 2.3687 | 10.0 | 3130 | 2.8573 | | 2.3221 | 11.0 | 3443 | 2.8398 | | 2.276 | 12.0 | 3756 | 2.8415 | | 2.2379 | 13.0 | 4069 | 2.8471 | | 2.2427 | 14.0 | 4382 | 2.8318 | | 2.1741 | 15.0 | 4695 | 2.8356 | | 2.1652 | 16.0 | 5008 | 2.8310 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
jirtan/ddpm-butterflies-128
jirtan
2022-07-28T15:30:13Z
5
0
diffusers
[ "diffusers", "tensorboard", "en", "dataset:huggan/smithsonian_butterflies_subset", "license:apache-2.0", "diffusers:DDPMPipeline", "region:us" ]
null
2022-07-28T14:42:17Z
--- language: en license: apache-2.0 library_name: diffusers tags: [] datasets: huggan/smithsonian_butterflies_subset metrics: [] --- <!-- This model card has been generated automatically according to the information the training script had access to. You should probably proofread and complete it, then remove this comment. --> # ddpm-butterflies-128 ## Model description This diffusion model is trained with the [🤗 Diffusers](https://github.com/huggingface/diffusers) library on the `huggan/smithsonian_butterflies_subset` dataset. ## Intended uses & limitations #### How to use ```python # TODO: add an example code snippet for running this diffusion pipeline ``` #### Limitations and bias [TODO: provide examples of latent issues and potential remediations] ## Training data [TODO: describe the data used to train the model] ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 16 - eval_batch_size: 16 - gradient_accumulation_steps: 1 - optimizer: AdamW with betas=(None, None), weight_decay=None and epsilon=None - lr_scheduler: None - lr_warmup_steps: 500 - ema_inv_gamma: None - ema_inv_gamma: None - ema_inv_gamma: None - mixed_precision: fp16 ### Training results 📈 [TensorBoard logs](https://huggingface.co/jirtan/ddpm-butterflies-128/tensorboard?#scalars)
Atharvgarg/bert-small2bert-small-finetuned-cnn_daily_mail-summarization-finetuned-bbc-news
Atharvgarg
2022-07-28T15:22:19Z
4
0
transformers
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "summarisation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2022-07-28T14:37:18Z
--- license: apache-2.0 tags: - summarisation - generated_from_trainer metrics: - rouge model-index: - name: bert-small2bert-small-finetuned-cnn_daily_mail-summarization-finetuned-bbc-news results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-small2bert-small-finetuned-cnn_daily_mail-summarization-finetuned-bbc-news This model is a fine-tuned version of [mrm8488/bert-small2bert-small-finetuned-cnn_daily_mail-summarization](https://huggingface.co/mrm8488/bert-small2bert-small-finetuned-cnn_daily_mail-summarization) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.6835 - Rouge1: 58.9345 - Rouge2: 47.1037 - Rougel: 40.9839 - Rougelsum: 57.6981 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5.6e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 8 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:| | 0.8246 | 1.0 | 223 | 0.7050 | 55.7882 | 42.9793 | 38.4511 | 54.3125 | | 0.6414 | 2.0 | 446 | 0.6834 | 55.149 | 42.664 | 38.3864 | 53.7712 | | 0.5603 | 3.0 | 669 | 0.6815 | 56.9756 | 44.8057 | 39.1377 | 55.5815 | | 0.5079 | 4.0 | 892 | 0.6749 | 57.7397 | 45.6267 | 40.0509 | 56.3886 | | 0.4622 | 5.0 | 1115 | 0.6781 | 58.07 | 45.9102 | 40.2704 | 56.7008 | | 0.4263 | 6.0 | 1338 | 0.6798 | 58.1215 | 45.976 | 40.256 | 56.8203 | | 0.399 | 7.0 | 1561 | 0.6798 | 58.5486 | 46.6901 | 40.8045 | 57.2947 | | 0.3815 | 8.0 | 1784 | 0.6835 | 58.9345 | 47.1037 | 40.9839 | 57.6981 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
Dugerij/Reinforce-pixelcopter
Dugerij
2022-07-28T14:45:45Z
0
0
null
[ "Pixelcopter-PLE-v0", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2022-07-28T14:45:39Z
--- tags: - Pixelcopter-PLE-v0 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-pixelcopter results: - metrics: - type: mean_reward value: 17.00 +/- 12.95 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: Pixelcopter-PLE-v0 type: Pixelcopter-PLE-v0 --- # **Reinforce** Agent playing **Pixelcopter-PLE-v0** This is a trained model of a **Reinforce** agent playing **Pixelcopter-PLE-v0** . To learn to use this model and train yours check Unit 5 of the Deep Reinforcement Learning Class: https://github.com/huggingface/deep-rl-class/tree/main/unit5
KBLab/albert-base-swedish-cased-alpha
KBLab
2022-07-28T14:08:17Z
11
2
transformers
[ "transformers", "pytorch", "albert", "sv", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:04Z
--- language: sv --- # Swedish BERT Models The National Library of Sweden / KBLab releases three pretrained language models based on BERT and ALBERT. The models are trained on approximately 15-20GB of text (200M sentences, 3000M tokens) from various sources (books, news, government publications, swedish wikipedia and internet forums) aiming to provide a representative BERT model for Swedish text. A more complete description will be published later on. The following three models are currently available: - **bert-base-swedish-cased** (*v1*) - A BERT trained with the same hyperparameters as first published by Google. - **bert-base-swedish-cased-ner** (*experimental*) - a BERT fine-tuned for NER using SUC 3.0. - **albert-base-swedish-cased-alpha** (*alpha*) - A first attempt at an ALBERT for Swedish. All models are cased and trained with whole word masking. ## Files | **name** | **files** | |---------------------------------|-----------| | bert-base-swedish-cased | [config](https://s3.amazonaws.com/models.huggingface.co/bert/KB/bert-base-swedish-cased/config.json), [vocab](https://s3.amazonaws.com/models.huggingface.co/bert/KB/bert-base-swedish-cased/vocab.txt), [pytorch_model.bin](https://s3.amazonaws.com/models.huggingface.co/bert/KB/bert-base-swedish-cased/pytorch_model.bin) | | bert-base-swedish-cased-ner | [config](https://s3.amazonaws.com/models.huggingface.co/bert/KB/bert-base-swedish-cased-ner/config.json), [vocab](https://s3.amazonaws.com/models.huggingface.co/bert/KB/bert-base-swedish-cased-ner/vocab.txt) [pytorch_model.bin](https://s3.amazonaws.com/models.huggingface.co/bert/KB/bert-base-swedish-cased-ner/pytorch_model.bin) | | albert-base-swedish-cased-alpha | [config](https://s3.amazonaws.com/models.huggingface.co/bert/KB/albert-base-swedish-cased-alpha/config.json), [sentencepiece model](https://s3.amazonaws.com/models.huggingface.co/bert/KB/albert-base-swedish-cased-alpha/spiece.model), [pytorch_model.bin](https://s3.amazonaws.com/models.huggingface.co/bert/KB/albert-base-swedish-cased-alpha/pytorch_model.bin) | TensorFlow model weights will be released soon. ## Usage requirements / installation instructions The examples below require Huggingface Transformers 2.4.1 and Pytorch 1.3.1 or greater. For Transformers<2.4.0 the tokenizer must be instantiated manually and the `do_lower_case` flag parameter set to `False` and `keep_accents` to `True` (for ALBERT). To create an environment where the examples can be run, run the following in an terminal on your OS of choice. ``` # git clone https://github.com/Kungbib/swedish-bert-models # cd swedish-bert-models # python3 -m venv venv # source venv/bin/activate # pip install --upgrade pip # pip install -r requirements.txt ``` ### BERT Base Swedish A standard BERT base for Swedish trained on a variety of sources. Vocabulary size is ~50k. Using Huggingface Transformers the model can be loaded in Python as follows: ```python from transformers import AutoModel,AutoTokenizer tok = AutoTokenizer.from_pretrained('KBLab/bert-base-swedish-cased') model = AutoModel.from_pretrained('KBLab/bert-base-swedish-cased') ``` ### BERT base fine-tuned for Swedish NER This model is fine-tuned on the SUC 3.0 dataset. Using the Huggingface pipeline the model can be easily instantiated. For Transformer<2.4.1 it seems the tokenizer must be loaded separately to disable lower-casing of input strings: ```python from transformers import pipeline nlp = pipeline('ner', model='KB/bert-base-swedish-cased-ner', tokenizer='KB/bert-base-swedish-cased-ner') nlp('Idag släpper KB tre språkmodeller.') ``` Running the Python code above should produce in something like the result below. Entity types used are `TME` for time, `PRS` for personal names, `LOC` for locations, `EVN` for events and `ORG` for organisations. These labels are subject to change. ```python [ { 'word': 'Idag', 'score': 0.9998126029968262, 'entity': 'TME' }, { 'word': 'KB', 'score': 0.9814832210540771, 'entity': 'ORG' } ] ``` The BERT tokenizer often splits words into multiple tokens, with the subparts starting with `##`, for example the string `Engelbert kör Volvo till Herrängens fotbollsklubb` gets tokenized as `Engel ##bert kör Volvo till Herr ##ängens fotbolls ##klubb`. To glue parts back together one can use something like this: ```python text = 'Engelbert tar Volvon till Tele2 Arena för att titta på Djurgården IF ' +\ 'som spelar fotboll i VM klockan två på kvällen.' l = [] for token in nlp(text): if token['word'].startswith('##'): l[-1]['word'] += token['word'][2:] else: l += [ token ] print(l) ``` Which should result in the following (though less cleanly formatted): ```python [ { 'word': 'Engelbert', 'score': 0.99..., 'entity': 'PRS'}, { 'word': 'Volvon', 'score': 0.99..., 'entity': 'OBJ'}, { 'word': 'Tele2', 'score': 0.99..., 'entity': 'LOC'}, { 'word': 'Arena', 'score': 0.99..., 'entity': 'LOC'}, { 'word': 'Djurgården', 'score': 0.99..., 'entity': 'ORG'}, { 'word': 'IF', 'score': 0.99..., 'entity': 'ORG'}, { 'word': 'VM', 'score': 0.99..., 'entity': 'EVN'}, { 'word': 'klockan', 'score': 0.99..., 'entity': 'TME'}, { 'word': 'två', 'score': 0.99..., 'entity': 'TME'}, { 'word': 'på', 'score': 0.99..., 'entity': 'TME'}, { 'word': 'kvällen', 'score': 0.54..., 'entity': 'TME'} ] ``` ### ALBERT base The easiest way to do this is, again, using Huggingface Transformers: ```python from transformers import AutoModel,AutoTokenizer tok = AutoTokenizer.from_pretrained('KBLab/albert-base-swedish-cased-alpha'), model = AutoModel.from_pretrained('KBLab/albert-base-swedish-cased-alpha') ``` ## Acknowledgements ❤️ - Resources from Stockholms University, Umeå University and Swedish Language Bank at Gothenburg University were used when fine-tuning BERT for NER. - Model pretraining was made partly in-house at the KBLab and partly (for material without active copyright) with the support of Cloud TPUs from Google's TensorFlow Research Cloud (TFRC). - Models are hosted on S3 by Huggingface 🤗
Nekoo/P0ken_picture
Nekoo
2022-07-28T13:33:38Z
0
0
null
[ "license:bigscience-bloom-rail-1.0", "region:us" ]
null
2022-07-28T13:33:38Z
--- license: bigscience-bloom-rail-1.0 ---
Perselope/thesis-audio-1
Perselope
2022-07-28T13:27:40Z
5
0
transformers
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2022-07-26T22:02:31Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: thesis-audio-1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # thesis-audio-1 This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4268 - Wer: 0.3395 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 30 ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 3.4633 | 4.0 | 500 | 1.4892 | 1.0006 | | 0.5377 | 8.0 | 1000 | 0.4046 | 0.4163 | | 0.1818 | 12.0 | 1500 | 0.4255 | 0.3850 | | 0.1024 | 16.0 | 2000 | 0.4574 | 0.3644 | | 0.0723 | 20.0 | 2500 | 0.4412 | 0.3550 | | 0.0542 | 24.0 | 3000 | 0.4095 | 0.3404 | | 0.0434 | 28.0 | 3500 | 0.4268 | 0.3395 | ### Framework versions - Transformers 4.17.0 - Pytorch 1.12.0+cu113 - Datasets 1.18.3 - Tokenizers 0.12.1
kabelomalapane/En-Zu_update
kabelomalapane
2022-07-28T13:24:27Z
9
0
transformers
[ "transformers", "pytorch", "tensorboard", "marian", "text2text-generation", "translation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-07-28T10:55:08Z
--- license: apache-2.0 tags: - translation - generated_from_trainer metrics: - bleu model-index: - name: En-Zu_update results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # En-Zu_update This model is a fine-tuned version of [kabelomalapane/test_model1.2_updated](https://huggingface.co/kabelomalapane/test_model1.2_updated) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.7101 - Bleu: 11.8551 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | |:-------------:|:-----:|:-----:|:---------------:|:-------:| | 1.9111 | 1.0 | 1173 | 1.7594 | 11.7012 | | 1.7191 | 2.0 | 2346 | 1.7279 | 12.0250 | | 1.5709 | 3.0 | 3519 | 1.7172 | 10.6222 | | 1.4924 | 4.0 | 4692 | 1.7042 | 11.4224 | | 1.4188 | 5.0 | 5865 | 1.7051 | 11.4330 | | 1.3566 | 6.0 | 7038 | 1.6972 | 11.5300 | | 1.3141 | 7.0 | 8211 | 1.7041 | 11.4339 | | 1.2641 | 8.0 | 9384 | 1.7064 | 11.4030 | | 1.2437 | 9.0 | 10557 | 1.7079 | 11.4014 | | 1.2333 | 10.0 | 11730 | 1.7101 | 11.5164 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
ICML2022/TimeIsMattEr
ICML2022
2022-07-28T12:00:17Z
0
3
null
[ "video-action-recognition", "dataset:HuggingFaceM4/something_something_v2", "license:cc-by-nc-4.0", "region:us" ]
null
2022-07-28T11:54:09Z
--- license: cc-by-nc-4.0 datasets: - HuggingFaceM4/something_something_v2 tags: - video-action-recognition metrics: - accuracy ---
amartyobanerjee/distilbert-base-uncased-whole-word-word-ids-finetuned-imdb
amartyobanerjee
2022-07-28T10:01:48Z
3
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "fill-mask", "generated_from_trainer", "dataset:imdb", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T09:53:39Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imdb model-index: - name: distilbert-base-uncased-whole-word-word-ids-finetuned-imdb results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-whole-word-word-ids-finetuned-imdb This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the imdb dataset. It achieves the following results on the evaluation set: - Loss: 0.6573 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.7261 | 1.0 | 157 | 0.6532 | | 0.6766 | 2.0 | 314 | 0.6514 | | 0.6677 | 3.0 | 471 | 0.6555 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
amartyobanerjee/distilbert-base-uncased-finetuned-imdb
amartyobanerjee
2022-07-28T09:45:35Z
3
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "fill-mask", "generated_from_trainer", "dataset:imdb", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-07-28T05:27:01Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imdb model-index: - name: distilbert-base-uncased-finetuned-imdb results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-imdb This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the imdb dataset. It achieves the following results on the evaluation set: - Loss: 2.4721 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.7086 | 1.0 | 157 | 2.4898 | | 2.5796 | 2.0 | 314 | 2.4230 | | 2.5269 | 3.0 | 471 | 2.4354 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
butchland/Optuna-ppo-LunarLander-v2
butchland
2022-07-28T09:41:11Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2022-07-28T09:34:51Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - metrics: - type: mean_reward value: 271.94 +/- 21.72 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
Ravindra001/bert-finetuned-ner
Ravindra001
2022-07-28T09:29:11Z
5
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "generated_from_trainer", "dataset:wikiann", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-05-25T06:09:39Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - wikiann metrics: - precision - recall - f1 - accuracy model-index: - name: bert-finetuned-ner results: - task: name: Token Classification type: token-classification dataset: name: wikiann type: wikiann args: en metrics: - name: Precision type: precision value: 0.819622641509434 - name: Recall type: recall value: 0.8444790046656299 - name: F1 type: f1 value: 0.8318651857525853 - name: Accuracy type: accuracy value: 0.9269227060339613 - task: type: token-classification name: Token Classification dataset: name: wikiann type: wikiann config: en split: test metrics: - name: Accuracy type: accuracy value: 0.8492771401033908 verified: true - name: Precision type: precision value: 0.857294905524994 verified: true - name: Recall type: recall value: 0.865900059186607 verified: true - name: F1 type: f1 value: 0.8615759964905745 verified: true - name: loss type: loss value: 1.054654836654663 verified: true --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-finetuned-ner This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the wikiann dataset. It achieves the following results on the evaluation set: - Loss: 0.3217 - Precision: 0.8196 - Recall: 0.8445 - F1: 0.8319 - Accuracy: 0.9269 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.2821 | 1.0 | 2500 | 0.2906 | 0.7983 | 0.8227 | 0.8103 | 0.9193 | | 0.2087 | 2.0 | 5000 | 0.2614 | 0.8030 | 0.8379 | 0.8201 | 0.9257 | | 0.1404 | 3.0 | 7500 | 0.3217 | 0.8196 | 0.8445 | 0.8319 | 0.9269 | ### Framework versions - Transformers 4.19.2 - Pytorch 1.11.0+cu113 - Datasets 2.2.2 - Tokenizers 0.12.1
AlbertShu/Reinforce-v0
AlbertShu
2022-07-28T09:22:30Z
0
0
null
[ "CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2022-07-28T09:22:20Z
--- tags: - CartPole-v1 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-v0 results: - metrics: - type: mean_reward value: 99.30 +/- 29.54 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 --- # **Reinforce** Agent playing **CartPole-v1** This is a trained model of a **Reinforce** agent playing **CartPole-v1** . To learn to use this model and train yours check Unit 5 of the Deep Reinforcement Learning Class: https://github.com/huggingface/deep-rl-class/tree/main/unit5
CompVis/ldm-celebahq-256
CompVis
2022-07-28T08:12:07Z
199
42
diffusers
[ "diffusers", "pytorch", "unconditional-image-generation", "arxiv:2112.10752", "license:apache-2.0", "diffusers:LDMPipeline", "region:us" ]
unconditional-image-generation
2022-07-15T17:28:35Z
--- license: apache-2.0 tags: - pytorch - diffusers - unconditional-image-generation --- # Latent Diffusion Models (LDM) **Paper**: [High-Resolution Image Synthesis with Latent Diffusion Models](https://arxiv.org/abs/2112.10752) **Abstract**: *By decomposing the image formation process into a sequential application of denoising autoencoders, diffusion models (DMs) achieve state-of-the-art synthesis results on image data and beyond. Additionally, their formulation allows for a guiding mechanism to control the image generation process without retraining. However, since these models typically operate directly in pixel space, optimization of powerful DMs often consumes hundreds of GPU days and inference is expensive due to sequential evaluations. To enable DM training on limited computational resources while retaining their quality and flexibility, we apply them in the latent space of powerful pretrained autoencoders. In contrast to previous work, training diffusion models on such a representation allows for the first time to reach a near-optimal point between complexity reduction and detail preservation, greatly boosting visual fidelity. By introducing cross-attention layers into the model architecture, we turn diffusion models into powerful and flexible generators for general conditioning inputs such as text or bounding boxes and high-resolution synthesis becomes possible in a convolutional manner. Our latent diffusion models (LDMs) achieve a new state of the art for image inpainting and highly competitive performance on various tasks, including unconditional image generation, semantic scene synthesis, and super-resolution, while significantly reducing computational requirements compared to pixel-based DMs.* **Authors** *Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, Björn Ommer* ## Usage ### Inference with a pipeline ```python !pip install diffusers from diffusers import DiffusionPipeline model_id = "CompVis/ldm-celebahq-256" # load model and scheduler pipeline = DiffusionPipeline.from_pretrained(model_id) # run pipeline in inference (sample random noise and denoise) image = pipeline(num_inference_steps=200)["sample"] # save image image[0].save("ldm_generated_image.png") ``` ### Inference with an unrolled loop ```python !pip install diffusers from diffusers import UNet2DModel, DDIMScheduler, VQModel import torch import PIL.Image import numpy as np import tqdm seed = 3 # load all models unet = UNet2DModel.from_pretrained("CompVis/ldm-celebahq-256", subfolder="unet") vqvae = VQModel.from_pretrained("CompVis/ldm-celebahq-256", subfolder="vqvae") scheduler = DDIMScheduler.from_config("CompVis/ldm-celebahq-256", subfolder="scheduler") # set to cuda torch_device = "cuda" if torch.cuda.is_available() else "cpu" unet.to(torch_device) vqvae.to(torch_device) # generate gaussian noise to be decoded generator = torch.manual_seed(seed) noise = torch.randn( (1, unet.in_channels, unet.sample_size, unet.sample_size), generator=generator, ).to(torch_device) # set inference steps for DDIM scheduler.set_timesteps(num_inference_steps=200) image = noise for t in tqdm.tqdm(scheduler.timesteps): # predict noise residual of previous image with torch.no_grad(): residual = unet(image, t)["sample"] # compute previous image x_t according to DDIM formula prev_image = scheduler.step(residual, t, image, eta=0.0)["prev_sample"] # x_t-1 -> x_t image = prev_image # decode image with vae with torch.no_grad(): image = vqvae.decode(image) # process image image_processed = image.cpu().permute(0, 2, 3, 1) image_processed = (image_processed + 1.0) * 127.5 image_processed = image_processed.clamp(0, 255).numpy().astype(np.uint8) image_pil = PIL.Image.fromarray(image_processed[0]) image_pil.save(f"generated_image_{seed}.png") ``` ## Samples 1. ![sample_0](https://huggingface.co/CompVis/latent-diffusion-celeba-256/resolve/main/images/generated_image_0.png) 2. ![sample_1](https://huggingface.co/CompVis/latent-diffusion-celeba-256/resolve/main/images/generated_image_1.png) 3. ![sample_2](https://huggingface.co/CompVis/latent-diffusion-celeba-256/resolve/main/images/generated_image_2.png) 4. ![sample_3](https://huggingface.co/CompVis/latent-diffusion-celeba-256/resolve/main/images/generated_image_3.png)
pkufool/icefall-asr-librispeech-pruned-stateless-streaming-conformer-rnnt4-2022-06-10
pkufool
2022-07-28T08:00:20Z
0
1
null
[ "tensorboard", "license:apache-2.0", "region:us" ]
null
2022-06-09T22:50:20Z
--- license: apache-2.0 --- The pretrained model (pruned_transducer_stateless4) in https://github.com/k2-fsa/icefall/pull/380 ### training ``` #!/usr/bin/env bash set -x K2_ROOT=/path/to/k2 ICEFALL=/path/to/icefall export PYTHONPATH=$K2_ROOT/k2/python:$PYTHONPATH export PYTHONPATH=$K2_ROOT/build/lib:$PYTHONPATH export PYTHONPATH=$ICEFALL:$PYTHONPATH export CUDA_VISIBLE_DEVICES="0,1,2,3" ./pruned_transducer_stateless4/train.py \ --exp-dir pruned_transducer_stateless4/exp \ --full-libri 1 \ --dynamic-chunk-training 1 \ --short-chunk-size 32 \ --num-left-chunks 4 \ --causal-convolution 1 \ --max-duration 300 \ --world-size 4 \ --start-epoch 1 \ --num-epochs 30 ``` ### decoding #### simulate streaming ``` #!/usr/bin/env bash set -x K2_ROOT=/path/to/k2 ICEFALL=/path/to/icefall export PYTHONPATH=$K2_ROOT/k2/python:$PYTHONPATH export PYTHONPATH=$K2_ROOT/build/lib:$PYTHONPATH export PYTHONPATH=$ICEFALL:$PYTHONPATH export CUDA_VISIBLE_DEVICES="0" for size in 1 2 4 8 16 32; do for left in 32 64 -1; do ./pruned_transducer_stateless4/decode.py \ --simulate-streaming 1 \ --decode-chunk-size ${size} \ --left-context ${left} \ --causal-convolution 1 \ --use-averaged-model 1 \ --epoch 29 \ --avg 6 \ --exp-dir ./pruned_transducer_stateless4/exp \ --max-sym-per-frame 1 \ --max-duration 1000 \ --decoding-method greedy_search done done ``` #### streaming ``` #!/usr/bin/env bash set -x K2_ROOT=/path/to/k2 ICEFALL=/path/to/icefall export PYTHONPATH=$K2_ROOT/k2/python:$PYTHONPATH export PYTHONPATH=$K2_ROOT/build/lib:$PYTHONPATH export PYTHONPATH=$ICEFALL:$PYTHONPATH export CUDA_VISIBLE_DEVICES="0" #left_context=32 #chunk_size=8 left_context=64 chunk_size=16 for right in 0 2 4 8; do ./pruned_transducer_stateless4/streaming_decode.py \ --left-context ${left_context} \ --decode-chunk-size ${chunk_size} \ --right-context ${right} \ --exp-dir ./pruned_transducer_stateless4/exp \ --use-averaged-model 1 \ --epoch 29 \ --avg 6 \ --num-decode-streams 1000 done ``` ### export for pretrained.pt ``` python pruned_transducer_stateless4/export.py \ --exp-dir ./pruned_transducer_stateless4/exp \ --epoch 29 \ --avg 6 \ --streaming-model 1 \ --causal-convolution 1 ``` for cpu_jit.pt ``` python pruned_transducer_stateless4/export.py \ --exp-dir ./pruned_transducer_stateless4/exp \ --epoch 29 \ --avg 6 \ --streaming-model 1 \ --causal-convolution 1 \ --jit 1 ```
seeksery/DialoGPT-calig3
seeksery
2022-07-28T03:16:28Z
9
0
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2022-07-28T00:28:21Z
--- tags: - conversational ---
Jmolano/bert-finetuned-ner
Jmolano
2022-07-28T02:51:07Z
4
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "generated_from_trainer", "dataset:conll2003", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-07-26T21:56:29Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - conll2003 metrics: - precision - recall - f1 - accuracy model-index: - name: bert-finetuned-ner results: - task: name: Token Classification type: token-classification dataset: name: conll2003 type: conll2003 config: conll2003 split: train args: conll2003 metrics: - name: Precision type: precision value: 0.9327383903487027 - name: Recall type: recall value: 0.9498485358465163 - name: F1 type: f1 value: 0.9412157091636788 - name: Accuracy type: accuracy value: 0.9860923058809677 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-finetuned-ner This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the conll2003 dataset. It achieves the following results on the evaluation set: - Loss: 0.0617 - Precision: 0.9327 - Recall: 0.9498 - F1: 0.9412 - Accuracy: 0.9861 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0868 | 1.0 | 1756 | 0.0697 | 0.9204 | 0.9297 | 0.9250 | 0.9807 | | 0.0342 | 2.0 | 3512 | 0.0647 | 0.9273 | 0.9465 | 0.9368 | 0.9853 | | 0.0175 | 3.0 | 5268 | 0.0617 | 0.9327 | 0.9498 | 0.9412 | 0.9861 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
AykeeSalazar/vc-bantai-vit-withoutAMBI-adunest-v1
AykeeSalazar
2022-07-28T02:45:09Z
54
0
transformers
[ "transformers", "pytorch", "tensorboard", "vit", "image-classification", "generated_from_trainer", "dataset:imagefolder", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2022-07-28T01:15:01Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imagefolder metrics: - accuracy model-index: - name: vc-bantai-vit-withoutAMBI-adunest-v1 results: - task: name: Image Classification type: image-classification dataset: name: imagefolder type: imagefolder args: Violation-Classification---Raw-6 metrics: - name: Accuracy type: accuracy value: 0.9181222707423581 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # vc-bantai-vit-withoutAMBI-adunest-v1 This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset. It achieves the following results on the evaluation set: - Loss: 0.3318 - Accuracy: 0.9181 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 200 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 0.23 | 100 | 0.3365 | 0.8581 | | No log | 0.45 | 200 | 0.3552 | 0.8472 | | No log | 0.68 | 300 | 0.3165 | 0.8581 | | No log | 0.91 | 400 | 0.2882 | 0.8690 | | 0.3813 | 1.13 | 500 | 0.2825 | 0.8745 | | 0.3813 | 1.36 | 600 | 0.2686 | 0.9007 | | 0.3813 | 1.59 | 700 | 0.2381 | 0.9017 | | 0.3813 | 1.81 | 800 | 0.3643 | 0.8734 | | 0.3813 | 2.04 | 900 | 0.2873 | 0.8930 | | 0.2736 | 2.27 | 1000 | 0.2236 | 0.9039 | | 0.2736 | 2.49 | 1100 | 0.2652 | 0.8723 | | 0.2736 | 2.72 | 1200 | 0.2793 | 0.8952 | | 0.2736 | 2.95 | 1300 | 0.2158 | 0.8974 | | 0.2736 | 3.17 | 1400 | 0.2410 | 0.8886 | | 0.2093 | 3.4 | 1500 | 0.2262 | 0.9017 | | 0.2093 | 3.63 | 1600 | 0.2110 | 0.9214 | | 0.2093 | 3.85 | 1700 | 0.2048 | 0.9138 | | 0.2093 | 4.08 | 1800 | 0.2044 | 0.9127 | | 0.2093 | 4.31 | 1900 | 0.2591 | 0.9007 | | 0.1764 | 4.54 | 2000 | 0.2466 | 0.8952 | | 0.1764 | 4.76 | 2100 | 0.2554 | 0.9017 | | 0.1764 | 4.99 | 2200 | 0.2145 | 0.9203 | | 0.1764 | 5.22 | 2300 | 0.3187 | 0.9039 | | 0.1764 | 5.44 | 2400 | 0.3336 | 0.9050 | | 0.1454 | 5.67 | 2500 | 0.2542 | 0.9127 | | 0.1454 | 5.9 | 2600 | 0.2796 | 0.8952 | | 0.1454 | 6.12 | 2700 | 0.2410 | 0.9181 | | 0.1454 | 6.35 | 2800 | 0.2503 | 0.9148 | | 0.1454 | 6.58 | 2900 | 0.2966 | 0.8996 | | 0.1216 | 6.8 | 3000 | 0.1978 | 0.9312 | | 0.1216 | 7.03 | 3100 | 0.2297 | 0.9214 | | 0.1216 | 7.26 | 3200 | 0.2768 | 0.9203 | | 0.1216 | 7.48 | 3300 | 0.3356 | 0.9083 | | 0.1216 | 7.71 | 3400 | 0.3415 | 0.9138 | | 0.1038 | 7.94 | 3500 | 0.2398 | 0.9061 | | 0.1038 | 8.16 | 3600 | 0.3347 | 0.8963 | | 0.1038 | 8.39 | 3700 | 0.2199 | 0.9203 | | 0.1038 | 8.62 | 3800 | 0.2943 | 0.9061 | | 0.1038 | 8.84 | 3900 | 0.2561 | 0.9181 | | 0.0925 | 9.07 | 4000 | 0.4170 | 0.8777 | | 0.0925 | 9.3 | 4100 | 0.3638 | 0.8974 | | 0.0925 | 9.52 | 4200 | 0.3233 | 0.9094 | | 0.0925 | 9.75 | 4300 | 0.3496 | 0.9203 | | 0.0925 | 9.98 | 4400 | 0.3621 | 0.8996 | | 0.0788 | 10.2 | 4500 | 0.3260 | 0.9116 | | 0.0788 | 10.43 | 4600 | 0.3979 | 0.9061 | | 0.0788 | 10.66 | 4700 | 0.3301 | 0.8974 | | 0.0788 | 10.88 | 4800 | 0.2197 | 0.9105 | | 0.0788 | 11.11 | 4900 | 0.3306 | 0.9148 | | 0.0708 | 11.34 | 5000 | 0.3318 | 0.9181 | ### Framework versions - Transformers 4.20.1 - Pytorch 1.12.0+cu113 - Datasets 2.3.2 - Tokenizers 0.12.1
jianzhnie/q-Taxi-v3
jianzhnie
2022-07-28T02:20:13Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2022-07-28T01:57:36Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-Taxi-v3 results: - metrics: - type: mean_reward value: 7.56 +/- 2.71 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 --- # **Q-Learning** Agent playing **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="jianzhnie/q-Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) evaluate_agent(env, model["max_steps"], model["n_eval_episodes"], model["qtable"], model["eval_seed"]) ```
huggingtweets/penguinnnno
huggingtweets
2022-07-28T01:35:06Z
3
0
transformers
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2022-07-28T01:07:43Z
--- language: en thumbnail: http://www.huggingtweets.com/penguinnnno/1658971968390/predictions.png tags: - huggingtweets widget: - text: "My dream is" --- <div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1452082178741968901/oERkhKFL_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">penguino</div> <div style="text-align: center; font-size: 14px;">@penguinnnno</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from penguino. | Data | penguino | | --- | --- | | Tweets downloaded | 1865 | | Retweets | 839 | | Short tweets | 377 | | Tweets kept | 649 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2hb9ovan/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @penguinnnno's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/4k058458) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/4k058458/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/penguinnnno') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
kabelomalapane/Af-En_update
kabelomalapane
2022-07-27T23:37:19Z
117
0
transformers
[ "transformers", "pytorch", "tensorboard", "marian", "text2text-generation", "translation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-07-27T20:53:09Z
--- license: apache-2.0 tags: - translation - generated_from_trainer metrics: - bleu model-index: - name: Af-En_update results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Af-En_update This model is a fine-tuned version of [Helsinki-NLP/opus-mt-af-en](https://huggingface.co/Helsinki-NLP/opus-mt-af-en) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.7197 - Bleu: 55.3346 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 15 ### Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | |:-------------:|:-----:|:-----:|:---------------:|:-------:| | 1.3745 | 1.0 | 2553 | 1.7537 | 51.9270 | | 1.0462 | 2.0 | 5106 | 1.6305 | 53.9359 | | 0.896 | 3.0 | 7659 | 1.6216 | 54.3049 | | 0.7824 | 4.0 | 10212 | 1.6108 | 54.9902 | | 0.6974 | 5.0 | 12765 | 1.6183 | 55.0265 | | 0.643 | 6.0 | 15318 | 1.6207 | 55.4137 | | 0.5635 | 7.0 | 17871 | 1.6276 | 55.1335 | | 0.5141 | 8.0 | 20424 | 1.6498 | 55.2215 | | 0.4681 | 9.0 | 22977 | 1.6678 | 55.2000 | | 0.4304 | 10.0 | 25530 | 1.6797 | 55.2748 | | 0.425 | 11.0 | 28083 | 1.7004 | 55.0478 | | 0.398 | 12.0 | 30636 | 1.7013 | 55.3591 | | 0.3759 | 13.0 | 33189 | 1.7082 | 55.3225 | | 0.3681 | 14.0 | 35742 | 1.7151 | 55.1793 | | 0.3571 | 15.0 | 38295 | 1.7197 | 55.2729 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1
mariastull/Reinforce-3
mariastull
2022-07-27T21:39:59Z
0
0
null
[ "Pong-PLE-v0", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2022-07-27T21:39:47Z
--- tags: - Pong-PLE-v0 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-3 results: - metrics: - type: mean_reward value: -16.00 +/- 0.00 name: mean_reward task: type: reinforcement-learning name: reinforcement-learning dataset: name: Pong-PLE-v0 type: Pong-PLE-v0 --- # **Reinforce** Agent playing **Pong-PLE-v0** This is a trained model of a **Reinforce** agent playing **Pong-PLE-v0** . To learn to use this model and train yours check Unit 5 of the Deep Reinforcement Learning Class: https://github.com/huggingface/deep-rl-class/tree/main/unit5
ejin/bert-base-cased-finetuned-ner
ejin
2022-07-27T21:16:41Z
3
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "generated_from_trainer", "dataset:conll2003", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2022-07-26T20:04:05Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - conll2003 metrics: - precision - recall - f1 - accuracy model-index: - name: bert-base-cased-finetuned-ner results: - task: name: Token Classification type: token-classification dataset: name: conll2003 type: conll2003 config: conll2003 split: train args: conll2003 metrics: - name: Precision type: precision value: 0.8940432730834298 - name: Recall type: recall value: 0.9008612955320294 - name: F1 type: f1 value: 0.8974393350315055 - name: Accuracy type: accuracy value: 0.9749955848590098 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-cased-finetuned-ner This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the conll2003 dataset. It achieves the following results on the evaluation set: - Loss: 0.0919 - Precision: 0.8940 - Recall: 0.9009 - F1: 0.8974 - Accuracy: 0.9750 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.1147 | 1.0 | 1756 | 0.0919 | 0.8940 | 0.9009 | 0.8974 | 0.9750 | ### Framework versions - Transformers 4.21.0 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1