modelId
stringlengths
5
139
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]date
2020-02-15 11:33:14
2025-08-29 06:27:22
downloads
int64
0
223M
likes
int64
0
11.7k
library_name
stringclasses
525 values
tags
listlengths
1
4.05k
pipeline_tag
stringclasses
55 values
createdAt
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-08-29 06:27:10
card
stringlengths
11
1.01M
Helsinki-NLP/opus-mt-de-ln
Helsinki-NLP
2023-08-16T11:28:12Z
127
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "ln", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-ln * source languages: de * target languages: ln * OPUS readme: [de-ln](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-ln/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-ln/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-ln/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-ln/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.de.ln | 26.7 | 0.504 |
Helsinki-NLP/opus-mt-de-kg
Helsinki-NLP
2023-08-16T11:28:11Z
113
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "kg", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-kg * source languages: de * target languages: kg * OPUS readme: [de-kg](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-kg/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-kg/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-kg/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-kg/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.de.kg | 24.4 | 0.474 |
Helsinki-NLP/opus-mt-de-it
Helsinki-NLP
2023-08-16T11:28:10Z
1,692
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "it", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-it * source languages: de * target languages: it * OPUS readme: [de-it](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-it/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-it/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-it/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-it/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.de.it | 45.3 | 0.671 |
Helsinki-NLP/opus-mt-de-iso
Helsinki-NLP
2023-08-16T11:28:09Z
122
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "iso", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-iso * source languages: de * target languages: iso * OPUS readme: [de-iso](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-iso/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-iso/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-iso/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-iso/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.de.iso | 21.4 | 0.389 |
Helsinki-NLP/opus-mt-de-ilo
Helsinki-NLP
2023-08-16T11:28:06Z
129
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "ilo", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-ilo * source languages: de * target languages: ilo * OPUS readme: [de-ilo](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-ilo/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-ilo/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-ilo/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-ilo/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.de.ilo | 29.8 | 0.533 |
Helsinki-NLP/opus-mt-de-hr
Helsinki-NLP
2023-08-16T11:28:02Z
198
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "hr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-hr * source languages: de * target languages: hr * OPUS readme: [de-hr](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-hr/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-26.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-hr/opus-2020-01-26.zip) * test set translations: [opus-2020-01-26.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-hr/opus-2020-01-26.test.txt) * test set scores: [opus-2020-01-26.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-hr/opus-2020-01-26.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.de.hr | 42.6 | 0.643 |
Helsinki-NLP/opus-mt-de-hil
Helsinki-NLP
2023-08-16T11:28:00Z
115
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "hil", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-hil * source languages: de * target languages: hil * OPUS readme: [de-hil](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-hil/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-hil/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-hil/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-hil/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.de.hil | 33.9 | 0.563 |
Helsinki-NLP/opus-mt-de-gil
Helsinki-NLP
2023-08-16T11:27:55Z
110
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "gil", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-gil * source languages: de * target languages: gil * OPUS readme: [de-gil](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-gil/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-gil/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-gil/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-gil/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.de.gil | 24.0 | 0.472 |
Helsinki-NLP/opus-mt-de-gaa
Helsinki-NLP
2023-08-16T11:27:54Z
114
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "gaa", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-gaa * source languages: de * target languages: gaa * OPUS readme: [de-gaa](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-gaa/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-gaa/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-gaa/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-gaa/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.de.gaa | 26.3 | 0.471 |
Nextcloud-AI/opus-mt-de-fr
Nextcloud-AI
2023-08-16T11:27:53Z
103
0
transformers
[ "transformers", "pytorch", "tf", "rust", "marian", "text2text-generation", "translation", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2024-02-23T10:38:23Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-fr * source languages: de * target languages: fr * OPUS readme: [de-fr](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-fr/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-fr/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-fr/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-fr/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | euelections_dev2019.transformer-align.de | 32.2 | 0.590 | | newssyscomb2009.de.fr | 26.8 | 0.553 | | news-test2008.de.fr | 26.4 | 0.548 | | newstest2009.de.fr | 25.6 | 0.539 | | newstest2010.de.fr | 29.1 | 0.572 | | newstest2011.de.fr | 26.9 | 0.551 | | newstest2012.de.fr | 27.7 | 0.554 | | newstest2013.de.fr | 29.5 | 0.560 | | newstest2019-defr.de.fr | 36.6 | 0.625 | | Tatoeba.de.fr | 49.2 | 0.664 |
Helsinki-NLP/opus-mt-de-fj
Helsinki-NLP
2023-08-16T11:27:52Z
117
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "fj", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-fj * source languages: de * target languages: fj * OPUS readme: [de-fj](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-fj/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-fj/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-fj/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-fj/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.de.fj | 24.6 | 0.470 |
Nextcloud-AI/opus-mt-de-fi
Nextcloud-AI
2023-08-16T11:27:51Z
106
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2024-02-23T10:38:13Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-fi * source languages: de * target languages: fi * OPUS readme: [de-fi](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-fi/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-fi/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-fi/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-fi/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.de.fi | 40.0 | 0.628 |
Helsinki-NLP/opus-mt-de-et
Helsinki-NLP
2023-08-16T11:27:49Z
121
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "et", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-et * source languages: de * target languages: et * OPUS readme: [de-et](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-et/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-et/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-et/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-et/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.de.et | 20.2 | 0.465 |
Helsinki-NLP/opus-mt-de-el
Helsinki-NLP
2023-08-16T11:27:44Z
181
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "el", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-el * source languages: de * target languages: el * OPUS readme: [de-el](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-el/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-el/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-el/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-el/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.de.el | 45.7 | 0.649 |
Helsinki-NLP/opus-mt-de-da
Helsinki-NLP
2023-08-16T11:27:40Z
5,637
1
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "da", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-da * source languages: de * target languages: da * OPUS readme: [de-da](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-da/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-29.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-da/opus-2020-01-29.zip) * test set translations: [opus-2020-01-29.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-da/opus-2020-01-29.test.txt) * test set scores: [opus-2020-01-29.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-da/opus-2020-01-29.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.de.da | 57.2 | 0.730 |
Helsinki-NLP/opus-mt-de-cs
Helsinki-NLP
2023-08-16T11:27:39Z
312
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "cs", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-cs * source languages: de * target languages: cs * OPUS readme: [de-cs](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-cs/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-cs/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-cs/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-cs/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | newssyscomb2009.de.cs | 22.4 | 0.499 | | news-test2008.de.cs | 20.2 | 0.487 | | newstest2009.de.cs | 20.9 | 0.485 | | newstest2010.de.cs | 22.7 | 0.510 | | newstest2011.de.cs | 21.2 | 0.487 | | newstest2012.de.cs | 20.9 | 0.479 | | newstest2013.de.cs | 23.0 | 0.500 | | newstest2019-decs.de.cs | 22.5 | 0.495 | | Tatoeba.de.cs | 42.2 | 0.625 |
Helsinki-NLP/opus-mt-de-crs
Helsinki-NLP
2023-08-16T11:27:38Z
118
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "crs", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-crs * source languages: de * target languages: crs * OPUS readme: [de-crs](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-crs/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-crs/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-crs/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-crs/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.de.crs | 24.1 | 0.429 |
Helsinki-NLP/opus-mt-de-ca
Helsinki-NLP
2023-08-16T11:27:37Z
181
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "ca", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - de - ca tags: - translation license: apache-2.0 --- ### deu-cat * source group: German * target group: Catalan * OPUS readme: [deu-cat](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/deu-cat/README.md) * model: transformer-align * source language(s): deu * target language(s): cat * model: transformer-align * pre-processing: normalization + SentencePiece (spm12k,spm12k) * download original weights: [opus-2020-06-16.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/deu-cat/opus-2020-06-16.zip) * test set translations: [opus-2020-06-16.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/deu-cat/opus-2020-06-16.test.txt) * test set scores: [opus-2020-06-16.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/deu-cat/opus-2020-06-16.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.deu.cat | 37.4 | 0.582 | ### System Info: - hf_name: deu-cat - source_languages: deu - target_languages: cat - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/deu-cat/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['de', 'ca'] - src_constituents: {'deu'} - tgt_constituents: {'cat'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm12k,spm12k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/deu-cat/opus-2020-06-16.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/deu-cat/opus-2020-06-16.test.txt - src_alpha3: deu - tgt_alpha3: cat - short_pair: de-ca - chrF2_score: 0.5820000000000001 - bleu: 37.4 - brevity_penalty: 0.956 - ref_len: 5507.0 - src_name: German - tgt_name: Catalan - train_date: 2020-06-16 - src_alpha2: de - tgt_alpha2: ca - prefer_old: False - long_pair: deu-cat - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Nextcloud-AI/opus-mt-de-ar
Nextcloud-AI
2023-08-16T11:27:30Z
105
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "ar", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2024-02-23T10:37:47Z
--- language: - de - ar tags: - translation license: apache-2.0 --- ### deu-ara * source group: German * target group: Arabic * OPUS readme: [deu-ara](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/deu-ara/README.md) * model: transformer-align * source language(s): deu * target language(s): afb apc ara ara_Latn arq arz * model: transformer-align * pre-processing: normalization + SentencePiece (spm32k,spm32k) * a sentence initial language token is required in the form of `>>id<<` (id = valid target language ID) * download original weights: [opus-2020-07-03.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/deu-ara/opus-2020-07-03.zip) * test set translations: [opus-2020-07-03.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/deu-ara/opus-2020-07-03.test.txt) * test set scores: [opus-2020-07-03.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/deu-ara/opus-2020-07-03.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.deu.ara | 19.7 | 0.486 | ### System Info: - hf_name: deu-ara - source_languages: deu - target_languages: ara - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/deu-ara/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['de', 'ar'] - src_constituents: {'deu'} - tgt_constituents: {'apc', 'ara', 'arq_Latn', 'arq', 'afb', 'ara_Latn', 'apc_Latn', 'arz'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/deu-ara/opus-2020-07-03.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/deu-ara/opus-2020-07-03.test.txt - src_alpha3: deu - tgt_alpha3: ara - short_pair: de-ar - chrF2_score: 0.486 - bleu: 19.7 - brevity_penalty: 0.993 - ref_len: 6324.0 - src_name: German - tgt_name: Arabic - train_date: 2020-07-03 - src_alpha2: de - tgt_alpha2: ar - prefer_old: False - long_pair: deu-ara - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-de-af
Helsinki-NLP
2023-08-16T11:27:29Z
259
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "af", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - de - af tags: - translation license: apache-2.0 --- ### deu-afr * source group: German * target group: Afrikaans * OPUS readme: [deu-afr](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/deu-afr/README.md) * model: transformer-align * source language(s): deu * target language(s): afr * model: transformer-align * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-06-17.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/deu-afr/opus-2020-06-17.zip) * test set translations: [opus-2020-06-17.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/deu-afr/opus-2020-06-17.test.txt) * test set scores: [opus-2020-06-17.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/deu-afr/opus-2020-06-17.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.deu.afr | 51.3 | 0.690 | ### System Info: - hf_name: deu-afr - source_languages: deu - target_languages: afr - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/deu-afr/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['de', 'af'] - src_constituents: {'deu'} - tgt_constituents: {'afr'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/deu-afr/opus-2020-06-17.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/deu-afr/opus-2020-06-17.test.txt - src_alpha3: deu - tgt_alpha3: afr - short_pair: de-af - chrF2_score: 0.69 - bleu: 51.3 - brevity_penalty: 1.0 - ref_len: 9507.0 - src_name: German - tgt_name: Afrikaans - train_date: 2020-06-17 - src_alpha2: de - tgt_alpha2: af - prefer_old: False - long_pair: deu-afr - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-de-ZH
Helsinki-NLP
2023-08-16T11:27:28Z
379
2
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "de", "zh", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-de-ZH * source languages: de * target languages: cmn,cn,yue,ze_zh,zh_cn,zh_CN,zh_HK,zh_tw,zh_TW,zh_yue,zhs,zht,zh * OPUS readme: [de-cmn+cn+yue+ze_zh+zh_cn+zh_CN+zh_HK+zh_tw+zh_TW+zh_yue+zhs+zht+zh](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de-cmn+cn+yue+ze_zh+zh_cn+zh_CN+zh_HK+zh_tw+zh_TW+zh_yue+zhs+zht+zh/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * a sentence initial language token is required in the form of `>>id<<` (id = valid target language ID) * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/de-cmn+cn+yue+ze_zh+zh_cn+zh_CN+zh_HK+zh_tw+zh_TW+zh_yue+zhs+zht+zh/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-cmn+cn+yue+ze_zh+zh_cn+zh_CN+zh_HK+zh_tw+zh_TW+zh_yue+zhs+zht+zh/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de-cmn+cn+yue+ze_zh+zh_cn+zh_CN+zh_HK+zh_tw+zh_TW+zh_yue+zhs+zht+zh/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | bible-uedin.de.zh | 24.4 | 0.335 |
Helsinki-NLP/opus-mt-da-eo
Helsinki-NLP
2023-08-16T11:27:22Z
108
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "da", "eo", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - da - eo tags: - translation license: apache-2.0 --- ### dan-epo * source group: Danish * target group: Esperanto * OPUS readme: [dan-epo](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/dan-epo/README.md) * model: transformer-align * source language(s): dan * target language(s): epo * model: transformer-align * pre-processing: normalization + SentencePiece (spm4k,spm4k) * download original weights: [opus-2020-06-16.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/dan-epo/opus-2020-06-16.zip) * test set translations: [opus-2020-06-16.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/dan-epo/opus-2020-06-16.test.txt) * test set scores: [opus-2020-06-16.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/dan-epo/opus-2020-06-16.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.dan.epo | 23.6 | 0.432 | ### System Info: - hf_name: dan-epo - source_languages: dan - target_languages: epo - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/dan-epo/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['da', 'eo'] - src_constituents: {'dan'} - tgt_constituents: {'epo'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm4k,spm4k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/dan-epo/opus-2020-06-16.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/dan-epo/opus-2020-06-16.test.txt - src_alpha3: dan - tgt_alpha3: epo - short_pair: da-eo - chrF2_score: 0.43200000000000005 - bleu: 23.6 - brevity_penalty: 0.9420000000000001 - ref_len: 69856.0 - src_name: Danish - tgt_name: Esperanto - train_date: 2020-06-16 - src_alpha2: da - tgt_alpha2: eo - prefer_old: False - long_pair: dan-epo - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-da-en
Helsinki-NLP
2023-08-16T11:27:21Z
86,194
5
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "da", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-da-en * source languages: da * target languages: en * OPUS readme: [da-en](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/da-en/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2019-12-18.zip](https://object.pouta.csc.fi/OPUS-MT-models/da-en/opus-2019-12-18.zip) * test set translations: [opus-2019-12-18.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/da-en/opus-2019-12-18.test.txt) * test set scores: [opus-2019-12-18.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/da-en/opus-2019-12-18.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.da.en | 63.6 | 0.769 |
Helsinki-NLP/opus-mt-da-de
Helsinki-NLP
2023-08-16T11:27:20Z
16,695
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "da", "de", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-da-de * source languages: da * target languages: de * OPUS readme: [da-de](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/da-de/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-26.zip](https://object.pouta.csc.fi/OPUS-MT-models/da-de/opus-2020-01-26.zip) * test set translations: [opus-2020-01-26.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/da-de/opus-2020-01-26.test.txt) * test set scores: [opus-2020-01-26.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/da-de/opus-2020-01-26.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.da.de | 57.4 | 0.740 |
Helsinki-NLP/opus-mt-cy-en
Helsinki-NLP
2023-08-16T11:27:19Z
4,822
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "cy", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-cy-en * source languages: cy * target languages: en * OPUS readme: [cy-en](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/cy-en/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2019-12-18.zip](https://object.pouta.csc.fi/OPUS-MT-models/cy-en/opus-2019-12-18.zip) * test set translations: [opus-2019-12-18.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/cy-en/opus-2019-12-18.test.txt) * test set scores: [opus-2019-12-18.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/cy-en/opus-2019-12-18.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.cy.en | 33.0 | 0.525 |
Helsinki-NLP/opus-mt-csg-es
Helsinki-NLP
2023-08-16T11:27:15Z
117
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "csg", "es", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-csg-es * source languages: csg * target languages: es * OPUS readme: [csg-es](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/csg-es/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-15.zip](https://object.pouta.csc.fi/OPUS-MT-models/csg-es/opus-2020-01-15.zip) * test set translations: [opus-2020-01-15.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/csg-es/opus-2020-01-15.test.txt) * test set scores: [opus-2020-01-15.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/csg-es/opus-2020-01-15.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.csg.es | 93.1 | 0.952 |
Helsinki-NLP/opus-mt-cs-sv
Helsinki-NLP
2023-08-16T11:27:13Z
126
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "cs", "sv", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-cs-sv * source languages: cs * target languages: sv * OPUS readme: [cs-sv](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/cs-sv/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/cs-sv/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/cs-sv/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/cs-sv/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.cs.sv | 30.6 | 0.527 |
Helsinki-NLP/opus-mt-cs-fi
Helsinki-NLP
2023-08-16T11:27:11Z
117
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "cs", "fi", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-cs-fi * source languages: cs * target languages: fi * OPUS readme: [cs-fi](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/cs-fi/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/cs-fi/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/cs-fi/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/cs-fi/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.cs.fi | 25.5 | 0.523 |
Helsinki-NLP/opus-mt-cs-eo
Helsinki-NLP
2023-08-16T11:27:10Z
111
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "cs", "eo", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - cs - eo tags: - translation license: apache-2.0 --- ### ces-epo * source group: Czech * target group: Esperanto * OPUS readme: [ces-epo](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ces-epo/README.md) * model: transformer-align * source language(s): ces * target language(s): epo * model: transformer-align * pre-processing: normalization + SentencePiece (spm4k,spm4k) * download original weights: [opus-2020-06-16.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ces-epo/opus-2020-06-16.zip) * test set translations: [opus-2020-06-16.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ces-epo/opus-2020-06-16.test.txt) * test set scores: [opus-2020-06-16.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ces-epo/opus-2020-06-16.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.ces.epo | 26.0 | 0.459 | ### System Info: - hf_name: ces-epo - source_languages: ces - target_languages: epo - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ces-epo/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['cs', 'eo'] - src_constituents: {'ces'} - tgt_constituents: {'epo'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm4k,spm4k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ces-epo/opus-2020-06-16.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ces-epo/opus-2020-06-16.test.txt - src_alpha3: ces - tgt_alpha3: epo - short_pair: cs-eo - chrF2_score: 0.45899999999999996 - bleu: 26.0 - brevity_penalty: 0.94 - ref_len: 24901.0 - src_name: Czech - tgt_name: Esperanto - train_date: 2020-06-16 - src_alpha2: cs - tgt_alpha2: eo - prefer_old: False - long_pair: ces-epo - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-cs-de
Helsinki-NLP
2023-08-16T11:27:08Z
204
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "cs", "de", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-cs-de * source languages: cs * target languages: de * OPUS readme: [cs-de](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/cs-de/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/cs-de/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/cs-de/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/cs-de/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | newssyscomb2009.cs.de | 22.0 | 0.525 | | news-test2008.cs.de | 21.1 | 0.520 | | newstest2009.cs.de | 22.2 | 0.525 | | newstest2010.cs.de | 22.1 | 0.527 | | newstest2011.cs.de | 21.6 | 0.515 | | newstest2012.cs.de | 22.2 | 0.516 | | newstest2013.cs.de | 24.8 | 0.538 | | newstest2019-csde.cs.de | 23.6 | 0.530 | | Tatoeba.cs.de | 51.6 | 0.687 |
Helsinki-NLP/opus-mt-crs-fr
Helsinki-NLP
2023-08-16T11:27:06Z
111
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "crs", "fr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-crs-fr * source languages: crs * target languages: fr * OPUS readme: [crs-fr](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/crs-fr/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/crs-fr/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/crs-fr/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/crs-fr/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.crs.fr | 29.4 | 0.475 |
Helsinki-NLP/opus-mt-crs-fi
Helsinki-NLP
2023-08-16T11:27:05Z
119
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "crs", "fi", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-crs-fi * source languages: crs * target languages: fi * OPUS readme: [crs-fi](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/crs-fi/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/crs-fi/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/crs-fi/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/crs-fi/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.crs.fi | 25.6 | 0.479 |
Helsinki-NLP/opus-mt-crs-es
Helsinki-NLP
2023-08-16T11:27:04Z
109
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "crs", "es", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-crs-es * source languages: crs * target languages: es * OPUS readme: [crs-es](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/crs-es/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-15.zip](https://object.pouta.csc.fi/OPUS-MT-models/crs-es/opus-2020-01-15.zip) * test set translations: [opus-2020-01-15.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/crs-es/opus-2020-01-15.test.txt) * test set scores: [opus-2020-01-15.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/crs-es/opus-2020-01-15.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.crs.es | 26.1 | 0.445 |
Helsinki-NLP/opus-mt-crs-de
Helsinki-NLP
2023-08-16T11:27:02Z
112
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "crs", "de", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-crs-de * source languages: crs * target languages: de * OPUS readme: [crs-de](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/crs-de/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/crs-de/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/crs-de/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/crs-de/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.crs.de | 20.4 | 0.397 |
Helsinki-NLP/opus-mt-chk-fr
Helsinki-NLP
2023-08-16T11:26:57Z
104
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "chk", "fr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-chk-fr * source languages: chk * target languages: fr * OPUS readme: [chk-fr](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/chk-fr/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/chk-fr/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/chk-fr/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/chk-fr/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.chk.fr | 22.4 | 0.387 |
Helsinki-NLP/opus-mt-chk-es
Helsinki-NLP
2023-08-16T11:26:56Z
109
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "chk", "es", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-chk-es * source languages: chk * target languages: es * OPUS readme: [chk-es](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/chk-es/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-15.zip](https://object.pouta.csc.fi/OPUS-MT-models/chk-es/opus-2020-01-15.zip) * test set translations: [opus-2020-01-15.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/chk-es/opus-2020-01-15.test.txt) * test set scores: [opus-2020-01-15.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/chk-es/opus-2020-01-15.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.chk.es | 20.8 | 0.374 |
Helsinki-NLP/opus-mt-chk-en
Helsinki-NLP
2023-08-16T11:26:55Z
117
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "chk", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-chk-en * source languages: chk * target languages: en * OPUS readme: [chk-en](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/chk-en/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2019-12-18.zip](https://object.pouta.csc.fi/OPUS-MT-models/chk-en/opus-2019-12-18.zip) * test set translations: [opus-2019-12-18.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/chk-en/opus-2019-12-18.test.txt) * test set scores: [opus-2019-12-18.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/chk-en/opus-2019-12-18.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.chk.en | 31.2 | 0.465 |
Helsinki-NLP/opus-mt-ceb-fr
Helsinki-NLP
2023-08-16T11:26:52Z
117
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ceb", "fr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-ceb-fr * source languages: ceb * target languages: fr * OPUS readme: [ceb-fr](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/ceb-fr/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/ceb-fr/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/ceb-fr/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/ceb-fr/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.ceb.fr | 30.0 | 0.491 |
Helsinki-NLP/opus-mt-ceb-es
Helsinki-NLP
2023-08-16T11:26:50Z
111
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ceb", "es", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-ceb-es * source languages: ceb * target languages: es * OPUS readme: [ceb-es](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/ceb-es/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-15.zip](https://object.pouta.csc.fi/OPUS-MT-models/ceb-es/opus-2020-01-15.zip) * test set translations: [opus-2020-01-15.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/ceb-es/opus-2020-01-15.test.txt) * test set scores: [opus-2020-01-15.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/ceb-es/opus-2020-01-15.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.ceb.es | 31.6 | 0.508 |
Helsinki-NLP/opus-mt-ceb-en
Helsinki-NLP
2023-08-16T11:26:49Z
1,276
1
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ceb", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - ceb - en tags: - translation license: apache-2.0 --- ### ceb-eng * source group: Cebuano * target group: English * OPUS readme: [ceb-eng](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ceb-eng/README.md) * model: transformer-align * source language(s): ceb * target language(s): eng * model: transformer-align * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-06-17.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ceb-eng/opus-2020-06-17.zip) * test set translations: [opus-2020-06-17.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ceb-eng/opus-2020-06-17.test.txt) * test set scores: [opus-2020-06-17.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ceb-eng/opus-2020-06-17.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.ceb.eng | 21.5 | 0.387 | ### System Info: - hf_name: ceb-eng - source_languages: ceb - target_languages: eng - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ceb-eng/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ceb', 'en'] - src_constituents: {'ceb'} - tgt_constituents: {'eng'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ceb-eng/opus-2020-06-17.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ceb-eng/opus-2020-06-17.test.txt - src_alpha3: ceb - tgt_alpha3: eng - short_pair: ceb-en - chrF2_score: 0.387 - bleu: 21.5 - brevity_penalty: 1.0 - ref_len: 2293.0 - src_name: Cebuano - tgt_name: English - train_date: 2020-06-17 - src_alpha2: ceb - tgt_alpha2: en - prefer_old: False - long_pair: ceb-eng - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ccs-en
Helsinki-NLP
2023-08-16T11:26:48Z
121
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ka", "ccs", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - ka - ccs - en tags: - translation license: apache-2.0 --- ### ccs-eng * source group: South Caucasian languages * target group: English * OPUS readme: [ccs-eng](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ccs-eng/README.md) * model: transformer * source language(s): kat * target language(s): eng * model: transformer * pre-processing: normalization + SentencePiece (spm12k,spm12k) * download original weights: [opus2m-2020-07-31.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ccs-eng/opus2m-2020-07-31.zip) * test set translations: [opus2m-2020-07-31.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ccs-eng/opus2m-2020-07-31.test.txt) * test set scores: [opus2m-2020-07-31.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ccs-eng/opus2m-2020-07-31.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.kat-eng.kat.eng | 18.0 | 0.357 | | Tatoeba-test.multi.eng | 18.0 | 0.357 | ### System Info: - hf_name: ccs-eng - source_languages: ccs - target_languages: eng - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ccs-eng/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ka', 'ccs', 'en'] - src_constituents: {'kat'} - tgt_constituents: {'eng'} - src_multilingual: True - tgt_multilingual: False - prepro: normalization + SentencePiece (spm12k,spm12k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ccs-eng/opus2m-2020-07-31.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ccs-eng/opus2m-2020-07-31.test.txt - src_alpha3: ccs - tgt_alpha3: eng - short_pair: ccs-en - chrF2_score: 0.35700000000000004 - bleu: 18.0 - brevity_penalty: 1.0 - ref_len: 5992.0 - src_name: South Caucasian languages - tgt_name: English - train_date: 2020-07-31 - src_alpha2: ccs - tgt_alpha2: en - prefer_old: False - long_pair: ccs-eng - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ca-pt
Helsinki-NLP
2023-08-16T11:26:45Z
132
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ca", "pt", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - ca - pt tags: - translation license: apache-2.0 --- ### cat-por * source group: Catalan * target group: Portuguese * OPUS readme: [cat-por](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/cat-por/README.md) * model: transformer-align * source language(s): cat * target language(s): por * model: transformer-align * pre-processing: normalization + SentencePiece (spm12k,spm12k) * download original weights: [opus-2020-06-17.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/cat-por/opus-2020-06-17.zip) * test set translations: [opus-2020-06-17.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/cat-por/opus-2020-06-17.test.txt) * test set scores: [opus-2020-06-17.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/cat-por/opus-2020-06-17.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.cat.por | 44.9 | 0.658 | ### System Info: - hf_name: cat-por - source_languages: cat - target_languages: por - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/cat-por/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ca', 'pt'] - src_constituents: {'cat'} - tgt_constituents: {'por'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm12k,spm12k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/cat-por/opus-2020-06-17.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/cat-por/opus-2020-06-17.test.txt - src_alpha3: cat - tgt_alpha3: por - short_pair: ca-pt - chrF2_score: 0.6579999999999999 - bleu: 44.9 - brevity_penalty: 0.953 - ref_len: 5847.0 - src_name: Catalan - tgt_name: Portuguese - train_date: 2020-06-17 - src_alpha2: ca - tgt_alpha2: pt - prefer_old: False - long_pair: cat-por - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ca-nl
Helsinki-NLP
2023-08-16T11:26:44Z
122
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ca", "nl", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - ca - nl tags: - translation license: apache-2.0 --- ### cat-nld * source group: Catalan * target group: Dutch * OPUS readme: [cat-nld](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/cat-nld/README.md) * model: transformer-align * source language(s): cat * target language(s): nld * model: transformer-align * pre-processing: normalization + SentencePiece (spm12k,spm12k) * download original weights: [opus-2020-06-16.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/cat-nld/opus-2020-06-16.zip) * test set translations: [opus-2020-06-16.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/cat-nld/opus-2020-06-16.test.txt) * test set scores: [opus-2020-06-16.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/cat-nld/opus-2020-06-16.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.cat.nld | 45.1 | 0.632 | ### System Info: - hf_name: cat-nld - source_languages: cat - target_languages: nld - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/cat-nld/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ca', 'nl'] - src_constituents: {'cat'} - tgt_constituents: {'nld'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm12k,spm12k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/cat-nld/opus-2020-06-16.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/cat-nld/opus-2020-06-16.test.txt - src_alpha3: cat - tgt_alpha3: nld - short_pair: ca-nl - chrF2_score: 0.632 - bleu: 45.1 - brevity_penalty: 0.965 - ref_len: 4157.0 - src_name: Catalan - tgt_name: Dutch - train_date: 2020-06-16 - src_alpha2: ca - tgt_alpha2: nl - prefer_old: False - long_pair: cat-nld - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ca-it
Helsinki-NLP
2023-08-16T11:26:42Z
169
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ca", "it", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - ca - it tags: - translation license: apache-2.0 --- ### cat-ita * source group: Catalan * target group: Italian * OPUS readme: [cat-ita](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/cat-ita/README.md) * model: transformer-align * source language(s): cat * target language(s): ita * model: transformer-align * pre-processing: normalization + SentencePiece (spm12k,spm12k) * download original weights: [opus-2020-06-16.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/cat-ita/opus-2020-06-16.zip) * test set translations: [opus-2020-06-16.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/cat-ita/opus-2020-06-16.test.txt) * test set scores: [opus-2020-06-16.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/cat-ita/opus-2020-06-16.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.cat.ita | 48.6 | 0.690 | ### System Info: - hf_name: cat-ita - source_languages: cat - target_languages: ita - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/cat-ita/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ca', 'it'] - src_constituents: {'cat'} - tgt_constituents: {'ita'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm12k,spm12k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/cat-ita/opus-2020-06-16.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/cat-ita/opus-2020-06-16.test.txt - src_alpha3: cat - tgt_alpha3: ita - short_pair: ca-it - chrF2_score: 0.69 - bleu: 48.6 - brevity_penalty: 0.985 - ref_len: 1995.0 - src_name: Catalan - tgt_name: Italian - train_date: 2020-06-16 - src_alpha2: ca - tgt_alpha2: it - prefer_old: False - long_pair: cat-ita - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ca-fr
Helsinki-NLP
2023-08-16T11:26:41Z
697
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ca", "fr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - ca - fr tags: - translation license: apache-2.0 --- ### cat-fra * source group: Catalan * target group: French * OPUS readme: [cat-fra](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/cat-fra/README.md) * model: transformer-align * source language(s): cat * target language(s): fra * model: transformer-align * pre-processing: normalization + SentencePiece (spm12k,spm12k) * download original weights: [opus-2020-06-16.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/cat-fra/opus-2020-06-16.zip) * test set translations: [opus-2020-06-16.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/cat-fra/opus-2020-06-16.test.txt) * test set scores: [opus-2020-06-16.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/cat-fra/opus-2020-06-16.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.cat.fra | 52.4 | 0.694 | ### System Info: - hf_name: cat-fra - source_languages: cat - target_languages: fra - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/cat-fra/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ca', 'fr'] - src_constituents: {'cat'} - tgt_constituents: {'fra'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm12k,spm12k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/cat-fra/opus-2020-06-16.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/cat-fra/opus-2020-06-16.test.txt - src_alpha3: cat - tgt_alpha3: fra - short_pair: ca-fr - chrF2_score: 0.6940000000000001 - bleu: 52.4 - brevity_penalty: 0.987 - ref_len: 5517.0 - src_name: Catalan - tgt_name: French - train_date: 2020-06-16 - src_alpha2: ca - tgt_alpha2: fr - prefer_old: False - long_pair: cat-fra - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ca-es
Helsinki-NLP
2023-08-16T11:26:40Z
832
1
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ca", "es", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-ca-es * source languages: ca * target languages: es * OPUS readme: [ca-es](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/ca-es/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-15.zip](https://object.pouta.csc.fi/OPUS-MT-models/ca-es/opus-2020-01-15.zip) * test set translations: [opus-2020-01-15.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/ca-es/opus-2020-01-15.test.txt) * test set scores: [opus-2020-01-15.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/ca-es/opus-2020-01-15.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.ca.es | 74.9 | 0.863 |
Helsinki-NLP/opus-mt-bzs-sv
Helsinki-NLP
2023-08-16T11:26:37Z
122
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bzs", "sv", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-bzs-sv * source languages: bzs * target languages: sv * OPUS readme: [bzs-sv](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/bzs-sv/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/bzs-sv/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/bzs-sv/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/bzs-sv/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.bzs.sv | 30.7 | 0.489 |
Helsinki-NLP/opus-mt-bzs-fi
Helsinki-NLP
2023-08-16T11:26:35Z
125
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bzs", "fi", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-bzs-fi * source languages: bzs * target languages: fi * OPUS readme: [bzs-fi](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/bzs-fi/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/bzs-fi/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/bzs-fi/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/bzs-fi/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.bzs.fi | 24.7 | 0.464 |
Helsinki-NLP/opus-mt-bn-en
Helsinki-NLP
2023-08-16T11:26:30Z
7,721
7
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bn", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - bn - en tags: - translation license: apache-2.0 --- ### ben-eng * source group: Bengali * target group: English * OPUS readme: [ben-eng](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ben-eng/README.md) * model: transformer-align * source language(s): ben * target language(s): eng * model: transformer-align * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-06-17.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ben-eng/opus-2020-06-17.zip) * test set translations: [opus-2020-06-17.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ben-eng/opus-2020-06-17.test.txt) * test set scores: [opus-2020-06-17.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ben-eng/opus-2020-06-17.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.ben.eng | 49.7 | 0.641 | ### System Info: - hf_name: ben-eng - source_languages: ben - target_languages: eng - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ben-eng/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['bn', 'en'] - src_constituents: {'ben'} - tgt_constituents: {'eng'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ben-eng/opus-2020-06-17.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ben-eng/opus-2020-06-17.test.txt - src_alpha3: ben - tgt_alpha3: eng - short_pair: bn-en - chrF2_score: 0.6409999999999999 - bleu: 49.7 - brevity_penalty: 0.976 - ref_len: 13978.0 - src_name: Bengali - tgt_name: English - train_date: 2020-06-17 - src_alpha2: bn - tgt_alpha2: en - prefer_old: False - long_pair: ben-eng - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-bi-en
Helsinki-NLP
2023-08-16T11:26:26Z
142
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bi", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-bi-en * source languages: bi * target languages: en * OPUS readme: [bi-en](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/bi-en/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/bi-en/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/bi-en/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/bi-en/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.bi.en | 30.3 | 0.458 |
Helsinki-NLP/opus-mt-bg-uk
Helsinki-NLP
2023-08-16T11:26:25Z
139
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bg", "uk", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - bg - uk tags: - translation license: apache-2.0 --- ### bul-ukr * source group: Bulgarian * target group: Ukrainian * OPUS readme: [bul-ukr](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/bul-ukr/README.md) * model: transformer-align * source language(s): bul * target language(s): ukr * model: transformer-align * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-06-17.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-ukr/opus-2020-06-17.zip) * test set translations: [opus-2020-06-17.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-ukr/opus-2020-06-17.test.txt) * test set scores: [opus-2020-06-17.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-ukr/opus-2020-06-17.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.bul.ukr | 49.2 | 0.683 | ### System Info: - hf_name: bul-ukr - source_languages: bul - target_languages: ukr - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/bul-ukr/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['bg', 'uk'] - src_constituents: {'bul', 'bul_Latn'} - tgt_constituents: {'ukr'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/bul-ukr/opus-2020-06-17.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/bul-ukr/opus-2020-06-17.test.txt - src_alpha3: bul - tgt_alpha3: ukr - short_pair: bg-uk - chrF2_score: 0.6829999999999999 - bleu: 49.2 - brevity_penalty: 0.983 - ref_len: 4932.0 - src_name: Bulgarian - tgt_name: Ukrainian - train_date: 2020-06-17 - src_alpha2: bg - tgt_alpha2: uk - prefer_old: False - long_pair: bul-ukr - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-bg-sv
Helsinki-NLP
2023-08-16T11:26:22Z
119
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bg", "sv", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-bg-sv * source languages: bg * target languages: sv * OPUS readme: [bg-sv](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/bg-sv/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/bg-sv/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/bg-sv/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/bg-sv/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.bg.sv | 29.1 | 0.494 |
Helsinki-NLP/opus-mt-bg-ru
Helsinki-NLP
2023-08-16T11:26:21Z
180
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bg", "ru", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - bg - ru tags: - translation license: apache-2.0 --- ### bul-rus * source group: Bulgarian * target group: Russian * OPUS readme: [bul-rus](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/bul-rus/README.md) * model: transformer * source language(s): bul bul_Latn * target language(s): rus * model: transformer * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-07-03.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-rus/opus-2020-07-03.zip) * test set translations: [opus-2020-07-03.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-rus/opus-2020-07-03.test.txt) * test set scores: [opus-2020-07-03.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-rus/opus-2020-07-03.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.bul.rus | 48.5 | 0.691 | ### System Info: - hf_name: bul-rus - source_languages: bul - target_languages: rus - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/bul-rus/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['bg', 'ru'] - src_constituents: {'bul', 'bul_Latn'} - tgt_constituents: {'rus'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/bul-rus/opus-2020-07-03.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/bul-rus/opus-2020-07-03.test.txt - src_alpha3: bul - tgt_alpha3: rus - short_pair: bg-ru - chrF2_score: 0.691 - bleu: 48.5 - brevity_penalty: 1.0 - ref_len: 7870.0 - src_name: Bulgarian - tgt_name: Russian - train_date: 2020-07-03 - src_alpha2: bg - tgt_alpha2: ru - prefer_old: False - long_pair: bul-rus - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-bg-fr
Helsinki-NLP
2023-08-16T11:26:19Z
160
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bg", "fr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - bg - fr tags: - translation license: apache-2.0 --- ### bul-fra * source group: Bulgarian * target group: French * OPUS readme: [bul-fra](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/bul-fra/README.md) * model: transformer * source language(s): bul * target language(s): fra * model: transformer * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-07-03.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-fra/opus-2020-07-03.zip) * test set translations: [opus-2020-07-03.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-fra/opus-2020-07-03.test.txt) * test set scores: [opus-2020-07-03.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-fra/opus-2020-07-03.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.bul.fra | 53.7 | 0.693 | ### System Info: - hf_name: bul-fra - source_languages: bul - target_languages: fra - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/bul-fra/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['bg', 'fr'] - src_constituents: {'bul', 'bul_Latn'} - tgt_constituents: {'fra'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/bul-fra/opus-2020-07-03.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/bul-fra/opus-2020-07-03.test.txt - src_alpha3: bul - tgt_alpha3: fra - short_pair: bg-fr - chrF2_score: 0.693 - bleu: 53.7 - brevity_penalty: 0.977 - ref_len: 3669.0 - src_name: Bulgarian - tgt_name: French - train_date: 2020-07-03 - src_alpha2: bg - tgt_alpha2: fr - prefer_old: False - long_pair: bul-fra - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-bg-fi
Helsinki-NLP
2023-08-16T11:26:18Z
147
1
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bg", "fi", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-bg-fi * source languages: bg * target languages: fi * OPUS readme: [bg-fi](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/bg-fi/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/bg-fi/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/bg-fi/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/bg-fi/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.bg.fi | 23.7 | 0.505 |
Helsinki-NLP/opus-mt-bg-es
Helsinki-NLP
2023-08-16T11:26:16Z
125
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bg", "es", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - bg - es tags: - translation license: apache-2.0 --- ### bul-spa * source group: Bulgarian * target group: Spanish * OPUS readme: [bul-spa](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/bul-spa/README.md) * model: transformer * source language(s): bul * target language(s): spa * model: transformer * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-07-03.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-spa/opus-2020-07-03.zip) * test set translations: [opus-2020-07-03.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-spa/opus-2020-07-03.test.txt) * test set scores: [opus-2020-07-03.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-spa/opus-2020-07-03.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.bul.spa | 49.1 | 0.661 | ### System Info: - hf_name: bul-spa - source_languages: bul - target_languages: spa - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/bul-spa/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['bg', 'es'] - src_constituents: {'bul', 'bul_Latn'} - tgt_constituents: {'spa'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/bul-spa/opus-2020-07-03.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/bul-spa/opus-2020-07-03.test.txt - src_alpha3: bul - tgt_alpha3: spa - short_pair: bg-es - chrF2_score: 0.6609999999999999 - bleu: 49.1 - brevity_penalty: 0.992 - ref_len: 1783.0 - src_name: Bulgarian - tgt_name: Spanish - train_date: 2020-07-03 - src_alpha2: bg - tgt_alpha2: es - prefer_old: False - long_pair: bul-spa - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-bg-eo
Helsinki-NLP
2023-08-16T11:26:15Z
114
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bg", "eo", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - bg - eo tags: - translation license: apache-2.0 --- ### bul-epo * source group: Bulgarian * target group: Esperanto * OPUS readme: [bul-epo](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/bul-epo/README.md) * model: transformer-align * source language(s): bul * target language(s): epo * model: transformer-align * pre-processing: normalization + SentencePiece (spm4k,spm4k) * download original weights: [opus-2020-06-16.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-epo/opus-2020-06-16.zip) * test set translations: [opus-2020-06-16.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-epo/opus-2020-06-16.test.txt) * test set scores: [opus-2020-06-16.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/bul-epo/opus-2020-06-16.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.bul.epo | 24.5 | 0.438 | ### System Info: - hf_name: bul-epo - source_languages: bul - target_languages: epo - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/bul-epo/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['bg', 'eo'] - src_constituents: {'bul', 'bul_Latn'} - tgt_constituents: {'epo'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm4k,spm4k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/bul-epo/opus-2020-06-16.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/bul-epo/opus-2020-06-16.test.txt - src_alpha3: bul - tgt_alpha3: epo - short_pair: bg-eo - chrF2_score: 0.43799999999999994 - bleu: 24.5 - brevity_penalty: 0.9670000000000001 - ref_len: 4043.0 - src_name: Bulgarian - tgt_name: Esperanto - train_date: 2020-06-16 - src_alpha2: bg - tgt_alpha2: eo - prefer_old: False - long_pair: bul-epo - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ber-fr
Helsinki-NLP
2023-08-16T11:26:12Z
133
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ber", "fr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-ber-fr * source languages: ber * target languages: fr * OPUS readme: [ber-fr](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/ber-fr/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/ber-fr/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/ber-fr/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/ber-fr/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.ber.fr | 60.2 | 0.754 |
Helsinki-NLP/opus-mt-ber-es
Helsinki-NLP
2023-08-16T11:26:11Z
119
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ber", "es", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-ber-es * source languages: ber * target languages: es * OPUS readme: [ber-es](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/ber-es/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-15.zip](https://object.pouta.csc.fi/OPUS-MT-models/ber-es/opus-2020-01-15.zip) * test set translations: [opus-2020-01-15.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/ber-es/opus-2020-01-15.test.txt) * test set scores: [opus-2020-01-15.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/ber-es/opus-2020-01-15.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.ber.es | 33.8 | 0.487 |
Helsinki-NLP/opus-mt-bem-fi
Helsinki-NLP
2023-08-16T11:26:07Z
113
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bem", "fi", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-bem-fi * source languages: bem * target languages: fi * OPUS readme: [bem-fi](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/bem-fi/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/bem-fi/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/bem-fi/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/bem-fi/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.bem.fi | 22.8 | 0.439 |
Helsinki-NLP/opus-mt-bcl-en
Helsinki-NLP
2023-08-16T11:25:59Z
241
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "bcl", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-bcl-en * source languages: bcl * target languages: en * OPUS readme: [bcl-en](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/bcl-en/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-02-11.zip](https://object.pouta.csc.fi/OPUS-MT-models/bcl-en/opus-2020-02-11.zip) * test set translations: [opus-2020-02-11.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/bcl-en/opus-2020-02-11.test.txt) * test set scores: [opus-2020-02-11.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/bcl-en/opus-2020-02-11.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.bcl.en | 56.1 | 0.697 |
Helsinki-NLP/opus-mt-ase-sv
Helsinki-NLP
2023-08-16T11:25:52Z
127
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ase", "sv", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-ase-sv * source languages: ase * target languages: sv * OPUS readme: [ase-sv](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/ase-sv/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/ase-sv/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/ase-sv/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/ase-sv/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.ase.sv | 39.7 | 0.576 |
Helsinki-NLP/opus-mt-ase-es
Helsinki-NLP
2023-08-16T11:25:50Z
119
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ase", "es", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-ase-es * source languages: ase * target languages: es * OPUS readme: [ase-es](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/ase-es/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/ase-es/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/ase-es/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/ase-es/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.ase.es | 31.7 | 0.498 |
Helsinki-NLP/opus-mt-ase-en
Helsinki-NLP
2023-08-16T11:25:49Z
137
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ase", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-ase-en * source languages: ase * target languages: en * OPUS readme: [ase-en](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/ase-en/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-20.zip](https://object.pouta.csc.fi/OPUS-MT-models/ase-en/opus-2020-01-20.zip) * test set translations: [opus-2020-01-20.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/ase-en/opus-2020-01-20.test.txt) * test set scores: [opus-2020-01-20.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/ase-en/opus-2020-01-20.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.ase.en | 99.5 | 0.997 |
Helsinki-NLP/opus-mt-art-en
Helsinki-NLP
2023-08-16T11:25:47Z
131
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "eo", "io", "art", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - eo - io - art - en tags: - translation license: apache-2.0 --- ### art-eng * source group: Artificial languages * target group: English * OPUS readme: [art-eng](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/art-eng/README.md) * model: transformer * source language(s): afh_Latn avk_Latn dws_Latn epo ido ido_Latn ile_Latn ina_Latn jbo jbo_Cyrl jbo_Latn ldn_Latn lfn_Cyrl lfn_Latn nov_Latn qya qya_Latn sjn_Latn tlh_Latn tzl tzl_Latn vol_Latn * target language(s): eng * model: transformer * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus2m-2020-07-31.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/art-eng/opus2m-2020-07-31.zip) * test set translations: [opus2m-2020-07-31.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/art-eng/opus2m-2020-07-31.test.txt) * test set scores: [opus2m-2020-07-31.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/art-eng/opus2m-2020-07-31.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.afh-eng.afh.eng | 1.2 | 0.099 | | Tatoeba-test.avk-eng.avk.eng | 0.4 | 0.105 | | Tatoeba-test.dws-eng.dws.eng | 1.6 | 0.076 | | Tatoeba-test.epo-eng.epo.eng | 34.6 | 0.530 | | Tatoeba-test.ido-eng.ido.eng | 12.7 | 0.310 | | Tatoeba-test.ile-eng.ile.eng | 4.6 | 0.218 | | Tatoeba-test.ina-eng.ina.eng | 5.8 | 0.254 | | Tatoeba-test.jbo-eng.jbo.eng | 0.2 | 0.115 | | Tatoeba-test.ldn-eng.ldn.eng | 0.7 | 0.083 | | Tatoeba-test.lfn-eng.lfn.eng | 1.8 | 0.172 | | Tatoeba-test.multi.eng | 11.6 | 0.287 | | Tatoeba-test.nov-eng.nov.eng | 5.1 | 0.215 | | Tatoeba-test.qya-eng.qya.eng | 0.7 | 0.113 | | Tatoeba-test.sjn-eng.sjn.eng | 0.9 | 0.090 | | Tatoeba-test.tlh-eng.tlh.eng | 0.2 | 0.124 | | Tatoeba-test.tzl-eng.tzl.eng | 1.4 | 0.109 | | Tatoeba-test.vol-eng.vol.eng | 0.5 | 0.115 | ### System Info: - hf_name: art-eng - source_languages: art - target_languages: eng - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/art-eng/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['eo', 'io', 'art', 'en'] - src_constituents: {'sjn_Latn', 'tzl', 'vol_Latn', 'qya', 'tlh_Latn', 'ile_Latn', 'ido_Latn', 'tzl_Latn', 'jbo_Cyrl', 'jbo', 'lfn_Latn', 'nov_Latn', 'dws_Latn', 'ldn_Latn', 'avk_Latn', 'lfn_Cyrl', 'ina_Latn', 'jbo_Latn', 'epo', 'afh_Latn', 'qya_Latn', 'ido'} - tgt_constituents: {'eng'} - src_multilingual: True - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/art-eng/opus2m-2020-07-31.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/art-eng/opus2m-2020-07-31.test.txt - src_alpha3: art - tgt_alpha3: eng - short_pair: art-en - chrF2_score: 0.287 - bleu: 11.6 - brevity_penalty: 1.0 - ref_len: 73037.0 - src_name: Artificial languages - tgt_name: English - train_date: 2020-07-31 - src_alpha2: art - tgt_alpha2: en - prefer_old: False - long_pair: art-eng - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ar-ru
Helsinki-NLP
2023-08-16T11:25:45Z
204
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ar", "ru", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - ar - ru tags: - translation license: apache-2.0 --- ### ara-rus * source group: Arabic * target group: Russian * OPUS readme: [ara-rus](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-rus/README.md) * model: transformer * source language(s): apc ara arz * target language(s): rus * model: transformer * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-07-03.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-rus/opus-2020-07-03.zip) * test set translations: [opus-2020-07-03.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-rus/opus-2020-07-03.test.txt) * test set scores: [opus-2020-07-03.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-rus/opus-2020-07-03.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.ara.rus | 42.5 | 0.605 | ### System Info: - hf_name: ara-rus - source_languages: ara - target_languages: rus - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-rus/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ar', 'ru'] - src_constituents: {'apc', 'ara', 'arq_Latn', 'arq', 'afb', 'ara_Latn', 'apc_Latn', 'arz'} - tgt_constituents: {'rus'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-rus/opus-2020-07-03.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-rus/opus-2020-07-03.test.txt - src_alpha3: ara - tgt_alpha3: rus - short_pair: ar-ru - chrF2_score: 0.605 - bleu: 42.5 - brevity_penalty: 0.97 - ref_len: 21830.0 - src_name: Arabic - tgt_name: Russian - train_date: 2020-07-03 - src_alpha2: ar - tgt_alpha2: ru - prefer_old: False - long_pair: ara-rus - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ar-pl
Helsinki-NLP
2023-08-16T11:25:44Z
143
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ar", "pl", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - ar - pl tags: - translation license: apache-2.0 --- ### ara-pol * source group: Arabic * target group: Polish * OPUS readme: [ara-pol](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-pol/README.md) * model: transformer * source language(s): ara arz * target language(s): pol * model: transformer * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-07-03.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-pol/opus-2020-07-03.zip) * test set translations: [opus-2020-07-03.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-pol/opus-2020-07-03.test.txt) * test set scores: [opus-2020-07-03.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-pol/opus-2020-07-03.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.ara.pol | 38.0 | 0.623 | ### System Info: - hf_name: ara-pol - source_languages: ara - target_languages: pol - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-pol/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ar', 'pl'] - src_constituents: {'apc', 'ara', 'arq_Latn', 'arq', 'afb', 'ara_Latn', 'apc_Latn', 'arz'} - tgt_constituents: {'pol'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-pol/opus-2020-07-03.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-pol/opus-2020-07-03.test.txt - src_alpha3: ara - tgt_alpha3: pol - short_pair: ar-pl - chrF2_score: 0.623 - bleu: 38.0 - brevity_penalty: 0.948 - ref_len: 1171.0 - src_name: Arabic - tgt_name: Polish - train_date: 2020-07-03 - src_alpha2: ar - tgt_alpha2: pl - prefer_old: False - long_pair: ara-pol - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ar-it
Helsinki-NLP
2023-08-16T11:25:43Z
250
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ar", "it", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - ar - it tags: - translation license: apache-2.0 --- ### ara-ita * source group: Arabic * target group: Italian * OPUS readme: [ara-ita](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-ita/README.md) * model: transformer * source language(s): ara * target language(s): ita * model: transformer * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-07-03.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-ita/opus-2020-07-03.zip) * test set translations: [opus-2020-07-03.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-ita/opus-2020-07-03.test.txt) * test set scores: [opus-2020-07-03.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-ita/opus-2020-07-03.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.ara.ita | 44.2 | 0.658 | ### System Info: - hf_name: ara-ita - source_languages: ara - target_languages: ita - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-ita/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ar', 'it'] - src_constituents: {'apc', 'ara', 'arq_Latn', 'arq', 'afb', 'ara_Latn', 'apc_Latn', 'arz'} - tgt_constituents: {'ita'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-ita/opus-2020-07-03.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-ita/opus-2020-07-03.test.txt - src_alpha3: ara - tgt_alpha3: ita - short_pair: ar-it - chrF2_score: 0.6579999999999999 - bleu: 44.2 - brevity_penalty: 0.9890000000000001 - ref_len: 1495.0 - src_name: Arabic - tgt_name: Italian - train_date: 2020-07-03 - src_alpha2: ar - tgt_alpha2: it - prefer_old: False - long_pair: ara-ita - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Nextcloud-AI/opus-mt-ar-it
Nextcloud-AI
2023-08-16T11:25:43Z
111
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ar", "it", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2024-02-23T10:37:30Z
--- language: - ar - it tags: - translation license: apache-2.0 --- ### ara-ita * source group: Arabic * target group: Italian * OPUS readme: [ara-ita](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-ita/README.md) * model: transformer * source language(s): ara * target language(s): ita * model: transformer * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-07-03.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-ita/opus-2020-07-03.zip) * test set translations: [opus-2020-07-03.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-ita/opus-2020-07-03.test.txt) * test set scores: [opus-2020-07-03.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-ita/opus-2020-07-03.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.ara.ita | 44.2 | 0.658 | ### System Info: - hf_name: ara-ita - source_languages: ara - target_languages: ita - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-ita/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ar', 'it'] - src_constituents: {'apc', 'ara', 'arq_Latn', 'arq', 'afb', 'ara_Latn', 'apc_Latn', 'arz'} - tgt_constituents: {'ita'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-ita/opus-2020-07-03.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-ita/opus-2020-07-03.test.txt - src_alpha3: ara - tgt_alpha3: ita - short_pair: ar-it - chrF2_score: 0.6579999999999999 - bleu: 44.2 - brevity_penalty: 0.9890000000000001 - ref_len: 1495.0 - src_name: Arabic - tgt_name: Italian - train_date: 2020-07-03 - src_alpha2: ar - tgt_alpha2: it - prefer_old: False - long_pair: ara-ita - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Nextcloud-AI/opus-mt-ar-es
Nextcloud-AI
2023-08-16T11:25:40Z
113
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ar", "es", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2024-02-23T10:37:11Z
--- language: - ar - es tags: - translation license: apache-2.0 --- ### ara-spa * source group: Arabic * target group: Spanish * OPUS readme: [ara-spa](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-spa/README.md) * model: transformer * source language(s): apc apc_Latn ara arq * target language(s): spa * model: transformer * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-07-03.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-spa/opus-2020-07-03.zip) * test set translations: [opus-2020-07-03.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-spa/opus-2020-07-03.test.txt) * test set scores: [opus-2020-07-03.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-spa/opus-2020-07-03.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.ara.spa | 46.0 | 0.641 | ### System Info: - hf_name: ara-spa - source_languages: ara - target_languages: spa - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-spa/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ar', 'es'] - src_constituents: {'apc', 'ara', 'arq_Latn', 'arq', 'afb', 'ara_Latn', 'apc_Latn', 'arz'} - tgt_constituents: {'spa'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-spa/opus-2020-07-03.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-spa/opus-2020-07-03.test.txt - src_alpha3: ara - tgt_alpha3: spa - short_pair: ar-es - chrF2_score: 0.6409999999999999 - bleu: 46.0 - brevity_penalty: 0.9620000000000001 - ref_len: 9708.0 - src_name: Arabic - tgt_name: Spanish - train_date: 2020-07-03 - src_alpha2: ar - tgt_alpha2: es - prefer_old: False - long_pair: ara-spa - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ar-es
Helsinki-NLP
2023-08-16T11:25:40Z
331
1
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ar", "es", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - ar - es tags: - translation license: apache-2.0 --- ### ara-spa * source group: Arabic * target group: Spanish * OPUS readme: [ara-spa](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-spa/README.md) * model: transformer * source language(s): apc apc_Latn ara arq * target language(s): spa * model: transformer * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-07-03.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-spa/opus-2020-07-03.zip) * test set translations: [opus-2020-07-03.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-spa/opus-2020-07-03.test.txt) * test set scores: [opus-2020-07-03.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-spa/opus-2020-07-03.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.ara.spa | 46.0 | 0.641 | ### System Info: - hf_name: ara-spa - source_languages: ara - target_languages: spa - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-spa/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ar', 'es'] - src_constituents: {'apc', 'ara', 'arq_Latn', 'arq', 'afb', 'ara_Latn', 'apc_Latn', 'arz'} - tgt_constituents: {'spa'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-spa/opus-2020-07-03.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-spa/opus-2020-07-03.test.txt - src_alpha3: ara - tgt_alpha3: spa - short_pair: ar-es - chrF2_score: 0.6409999999999999 - bleu: 46.0 - brevity_penalty: 0.9620000000000001 - ref_len: 9708.0 - src_name: Arabic - tgt_name: Spanish - train_date: 2020-07-03 - src_alpha2: ar - tgt_alpha2: es - prefer_old: False - long_pair: ara-spa - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ar-eo
Helsinki-NLP
2023-08-16T11:25:37Z
132
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ar", "eo", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - ar - eo tags: - translation license: apache-2.0 --- ### ara-epo * source group: Arabic * target group: Esperanto * OPUS readme: [ara-epo](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-epo/README.md) * model: transformer-align * source language(s): apc apc_Latn ara arq arq_Latn arz * target language(s): epo * model: transformer-align * pre-processing: normalization + SentencePiece (spm4k,spm4k) * download original weights: [opus-2020-06-16.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-epo/opus-2020-06-16.zip) * test set translations: [opus-2020-06-16.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-epo/opus-2020-06-16.test.txt) * test set scores: [opus-2020-06-16.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-epo/opus-2020-06-16.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.ara.epo | 18.9 | 0.376 | ### System Info: - hf_name: ara-epo - source_languages: ara - target_languages: epo - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-epo/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ar', 'eo'] - src_constituents: {'apc', 'ara', 'arq_Latn', 'arq', 'afb', 'ara_Latn', 'apc_Latn', 'arz'} - tgt_constituents: {'epo'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm4k,spm4k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-epo/opus-2020-06-16.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-epo/opus-2020-06-16.test.txt - src_alpha3: ara - tgt_alpha3: epo - short_pair: ar-eo - chrF2_score: 0.376 - bleu: 18.9 - brevity_penalty: 0.948 - ref_len: 4506.0 - src_name: Arabic - tgt_name: Esperanto - train_date: 2020-06-16 - src_alpha2: ar - tgt_alpha2: eo - prefer_old: False - long_pair: ara-epo - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ar-en
Helsinki-NLP
2023-08-16T11:25:35Z
377,509
39
transformers
[ "transformers", "pytorch", "tf", "rust", "marian", "text2text-generation", "translation", "ar", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-ar-en * source languages: ar * target languages: en * OPUS readme: [ar-en](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/ar-en/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2019-12-18.zip](https://object.pouta.csc.fi/OPUS-MT-models/ar-en/opus-2019-12-18.zip) * test set translations: [opus-2019-12-18.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/ar-en/opus-2019-12-18.test.txt) * test set scores: [opus-2019-12-18.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/ar-en/opus-2019-12-18.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.ar.en | 49.4 | 0.661 |
Nextcloud-AI/opus-mt-ar-en
Nextcloud-AI
2023-08-16T11:25:35Z
103
1
transformers
[ "transformers", "pytorch", "tf", "rust", "marian", "text2text-generation", "translation", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2024-02-23T10:36:54Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-ar-en * source languages: ar * target languages: en * OPUS readme: [ar-en](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/ar-en/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2019-12-18.zip](https://object.pouta.csc.fi/OPUS-MT-models/ar-en/opus-2019-12-18.zip) * test set translations: [opus-2019-12-18.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/ar-en/opus-2019-12-18.test.txt) * test set scores: [opus-2019-12-18.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/ar-en/opus-2019-12-18.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.ar.en | 49.4 | 0.661 |
Nextcloud-AI/opus-mt-ar-de
Nextcloud-AI
2023-08-16T11:25:33Z
104
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ar", "de", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2024-02-23T10:36:36Z
--- language: - ar - de tags: - translation license: apache-2.0 --- ### ara-deu * source group: Arabic * target group: German * OPUS readme: [ara-deu](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-deu/README.md) * model: transformer-align * source language(s): afb apc ara ara_Latn arq arz * target language(s): deu * model: transformer-align * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-07-03.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-deu/opus-2020-07-03.zip) * test set translations: [opus-2020-07-03.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-deu/opus-2020-07-03.test.txt) * test set scores: [opus-2020-07-03.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-deu/opus-2020-07-03.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.ara.deu | 44.7 | 0.629 | ### System Info: - hf_name: ara-deu - source_languages: ara - target_languages: deu - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-deu/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ar', 'de'] - src_constituents: {'apc', 'ara', 'arq_Latn', 'arq', 'afb', 'ara_Latn', 'apc_Latn', 'arz'} - tgt_constituents: {'deu'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-deu/opus-2020-07-03.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-deu/opus-2020-07-03.test.txt - src_alpha3: ara - tgt_alpha3: deu - short_pair: ar-de - chrF2_score: 0.629 - bleu: 44.7 - brevity_penalty: 0.986 - ref_len: 8371.0 - src_name: Arabic - tgt_name: German - train_date: 2020-07-03 - src_alpha2: ar - tgt_alpha2: de - prefer_old: False - long_pair: ara-deu - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-ar-de
Helsinki-NLP
2023-08-16T11:25:33Z
757
1
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "ar", "de", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - ar - de tags: - translation license: apache-2.0 --- ### ara-deu * source group: Arabic * target group: German * OPUS readme: [ara-deu](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-deu/README.md) * model: transformer-align * source language(s): afb apc ara ara_Latn arq arz * target language(s): deu * model: transformer-align * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-07-03.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-deu/opus-2020-07-03.zip) * test set translations: [opus-2020-07-03.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-deu/opus-2020-07-03.test.txt) * test set scores: [opus-2020-07-03.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/ara-deu/opus-2020-07-03.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.ara.deu | 44.7 | 0.629 | ### System Info: - hf_name: ara-deu - source_languages: ara - target_languages: deu - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/ara-deu/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['ar', 'de'] - src_constituents: {'apc', 'ara', 'arq_Latn', 'arq', 'afb', 'ara_Latn', 'apc_Latn', 'arz'} - tgt_constituents: {'deu'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-deu/opus-2020-07-03.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/ara-deu/opus-2020-07-03.test.txt - src_alpha3: ara - tgt_alpha3: deu - short_pair: ar-de - chrF2_score: 0.629 - bleu: 44.7 - brevity_penalty: 0.986 - ref_len: 8371.0 - src_name: Arabic - tgt_name: German - train_date: 2020-07-03 - src_alpha2: ar - tgt_alpha2: de - prefer_old: False - long_pair: ara-deu - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-am-sv
Helsinki-NLP
2023-08-16T11:25:32Z
122
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "am", "sv", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-am-sv * source languages: am * target languages: sv * OPUS readme: [am-sv](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/am-sv/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/am-sv/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/am-sv/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/am-sv/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.am.sv | 21.0 | 0.377 |
Helsinki-NLP/opus-mt-af-nl
Helsinki-NLP
2023-08-16T11:25:25Z
144
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "af", "nl", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - af - nl tags: - translation license: apache-2.0 --- ### afr-nld * source group: Afrikaans * target group: Dutch * OPUS readme: [afr-nld](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/afr-nld/README.md) * model: transformer-align * source language(s): afr * target language(s): nld * model: transformer-align * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-06-17.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/afr-nld/opus-2020-06-17.zip) * test set translations: [opus-2020-06-17.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/afr-nld/opus-2020-06-17.test.txt) * test set scores: [opus-2020-06-17.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/afr-nld/opus-2020-06-17.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.afr.nld | 55.2 | 0.715 | ### System Info: - hf_name: afr-nld - source_languages: afr - target_languages: nld - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/afr-nld/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['af', 'nl'] - src_constituents: {'afr'} - tgt_constituents: {'nld'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/afr-nld/opus-2020-06-17.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/afr-nld/opus-2020-06-17.test.txt - src_alpha3: afr - tgt_alpha3: nld - short_pair: af-nl - chrF2_score: 0.715 - bleu: 55.2 - brevity_penalty: 0.995 - ref_len: 6710.0 - src_name: Afrikaans - tgt_name: Dutch - train_date: 2020-06-17 - src_alpha2: af - tgt_alpha2: nl - prefer_old: False - long_pair: afr-nld - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-af-fr
Helsinki-NLP
2023-08-16T11:25:24Z
129
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "af", "fr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-af-fr * source languages: af * target languages: fr * OPUS readme: [af-fr](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/af-fr/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-01-08.zip](https://object.pouta.csc.fi/OPUS-MT-models/af-fr/opus-2020-01-08.zip) * test set translations: [opus-2020-01-08.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/af-fr/opus-2020-01-08.test.txt) * test set scores: [opus-2020-01-08.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/af-fr/opus-2020-01-08.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | JW300.af.fr | 35.3 | 0.543 |
Helsinki-NLP/opus-mt-af-es
Helsinki-NLP
2023-08-16T11:25:22Z
123
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "af", "es", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- language: - af - es tags: - translation license: apache-2.0 --- ### afr-spa * source group: Afrikaans * target group: Spanish * OPUS readme: [afr-spa](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/afr-spa/README.md) * model: transformer-align * source language(s): afr * target language(s): spa * model: transformer-align * pre-processing: normalization + SentencePiece (spm32k,spm32k) * download original weights: [opus-2020-06-17.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/afr-spa/opus-2020-06-17.zip) * test set translations: [opus-2020-06-17.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/afr-spa/opus-2020-06-17.test.txt) * test set scores: [opus-2020-06-17.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/afr-spa/opus-2020-06-17.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba-test.afr.spa | 49.9 | 0.680 | ### System Info: - hf_name: afr-spa - source_languages: afr - target_languages: spa - opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/afr-spa/README.md - original_repo: Tatoeba-Challenge - tags: ['translation'] - languages: ['af', 'es'] - src_constituents: {'afr'} - tgt_constituents: {'spa'} - src_multilingual: False - tgt_multilingual: False - prepro: normalization + SentencePiece (spm32k,spm32k) - url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/afr-spa/opus-2020-06-17.zip - url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/afr-spa/opus-2020-06-17.test.txt - src_alpha3: afr - tgt_alpha3: spa - short_pair: af-es - chrF2_score: 0.68 - bleu: 49.9 - brevity_penalty: 1.0 - ref_len: 2783.0 - src_name: Afrikaans - tgt_name: Spanish - train_date: 2020-06-17 - src_alpha2: af - tgt_alpha2: es - prefer_old: False - long_pair: afr-spa - helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535 - transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b - port_machine: brutasse - port_time: 2020-08-21-14:41
Helsinki-NLP/opus-mt-af-en
Helsinki-NLP
2023-08-16T11:25:20Z
4,596
0
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "af", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-af-en * source languages: af * target languages: en * OPUS readme: [af-en](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/af-en/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * download original weights: [opus-2019-12-18.zip](https://object.pouta.csc.fi/OPUS-MT-models/af-en/opus-2019-12-18.zip) * test set translations: [opus-2019-12-18.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/af-en/opus-2019-12-18.test.txt) * test set scores: [opus-2019-12-18.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/af-en/opus-2019-12-18.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.af.en | 60.8 | 0.736 |
Helsinki-NLP/opus-mt-SCANDINAVIA-SCANDINAVIA
Helsinki-NLP
2023-08-16T11:25:15Z
180
1
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "scandinavia", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-SCANDINAVIA-SCANDINAVIA * source languages: da,fo,is,no,nb,nn,sv * target languages: da,fo,is,no,nb,nn,sv * OPUS readme: [da+fo+is+no+nb+nn+sv-da+fo+is+no+nb+nn+sv](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/da+fo+is+no+nb+nn+sv-da+fo+is+no+nb+nn+sv/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * a sentence initial language token is required in the form of `>>id<<` (id = valid target language ID) * download original weights: [opus-2019-12-18.zip](https://object.pouta.csc.fi/OPUS-MT-models/da+fo+is+no+nb+nn+sv-da+fo+is+no+nb+nn+sv/opus-2019-12-18.zip) * test set translations: [opus-2019-12-18.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/da+fo+is+no+nb+nn+sv-da+fo+is+no+nb+nn+sv/opus-2019-12-18.test.txt) * test set scores: [opus-2019-12-18.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/da+fo+is+no+nb+nn+sv-da+fo+is+no+nb+nn+sv/opus-2019-12-18.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.da.sv | 69.2 | 0.811 |
Helsinki-NLP/opus-mt-ROMANCE-en
Helsinki-NLP
2023-08-16T11:25:14Z
88,075
8
transformers
[ "transformers", "pytorch", "tf", "rust", "marian", "text2text-generation", "translation", "roa", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-ROMANCE-en * source languages: fr,fr_BE,fr_CA,fr_FR,wa,frp,oc,ca,rm,lld,fur,lij,lmo,es,es_AR,es_CL,es_CO,es_CR,es_DO,es_EC,es_ES,es_GT,es_HN,es_MX,es_NI,es_PA,es_PE,es_PR,es_SV,es_UY,es_VE,pt,pt_br,pt_BR,pt_PT,gl,lad,an,mwl,it,it_IT,co,nap,scn,vec,sc,ro,la * target languages: en * OPUS readme: [fr+fr_BE+fr_CA+fr_FR+wa+frp+oc+ca+rm+lld+fur+lij+lmo+es+es_AR+es_CL+es_CO+es_CR+es_DO+es_EC+es_ES+es_GT+es_HN+es_MX+es_NI+es_PA+es_PE+es_PR+es_SV+es_UY+es_VE+pt+pt_br+pt_BR+pt_PT+gl+lad+an+mwl+it+it_IT+co+nap+scn+vec+sc+ro+la-en](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/fr+fr_BE+fr_CA+fr_FR+wa+frp+oc+ca+rm+lld+fur+lij+lmo+es+es_AR+es_CL+es_CO+es_CR+es_DO+es_EC+es_ES+es_GT+es_HN+es_MX+es_NI+es_PA+es_PE+es_PR+es_SV+es_UY+es_VE+pt+pt_br+pt_BR+pt_PT+gl+lad+an+mwl+it+it_IT+co+nap+scn+vec+sc+ro+la-en/README.md) * dataset: opus * model: transformer * pre-processing: normalization + SentencePiece * download original weights: [opus-2020-04-01.zip](https://object.pouta.csc.fi/OPUS-MT-models/fr+fr_BE+fr_CA+fr_FR+wa+frp+oc+ca+rm+lld+fur+lij+lmo+es+es_AR+es_CL+es_CO+es_CR+es_DO+es_EC+es_ES+es_GT+es_HN+es_MX+es_NI+es_PA+es_PE+es_PR+es_SV+es_UY+es_VE+pt+pt_br+pt_BR+pt_PT+gl+lad+an+mwl+it+it_IT+co+nap+scn+vec+sc+ro+la-en/opus-2020-04-01.zip) * test set translations: [opus-2020-04-01.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/fr+fr_BE+fr_CA+fr_FR+wa+frp+oc+ca+rm+lld+fur+lij+lmo+es+es_AR+es_CL+es_CO+es_CR+es_DO+es_EC+es_ES+es_GT+es_HN+es_MX+es_NI+es_PA+es_PE+es_PR+es_SV+es_UY+es_VE+pt+pt_br+pt_BR+pt_PT+gl+lad+an+mwl+it+it_IT+co+nap+scn+vec+sc+ro+la-en/opus-2020-04-01.test.txt) * test set scores: [opus-2020-04-01.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/fr+fr_BE+fr_CA+fr_FR+wa+frp+oc+ca+rm+lld+fur+lij+lmo+es+es_AR+es_CL+es_CO+es_CR+es_DO+es_EC+es_ES+es_GT+es_HN+es_MX+es_NI+es_PA+es_PE+es_PR+es_SV+es_UY+es_VE+pt+pt_br+pt_BR+pt_PT+gl+lad+an+mwl+it+it_IT+co+nap+scn+vec+sc+ro+la-en/opus-2020-04-01.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.fr.en | 62.2 | 0.750 |
CyberHarem/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon
CyberHarem
2023-08-16T11:23:05Z
0
0
null
[ "art", "text-to-image", "dataset:CyberHarem/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon", "license:mit", "region:us" ]
text-to-image
2023-08-16T11:18:19Z
--- license: mit datasets: - CyberHarem/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon pipeline_tag: text-to-image tags: - art --- # Lora of liliruca_arde_isitwrongtotrytopickupgirlsinadungeon This model is trained with [HCP-Diffusion](https://github.com/7eu7d7/HCP-Diffusion). And the auto-training framework is maintained by [DeepGHS Team](https://huggingface.co/deepghs). After downloading the pt and safetensors files for the specified step, you need to use them simultaneously. The pt file will be used as an embedding, while the safetensors file will be loaded for Lora. For example, if you want to use the model from step 1500, you need to download `1500/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.pt` as the embedding and `1500/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.safetensors` for loading Lora. By using both files together, you can generate images for the desired characters. **The trigger word is `liliruca_arde_isitwrongtotrytopickupgirlsinadungeon`.** These are available steps: | Steps | pattern_1 | pattern_2 | pattern_3 | pattern_4 | bikini | free | nude | Download | |--------:|:-----------------------------------------------|:-----------------------------------------------|:-----------------------------------------------|:----------------------------------------------------|:-----------------------------------------|:-------------------------------------|:-----------------------------------------------|:-------------------------------------------------------------------------| | 1500 | ![pattern_1-1500](1500/previews/pattern_1.png) | ![pattern_2-1500](1500/previews/pattern_2.png) | ![pattern_3-1500](1500/previews/pattern_3.png) | [<NSFW, click to see>](1500/previews/pattern_4.png) | ![bikini-1500](1500/previews/bikini.png) | ![free-1500](1500/previews/free.png) | [<NSFW, click to see>](1500/previews/nude.png) | [Download](1500/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 1400 | ![pattern_1-1400](1400/previews/pattern_1.png) | ![pattern_2-1400](1400/previews/pattern_2.png) | ![pattern_3-1400](1400/previews/pattern_3.png) | [<NSFW, click to see>](1400/previews/pattern_4.png) | ![bikini-1400](1400/previews/bikini.png) | ![free-1400](1400/previews/free.png) | [<NSFW, click to see>](1400/previews/nude.png) | [Download](1400/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 1300 | ![pattern_1-1300](1300/previews/pattern_1.png) | ![pattern_2-1300](1300/previews/pattern_2.png) | ![pattern_3-1300](1300/previews/pattern_3.png) | [<NSFW, click to see>](1300/previews/pattern_4.png) | ![bikini-1300](1300/previews/bikini.png) | ![free-1300](1300/previews/free.png) | [<NSFW, click to see>](1300/previews/nude.png) | [Download](1300/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 1200 | ![pattern_1-1200](1200/previews/pattern_1.png) | ![pattern_2-1200](1200/previews/pattern_2.png) | ![pattern_3-1200](1200/previews/pattern_3.png) | [<NSFW, click to see>](1200/previews/pattern_4.png) | ![bikini-1200](1200/previews/bikini.png) | ![free-1200](1200/previews/free.png) | [<NSFW, click to see>](1200/previews/nude.png) | [Download](1200/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 1100 | ![pattern_1-1100](1100/previews/pattern_1.png) | ![pattern_2-1100](1100/previews/pattern_2.png) | ![pattern_3-1100](1100/previews/pattern_3.png) | [<NSFW, click to see>](1100/previews/pattern_4.png) | ![bikini-1100](1100/previews/bikini.png) | ![free-1100](1100/previews/free.png) | [<NSFW, click to see>](1100/previews/nude.png) | [Download](1100/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 1000 | ![pattern_1-1000](1000/previews/pattern_1.png) | ![pattern_2-1000](1000/previews/pattern_2.png) | ![pattern_3-1000](1000/previews/pattern_3.png) | [<NSFW, click to see>](1000/previews/pattern_4.png) | ![bikini-1000](1000/previews/bikini.png) | ![free-1000](1000/previews/free.png) | [<NSFW, click to see>](1000/previews/nude.png) | [Download](1000/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 900 | ![pattern_1-900](900/previews/pattern_1.png) | ![pattern_2-900](900/previews/pattern_2.png) | ![pattern_3-900](900/previews/pattern_3.png) | [<NSFW, click to see>](900/previews/pattern_4.png) | ![bikini-900](900/previews/bikini.png) | ![free-900](900/previews/free.png) | [<NSFW, click to see>](900/previews/nude.png) | [Download](900/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 800 | ![pattern_1-800](800/previews/pattern_1.png) | ![pattern_2-800](800/previews/pattern_2.png) | ![pattern_3-800](800/previews/pattern_3.png) | [<NSFW, click to see>](800/previews/pattern_4.png) | ![bikini-800](800/previews/bikini.png) | ![free-800](800/previews/free.png) | [<NSFW, click to see>](800/previews/nude.png) | [Download](800/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 700 | ![pattern_1-700](700/previews/pattern_1.png) | ![pattern_2-700](700/previews/pattern_2.png) | ![pattern_3-700](700/previews/pattern_3.png) | [<NSFW, click to see>](700/previews/pattern_4.png) | ![bikini-700](700/previews/bikini.png) | ![free-700](700/previews/free.png) | [<NSFW, click to see>](700/previews/nude.png) | [Download](700/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 600 | ![pattern_1-600](600/previews/pattern_1.png) | ![pattern_2-600](600/previews/pattern_2.png) | ![pattern_3-600](600/previews/pattern_3.png) | [<NSFW, click to see>](600/previews/pattern_4.png) | ![bikini-600](600/previews/bikini.png) | ![free-600](600/previews/free.png) | [<NSFW, click to see>](600/previews/nude.png) | [Download](600/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 500 | ![pattern_1-500](500/previews/pattern_1.png) | ![pattern_2-500](500/previews/pattern_2.png) | ![pattern_3-500](500/previews/pattern_3.png) | [<NSFW, click to see>](500/previews/pattern_4.png) | ![bikini-500](500/previews/bikini.png) | ![free-500](500/previews/free.png) | [<NSFW, click to see>](500/previews/nude.png) | [Download](500/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 400 | ![pattern_1-400](400/previews/pattern_1.png) | ![pattern_2-400](400/previews/pattern_2.png) | ![pattern_3-400](400/previews/pattern_3.png) | [<NSFW, click to see>](400/previews/pattern_4.png) | ![bikini-400](400/previews/bikini.png) | ![free-400](400/previews/free.png) | [<NSFW, click to see>](400/previews/nude.png) | [Download](400/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 300 | ![pattern_1-300](300/previews/pattern_1.png) | ![pattern_2-300](300/previews/pattern_2.png) | ![pattern_3-300](300/previews/pattern_3.png) | [<NSFW, click to see>](300/previews/pattern_4.png) | ![bikini-300](300/previews/bikini.png) | ![free-300](300/previews/free.png) | [<NSFW, click to see>](300/previews/nude.png) | [Download](300/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 200 | ![pattern_1-200](200/previews/pattern_1.png) | ![pattern_2-200](200/previews/pattern_2.png) | ![pattern_3-200](200/previews/pattern_3.png) | [<NSFW, click to see>](200/previews/pattern_4.png) | ![bikini-200](200/previews/bikini.png) | ![free-200](200/previews/free.png) | [<NSFW, click to see>](200/previews/nude.png) | [Download](200/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) | | 100 | ![pattern_1-100](100/previews/pattern_1.png) | ![pattern_2-100](100/previews/pattern_2.png) | ![pattern_3-100](100/previews/pattern_3.png) | [<NSFW, click to see>](100/previews/pattern_4.png) | ![bikini-100](100/previews/bikini.png) | ![free-100](100/previews/free.png) | [<NSFW, click to see>](100/previews/nude.png) | [Download](100/liliruca_arde_isitwrongtotrytopickupgirlsinadungeon.zip) |
Helsinki-NLP/opus-mt-NORTH_EU-NORTH_EU
Helsinki-NLP
2023-08-16T11:22:36Z
158
2
transformers
[ "transformers", "pytorch", "tf", "marian", "text2text-generation", "translation", "multilingual", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2022-03-02T23:29:04Z
--- tags: - translation license: apache-2.0 --- ### opus-mt-NORTH_EU-NORTH_EU * source languages: de,nl,fy,af,da,fo,is,no,nb,nn,sv * target languages: de,nl,fy,af,da,fo,is,no,nb,nn,sv * OPUS readme: [de+nl+fy+af+da+fo+is+no+nb+nn+sv-de+nl+fy+af+da+fo+is+no+nb+nn+sv](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/de+nl+fy+af+da+fo+is+no+nb+nn+sv-de+nl+fy+af+da+fo+is+no+nb+nn+sv/README.md) * dataset: opus * model: transformer-align * pre-processing: normalization + SentencePiece * a sentence initial language token is required in the form of `>>id<<` (id = valid target language ID) * download original weights: [opus-2020-01-15.zip](https://object.pouta.csc.fi/OPUS-MT-models/de+nl+fy+af+da+fo+is+no+nb+nn+sv-de+nl+fy+af+da+fo+is+no+nb+nn+sv/opus-2020-01-15.zip) * test set translations: [opus-2020-01-15.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/de+nl+fy+af+da+fo+is+no+nb+nn+sv-de+nl+fy+af+da+fo+is+no+nb+nn+sv/opus-2020-01-15.test.txt) * test set scores: [opus-2020-01-15.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/de+nl+fy+af+da+fo+is+no+nb+nn+sv-de+nl+fy+af+da+fo+is+no+nb+nn+sv/opus-2020-01-15.eval.txt) ## Benchmarks | testset | BLEU | chr-F | |-----------------------|-------|-------| | Tatoeba.de.sv | 48.1 | 0.663 |
davidkim205/komt-Llama-2-13b-hf-lora
davidkim205
2023-08-16T11:17:31Z
5
0
peft
[ "peft", "facebook", "meta", "pytorch", "llama", "llama-2", "llama-2-chat", "text-generation", "en", "ko", "license:apache-2.0", "region:us" ]
text-generation
2023-08-16T10:28:21Z
--- language: - en - ko pipeline_tag: text-generation inference: false tags: - facebook - meta - pytorch - llama - llama-2 - llama-2-chat license: apache-2.0 library_name: peft --- # komt-Llama-2-13b-hf-lora This model fine-tuned the aaa model using PEFT-LoRA. The "komt-Llama-2-13b-hf-lora" model was developed using a multi-task instruction technique aimed at enhancing Korean language performance. For more details, please refer to the GitHub Repository. Please refer below for more detailed information. For more detailed information, please refer to the https://huggingface.co/davidkim205/komt-Llama-2-13b-hf. ## Model Details * **Model Developers** : davidkim(changyeon kim) * **Repository** : https://github.com/davidkim205/komt * **Lora target modules** : q_proj, o_proj, v_proj, gate_proj, down_proj, k_proj, up_proj * **Model Size** : 120MB * **Model Architecture** : komt-Llama-2-13b is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning by multi-task instruction ## Dataset korean multi-task instruction dataset ## Prompt Template ``` ### instruction: {prompt} ### Response: ``` Examples: ``` ### instruction: 영화 해리포터 시리즈 배급사가 어디야 ### Response: ``` response: ``` ### instruction: 영화 해리포터 시리즈 배급사가 어디야 ### Response: 워너 브라더스\n그 외에도 키네스 같은 작은 역할을 맡은 사람들도 있었다. 데이비드 알바레즈는 볼드모트 군단의 보로미르 7개 학교 중 하나인 필립스부르그 보로미르 7개 학교의 교장이며, 필리우스 플리트우드는 필립스부르그 보로미르 7개 학교의 교사이다. 필립스부르그는 보로미르 7개 학교 중 하나로, 보로미르 전설에 따르면 보로미르 7개 학교의 교장들이 죽으면 세계를 멸망시킨다는 점을 알고 있다. 《해리 포터와 혼혈 왕자》에서 데스 이터는 해리 포터를 죽이려고 하지만 실패하고 해리 포터는 데스 이터를 죽이려고 하지만 또 실패한다. 《해리 포터와 아즈카반의 죄수》에서 데스 이터는 아즈카반의 죄수로 등장하여 해리 포터에게 복수를 하고자 하지만 해리 포터와 론 위즐리의 발 빠른 대처로 실패하고 해리 포터와 론 위즐리는 데스 이터를 죽이려고 하지만 실패하고 그 결과 데스 이터는 다시 기저승에 봉인된다. 《해리 포터와 죽음의 성물》에서 데스 이터는 부활하여 죽음의 성물 1부에 등장하였으며, 해리 포터와 죽음의 성물 2부에서 데스 이터는 죽음의 성물 2부의 내용을 방해하는 간첩 역할을 한다. 데스 이터는 영화 해리포터와 죽음의 성물 1부에서 데스 이터의 역할을 맡은 배우 스티븐 피카드가 연기한다. ``` ## Usage After downloading from GitHub, please install as follows: ``` git clone https://github.com/davidkim205/komt cd komt pip install -r lora/requirements_lora.txt ``` * Requirements Python >=3.8. Linux distribution (Ubuntu, MacOS, etc.) + CUDA > 10.0. Refer https://github.com/TimDettmers/bitsandbytes#tldr ``` import torch from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig from transformers import StoppingCriteria, StoppingCriteriaList from transformers import TextStreamer, GenerationConfig from peft import PeftModel, PeftConfig class LocalStoppingCriteria(StoppingCriteria): def __init__(self, tokenizer, stop_words = []): super().__init__() stops = [tokenizer(stop_word, return_tensors='pt', add_special_tokens = False)['input_ids'].squeeze() for stop_word in stop_words] print('stop_words', stop_words) print('stop_words_ids', stops) self.stop_words = stop_words self.stops = [stop.cuda() for stop in stops] self.tokenizer = tokenizer def _compare_token(self, input_ids): for stop in self.stops: if len(stop.size()) != 1: continue stop_len = len(stop) if torch.all((stop == input_ids[0][-stop_len:])).item(): return True return False def _compare_decode(self, input_ids): input_str = self.tokenizer.decode(input_ids[0]) for stop_word in self.stop_words: if input_str.endswith(stop_word): return True return False def __call__(self, input_ids: torch.LongTensor, scores: torch.FloatTensor): input_str = self.tokenizer.decode(input_ids[0]) for stop_word in self.stop_words: if input_str.endswith(stop_word): return True return False # # config peft_model_name = 'davidkim205/komt-Llama-2-7b-chat-hf-lora' model_name = 'davidkim205/komt-Llama-2-7b-chat-hf' instruction_prefix = "### instruction: " input_prefix = "### input: " answer_prefix = "### Response: " endoftext = "<|end|>" stop_words = [endoftext, '<s>', '###'] generation_config = GenerationConfig( temperature=0.9, top_p=0.7, top_k=100, max_new_tokens=2048, early_stopping=True, do_sample=True, ) # # create model config = PeftConfig.from_pretrained(peft_model_name) bnb_config = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type="nf4", bnb_4bit_compute_dtype=torch.bfloat16 ) model = AutoModelForCausalLM.from_pretrained(model_name, quantization_config=bnb_config, device_map="auto") model = PeftModel.from_pretrained(model, peft_model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) stopping_criteria = StoppingCriteriaList([LocalStoppingCriteria(tokenizer=tokenizer, stop_words=stop_words)]) streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True) model.eval() # # generate prompt = f"### instruction: 영화 해리포터 시리즈 배급사가 어디야\n\n### Response:" gened = model.generate( **tokenizer( prompt, return_tensors='pt', return_token_type_ids=False ).to('cuda'), generation_config=generation_config, eos_token_id=model.config.eos_token_id, stopping_criteria=stopping_criteria, streamer=streamer ) output_text = tokenizer.decode(gened[0], skip_special_tokens=True) print('--------------------') print(output_text) ``` response: ``` 워너 브라더스\n그 외에도 키네스 같은 작은 역할을 맡은 사람들도 있었다. 데이비드 알바레즈는 볼드모트 군단의 보로미르 7개 학교 중 하나인 필립스부르그 보로미르 7개 학교의 교장이며, 필리우스 플리트우드는 필립스부르그 보로미르 7개 학교의 교사이다. 필립스부르그는 보로미르 7개 학교 중 하나로, 보로미르 전설에 따르면 보로미르 7개 학교의 교장들이 죽으면 세계를 멸망시킨다는 점을 알고 있다. 《해리 포터와 혼혈 왕자》에서 데스 이터는 해리 포터를 죽이려고 하지만 실패하고 해리 포터는 데스 이터를 죽이려고 하지만 또 실패한다. 《해리 포터와 아즈카반의 죄수》에서 데스 이터는 아즈카반의 죄수로 등장하여 해리 포터에게 복수를 하고자 하지만 해리 포터와 론 위즐리의 발 빠른 대처로 실패하고 해리 포터와 론 위즐리는 데스 이터를 죽이려고 하지만 실패하고 그 결과 데스 이터는 다시 기저승에 봉인된다. 《해리 포터와 죽음의 성물》에서 데스 이터는 부활하여 죽음의 성물 1부에 등장하였으며, 해리 포터와 죽음의 성물 2부에서 데스 이터는 죽음의 성물 2부의 내용을 방해하는 간첩 역할을 한다. 데스 이터는 영화 해리포터와 죽음의 성물 1부에서 데스 이터의 역할을 맡은 배우 스티븐 피카드가 연기한다. ``` ## Hardware and Software - nvidia driver : 535.54.03 - CUDA Version: 12.2 - ## Training procedure The following `bitsandbytes` quantization config was used during training: - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: True - bnb_4bit_compute_dtype: bfloat16 ### Framework versions - PEFT 0.5.0.dev0
C-guy/Lora-V
C-guy
2023-08-16T10:53:59Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-07-07T17:02:06Z
--- license: creativeml-openrail-m ---
mindreader/llama-recipe-7b-1epoch-8batch
mindreader
2023-08-16T10:52:06Z
4
0
peft
[ "peft", "pytorch", "llama", "region:us" ]
null
2023-08-15T06:37:14Z
--- library_name: peft --- !python llama-recipes/llama_finetuning.py \ --use_peft \ --num_epochs 1 \ --peft_method lora \ --run_validation false \ --quantization \ --dataset alpaca_dataset \ --model_name meta-llama/Llama-2-7b-chat-hf \ --save_model \ --save_optimizer \ --batch_size_training 8 \ --output_dir ./save ## Training procedure The following `bitsandbytes` quantization config was used during training: - load_in_8bit: True - load_in_4bit: False - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: fp4 - bnb_4bit_use_double_quant: False - bnb_4bit_compute_dtype: float32 ### Framework versions - PEFT 0.5.0.dev0
AptaArkana/indonesian_toxic_classification
AptaArkana
2023-08-16T10:20:10Z
103
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-08-16T07:58:01Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy - precision - recall - f1 model-index: - name: kata_kasar_test results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # kata_kasar_test This model is a fine-tuned version of [indolem/indobertweet-base-uncased](https://huggingface.co/indolem/indobertweet-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0099 - Accuracy: 0.9963 - Precision: 0.9926 - Recall: 1.0 - F1: 0.9963 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:| | 0.053 | 1.0 | 547 | 0.0215 | 0.9963 | 0.9944 | 0.9981 | 0.9963 | | 0.0043 | 2.0 | 1094 | 0.0099 | 0.9963 | 0.9926 | 1.0 | 0.9963 | ### Framework versions - Transformers 4.26.1 - Pytorch 1.13.1+cpu - Datasets 2.11.0 - Tokenizers 0.13.2
DrasticActions/gok-so-vits-svc-models
DrasticActions
2023-08-16T10:11:14Z
0
0
null
[ "license:cc", "region:us" ]
null
2023-08-15T12:54:36Z
--- license: cc --- # GOK so-vits-svc models ## How to use - Install https://github.com/voicepaw/so-vits-svc-fork - [Download the models](https://huggingface.co/DrasticActions/gok-so-vits-svc-models/tree/main/Models) from this repo - Open the svc GUI - Under Paths, set the model path to the specific "G_*.pth" file you want to use. - Set the config path to the config.json from the same model path folder. - The Input Audio should only be a (single) speaker's voice. For that, you can use https://ultimatevocalremover.com/ - To create the file, click "Infer"
FedJudge/fedjudge-base-7b
FedJudge
2023-08-16T09:52:52Z
0
1
null
[ "license:apache-2.0", "region:us" ]
null
2023-08-16T08:03:12Z
--- license: apache-2.0 --- # FedJudge: Federated Legal Large Language Model 近期Large Language Model(LLM)卓越表现推动了其在智慧司法领域的应用。目前已经涌现出许多优秀的法律大模型(Legal LLM),如<a href="https://github.com/AndrewZhe/lawyer-llama" target="_blank">Lawyer LLaMA</a>和<a href="https://github.com/PKU-YuanGroup/ChatLaw" target="_blank">ChatLaw</a>等。 然而,尽管Legal LLM在集中式数据训练的环境中取得了显著的成果,却少有人深入探究其在联邦学习场景下的应用。 在法律领域,联邦学习为Legal LLM的应用带来了一系列潜在的好处和机遇。首先,法律数据的隐私性是一个至关重要的问题。大量的法律数据分布在法院、检察院、咨询公司和法律教育培训机构等组织中,这些数据包含了个人的敏感信息。通过采用联邦学习,Legal LLM可以在本地设备上进行训练,在中心服务器进行参数的聚合和分发,避免了原始数据的共享,有效保护了用户的隐私。 此外,法律领域中存在着数据稀缺的情况。某些特定领域的法律数据可能非常有限,例如针对特定罕见案例的判例数据或特定地区的法律实践数据。在传统的集中式学习中,这些数据可能无法充分利用。而采用联邦学习,可以在本地设备上进行模型训练,利用分散的数据资源,从而提高模型的性能和泛化能力。 为此,我们提出了**FedJudge**,一个联邦法律大模型(Federated Legal Large Language Model)。 ## 训练数据 在本项目中,由于真实场景下的法律数据难以获取,我们对公开的法律语料进行了分类和汇总,并得到了以下三个本地客户端的数据集来模拟联邦场景: - 法院客户端:收集了由<a href="https://github.com/yuelinan/C3VG" target="_blank">C3VG</a>公开的裁判文书数据集,作为法院客户端的本地数据 - 法律咨询公司客户端:收集了由<a href="https://github.com/AndrewZhe/lawyer-llama" target="_blank">Lawyer LLaMA</a>公开的法律咨询数据集,作为法律咨询公司客户端的本地数据 - 法律教培机构客户端:收集了由<a href="https://github.com/AndrewZhe/lawyer-llama" target="_blank">Lawyer LLaMA</a>公开的法考数据集,作为法律教培机构客户端的本地数据 ## 模型训练 我们选取<a href="https://github.com/baichuan-inc/Baichuan-7B" target="_blank">baichuan-7b</a>作为FedJudge的基座模型。考虑到LLM在通信过程中的cost,我们使用baichuan-7b + lora的训练方法,在通信过程中,仅传输lora的参数。 ## 效果展示 <details><summary>模型效果比较</summary> | Question | Baichuan-7B | Center | FedJudge | | ------------------------------------------------------------ | ------------------------------------------------------------ | ------------------------------------------------------------ | ------------------------------------------------------------ | |假设你是一名律师,请回答以下向你咨询的问题:在法律中定金与订金的区别是什么|1、首先我们先来了解一下什么是“定金”。根据《中华人民共和国民法典》第五百八十六条规定:“当事人可以约定一方向对方给付定金作为债权的担保。定金合同自实际交付定金时成立。债务人履行债务后,定金应当抵作价款或者收回。给付定金的一方不履行约定的债务的,无权要求返还定金;收受定金的一方不履行约定的债务的,应当双倍返还定金。”2、那么什么又是“订金”呢?根据《最高人民法院关于适用<中华人民共和国民事诉讼法>若干问题的解释》第一百二十一条的规定,“当事人申请鉴定,可以在举证期限届满前提出。申请鉴定的事项与待证事实无关联,或者对证明待证事实无意义的,人民法院不予准许。”3、所以如果要问两者有什么不同的话,答案就是:没有!|在法律中,订金和定金是两种不同的概念。订金是指预先支付的一笔款项,通常是为了保证合同的履行而签订的一种担保形式。而定金则是一种具有法律效力的保证金,是在合同履行过程中作为违约责任的一种承担方式。因此,如果一方违反了合同约定,另一方可以要求其返还已经交付的定金,而不能要求对方退还订金。|首先,这两个词语都是指合同中的预付款。但是它们之间的区别在于:1.性质不同:订金是双方约定的一种支付方式;而定金是一种担保形式,具有一定的法律效力。2.适用范围不同:订金一般适用于小额交易或短期租赁等情形;而定金则适用于大额交易、长期租赁等情况。3.违约责任不同:订金不具备强制性,如果一方违约,另一方可以要求其退还订金;而定金一旦交付给对方,就视为对对方的债权债务关系的认可,除非有法定事由或者当事人协商一致,否则不得随意解除。4.返还条件不同:订金没有明确的规定,只要符合相关规定即可全额返还;而定金只有在出现法定事由时才能予以扣除,剩余部分应当返还。5.其他方面不同:订金通常不会产生利息费用,但定金可能会产生相应的利息费用。因此,如果你需要使用定金作为一种担保手段来保障自己的权益,那么一定要了解清楚相关的法律法规以及具体的操作流程,以免造成损失。| |假设你是一名律师,请回答以下向你咨询的问题:盗窃罪的犯罪客体是什么,盗窃罪的犯罪主体|1、盗窃罪侵犯的是公私财物的所有权。2、盗窃罪的犯罪主体是达到刑事责任年龄且具备刑事责任能力的自然人或单位。|盗窃罪的犯罪客体是财产。即通过非法手段占有他人财物的行为。而盗窃罪的犯罪主体是指实施盗窃行为的人。他们可能是个人、团伙或者组织等。在我国,盗窃罪是一种侵犯财产罪,属于刑法中的轻罪之一。|盗窃罪是指以非法占有为目的,秘密窃取公私财物的行为。其侵犯的客体是财产所有权和国家对社会经济活动的管理秩序。因此,盗窃罪的犯罪客体就是指被盗窃的财产的所有权以及国家的正常经济活动和社会生活秩序。\n\n盗窃罪的犯罪主体是一般主体,即年满16周岁以上、具有刑事责任能力的自然人都可以成为盗窃罪的犯罪主体。| |假设你是一名律师,请回答以下向你咨询的问题:结婚二+多年没在一起算是自动离婚不|1、如果双方都同意离婚的话,可以协议离婚。2、如果一方不同意离婚的,另一方只能起诉到法院要求判决离婚了。3、诉讼离婚需要准备身份证复印件一份;户口本原件(首页和本人页)及复印件各一张;结婚证原件及复印件两份;夫妻共同财产清单一式三份等材料提交给立案庭即可。4、法律依据《中华人民共和国民法典》第一千零七十九条 【诉讼离婚】夫妻一方要求离婚的,可以由有关组织进行调解或者直接向人民法院提起离婚诉讼。人民法院审理离婚案件,应当进行调解;如果感情确已破裂,调解无效的,应当准予离婚。有下列情形之一,调解无效的,应当准予离婚:(一)重婚或者与他人同居;(二)实施家庭暴力或者虐待、遗弃家庭成员;(三)有赌博、吸毒等恶习屡教不改;(四)因感情不和分居满二年;(五)其他导致夫妻感情破裂的情形。一方被宣告失踪,另一方提起离婚诉讼的,应当准予离婚。经人民法院判决不准离婚后,双方又分居满一年,一方再次提起离婚诉讼的,应当准予离婚。|根据中国婚姻法的规定,夫妻双方如果连续分居满两年以上,可以视为感情破裂,可以协议离婚或者起诉离婚。但是,如果夫妻之间没有达成协议,也可以通过法院判决离婚。因此,如果你和你的配偶已经连续分居超过了两年的时间,你可以考虑通过法律途径来解决你们的婚姻问题。至于是否算作自动离婚,需要具体情况具体分析,建议您咨询专业律师以获取更准确的答案。|如果双方没有办理离婚手续的话,那么婚姻关系仍然存在。但是如果你想解除这种状态下的婚姻关系,可以考虑通过协议或者诉讼的方式来解决。在协议中,你可以选择是否要分割财产、子女抚养等问题;而在诉讼中,法院会根据具体情况作出判决。无论哪种方式,都需要提供相关的证据和材料证明你们的婚姻已经无法继续下去。| </details> ## 使用方法 ```python import torch from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftModel tokenizer = AutoTokenizer.from_pretrained("baichuan-inc/baichuan-7B", trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained("baichuan-inc/baichuan-7B", device_map="auto", trust_remote_code=True) peft_model = PeftModel.from_pretrained(model, 'FedJudge/fedjudge-base-7b',torch_dtype=torch.float32).half() data = '假设你是一名律师,请回答以下向你咨询的问题:在法律中定金与订金的区别是什么?' inputs = tokenizer(data, return_tensors='pt') inputs = inputs.to('cuda:0') pred = peft_model.generate(**inputs, max_new_tokens=500,repetition_penalty=1.1) pred_result = tokenizer.decode(pred.cpu()[0], skip_special_tokens=True) print(pred_result.split(data)[-1]) ``` ## 局限性 - 在本项目中,模型的输出结果并非专业法律咨询的结果,可能会包含错误或不准确的内容。因此,如果您需要法律援助或专业法律咨询,请务必向法律专业人士寻求帮助。 - 本项目暂时仅支持单轮对话。 更多详情请查看[FedJudge项目](https://github.com/yuelinan/FedJudge)。
ailabturkiye/epha
ailabturkiye
2023-08-16T09:52:45Z
0
0
null
[ "music", "tr", "license:openrail", "region:us" ]
null
2023-08-16T09:48:53Z
--- license: openrail language: - tr tags: - music --- Epha'nın videosunun sesiyle oluşturulan ses modeli. Train benim tarafımdan yapılmıştır.
huangyuyang/Qwen-7B-Chat-int8.flm
huangyuyang
2023-08-16T09:51:01Z
0
4
null
[ "license:apache-2.0", "region:us" ]
null
2023-08-16T09:06:35Z
--- license: apache-2.0 --- fastllm model for Qwen-7B-Chat-int8 Github address: https://github.com/ztxz16/fastllm
HZ0504/sdxl_pet_lora_wangbaozi
HZ0504
2023-08-16T09:42:49Z
2
1
diffusers
[ "diffusers", "text-to-image", "autotrain", "base_model:stabilityai/stable-diffusion-xl-base-1.0", "base_model:finetune:stabilityai/stable-diffusion-xl-base-1.0", "region:us" ]
text-to-image
2023-08-16T07:55:42Z
--- base_model: stabilityai/stable-diffusion-xl-base-1.0 instance_prompt: photo of wangpaofu tags: - text-to-image - diffusers - autotrain inference: true --- # DreamBooth trained by AutoTrain Test enoder was not trained.
PhysHunter/whisper-tiny-en
PhysHunter
2023-08-16T09:16:59Z
85
0
transformers
[ "transformers", "pytorch", "tensorboard", "whisper", "automatic-speech-recognition", "generated_from_trainer", "dataset:PolyAI/minds14", "base_model:openai/whisper-tiny", "base_model:finetune:openai/whisper-tiny", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2023-08-16T08:05:42Z
--- license: apache-2.0 base_model: openai/whisper-tiny tags: - generated_from_trainer datasets: - PolyAI/minds14 metrics: - wer model-index: - name: whisper-tiny-en results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: PolyAI/minds14 type: PolyAI/minds14 config: en-US split: train args: en-US metrics: - name: Wer type: wer value: 0.3020257826887661 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # whisper-tiny-en This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the PolyAI/minds14 dataset. It achieves the following results on the evaluation set: - Loss: 0.5262 - Wer Ortho: 0.3119 - Wer: 0.3020 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 20 - training_steps: 200 ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:| | 0.4823 | 3.57 | 50 | 0.5315 | 0.3202 | 0.3088 | | 0.1361 | 7.14 | 100 | 0.4843 | 0.3253 | 0.3161 | | 0.0563 | 10.71 | 150 | 0.5113 | 0.3106 | 0.3020 | | 0.0374 | 14.29 | 200 | 0.5262 | 0.3119 | 0.3020 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu118 - Datasets 2.14.4 - Tokenizers 0.13.3
fp16-guy/YesMix_2.0_fp16_cleaned
fp16-guy
2023-08-16T09:12:55Z
0
1
null
[ "text-to-image", "region:us" ]
text-to-image
2023-08-05T09:17:35Z
--- pipeline_tag: text-to-image --- 【Checkpoint】YesMix, but fp16/cleaned - smaller size, same result. ======== /// **[**original checkpoint link**](https://civitai.com/models/9139/checkpointyesmix)** *(all rights to the model belong to zakp)* --- *[*grid 01*](https://huggingface.co/datasets/fp16-guy/grids/blob/main/yesmix%2020%2001%2020230805110225-111-CheckpointYesmix_v20-Euler%20a-6.png) *(1.99gb version)* *[*grid 02*](https://huggingface.co/datasets/fp16-guy/grids/blob/main/yesmix%2020%2002%2020230805110340-111-CheckpointYesmix_v20-Euler%20a-6.png) *(1.83gb version - no vae)* *[*grid 03*](https://huggingface.co/datasets/fp16-guy/grids_inp/blob/main/CheckpointYesmix_v20%20inp%2001%2020230815215238-111-CheckpointYesmix_v20_fp16-Euler%20a-5.5.png) *(1.99gb inpainting version)* *[*grid 04*](https://huggingface.co/datasets/fp16-guy/grids_inp/blob/main/CheckpointYesmix_v20%20inp%2002%2020230816120951-111-CheckpointYesmix_v20_fp16_no_vae-Euler%20a-5.5.png) *(1.83gb inpainting version - no vae)*
bigcode/santacoderpack
bigcode
2023-08-16T08:37:40Z
179
5
transformers
[ "transformers", "pytorch", "gpt_bigcode", "fill-mask", "code", "text-generation", "dataset:bigcode/commitpack-subset-cf", "arxiv:2308.07124", "license:bigcode-openrail-m", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2023-08-10T14:35:42Z
--- pipeline_tag: text-generation inference: true widget: - text: '<commit_before>def has_close_elements(numbers: List[float], threshold: float) -> bool:\n for idx, elem in enumerate(numbers):\n for idx2, elem2 in enumerate(numbers):\n if idx != idx2:\n distance = elem - elem2\n if distance < threshold:\n return True\n\n return False<commit_message>Fix bugs in has_close_elements.<commit_after>' example_title: Fix has_close_elements group: Python license: bigcode-openrail-m datasets: - bigcode/commitpack-subset-cf metrics: - code_eval library_name: transformers tags: - code model-index: - name: SantaCoderPack results: - task: type: text-generation dataset: type: bigcode/humanevalpack name: HumanEvalFix Python metrics: - name: pass@1 type: pass@1 value: 3.2 verified: false - task: type: text-generation dataset: type: bigcode/humanevalpack name: HumanEvalFix JavaScript metrics: - name: pass@1 type: pass@1 value: 4.9 verified: false - task: type: text-generation dataset: type: bigcode/humanevalpack name: HumanEvalFix Java metrics: - name: pass@1 type: pass@1 value: 1.8 verified: false - task: type: text-generation dataset: type: bigcode/humanevalpack name: HumanEvalFix Go metrics: - name: pass@1 type: pass@1 value: 3.6 verified: false - task: type: text-generation dataset: type: bigcode/humanevalpack name: HumanEvalFix C++ metrics: - name: pass@1 type: pass@1 value: 4.2 verified: false - task: type: text-generation dataset: type: bigcode/humanevalpack name: HumanEvalFix Rust metrics: - name: pass@1 type: pass@1 value: 1.7 verified: false - task: type: text-generation dataset: type: bigcode/humanevalpack name: HumanEvalFix Average metrics: - name: pass@1 type: pass@1 value: 3.3 verified: false --- ![Octopack](https://github.com/bigcode-project/octopack/blob/31f3320f098703c7910e43492c39366eeea68d83/banner.png?raw=true) # Table of Contents 1. [Model Summary](#model-summary) 2. [Use](#use) 3. [Training](#training) 4. [Citation](#citation) # Model Summary SantaCoderPack is an pre-trained model with the same architecture of SantaCoder on <th><a href=https://huggingface.co/datasets/bigcode/commitpack>CommitPack</a> using this format: ``` <commit_before>code_before<commit_msg>message<commit_after>code_after ``` - **Repository:** [bigcode/octopack](https://github.com/bigcode-project/octopack) - **Paper:** [OctoPack: Instruction Tuning Code Large Language Models](https://arxiv.org/abs/2308.07124) - **Languages:** Python, JavaScript, Java, C++, Go, Rust - **SantaCoderPack:** <table> <tr> <th>Data</t> <th><a href=https://huggingface.co/datasets/bigcode/commitpack>CommitPack</a></th> <td>4TB of GitHub commits across 350 programming languages</td> </tr> <tr> <th>Model</t> <th><a href=https://huggingface.co/bigcode/octocoder>SantaCoderPack</a></th> <td>SantaCoderPack (1.1B parameters) pre-trained on CommitPack</td> </tr> <tr> <th>Evaluation&nbsp;&nbsp;</t> <th><a href=https://huggingface.co/datasets/bigcode/humanevalpack>HumanEvalPack/HumanEvalFix</a></th> <td>Extension of OpenAI's HumanEval to HumanEvalFix</td> </tr> </table> # Use ## Intended use The model follows instructions provided in the input. We recommend prefacing your input with "<commit_before>def has_close_elements(numbers: List[float], threshold: float) -> bool:\n for idx, elem in enumerate(numbers):\n for idx2, elem2 in enumerate(numbers):\n if idx != idx2:\n distance = elem - elem2\n if distance < threshold:\n return True\n\n return False<commit_message>Fix bugs in has_close_elements.<commit_after>" **Feel free to share your generations in the Community tab!** ## Generation ```python # pip install -q transformers from transformers import AutoModelForCausalLM, AutoTokenizer checkpoint = "bigcode/santacoderpack" device = "cuda" # for GPU usage or "cpu" for CPU usage tokenizer = AutoTokenizer.from_pretrained(checkpoint) model = AutoModelForCausalLM.from_pretrained(checkpoint).to(device) inputs = tokenizer.encode("Q<commit_before>def has_close_elements(numbers: List[float], threshold: float) -> bool:\n for idx, elem in enumerate(numbers):\n for idx2, elem2 in enumerate(numbers):\n if idx != idx2:\n distance = elem - elem2\n if distance < threshold:\n return True\n\n return False<commit_message>Fix bugs in has_close_elements.<commit_after>", return_tensors="pt").to(device) outputs = model.generate(inputs) print(tokenizer.decode(outputs[0])) ``` # Training ## Model - **Architecture:** GPT-2 model with multi-query attention - **Steps:** 250k pretraining - **Pretraining tokens:** 131B - **Precision:** bfloat16 ## Hardware - **Pretraining:** - **GPUs:** 32 Tesla A100 - **Training time:** 15 days ## Software - **Orchestration:** [Megatron-LM/Transformers](https://github.com/bigcode-project/santacoderpack#training) - **Neural networks:** [PyTorch](https://github.com/pytorch/pytorch) # Citation ```bibtex @article{muennighoff2023octopack, title={OctoPack: Instruction Tuning Code Large Language Models}, author={Niklas Muennighoff and Qian Liu and Armel Zebaze and Qinkai Zheng and Binyuan Hui and Terry Yue Zhuo and Swayam Singh and Xiangru Tang and Leandro von Werra and Shayne Longpre}, journal={arXiv preprint arXiv:2308.07124}, year={2023} } ```
sbpark/xlm-roberta-base-finetuned-panx-all
sbpark
2023-08-16T08:11:34Z
108
0
transformers
[ "transformers", "pytorch", "xlm-roberta", "token-classification", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2023-08-16T08:04:14Z
--- license: mit tags: - generated_from_trainer metrics: - f1 model-index: - name: xlm-roberta-base-finetuned-panx-all results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-panx-all This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.1650 - F1: 0.8569 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 24 - eval_batch_size: 24 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.2962 | 1.0 | 715 | 0.1819 | 0.8285 | | 0.1477 | 2.0 | 1430 | 0.1723 | 0.8500 | | 0.0954 | 3.0 | 2145 | 0.1650 | 0.8569 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.0 - Datasets 2.1.0 - Tokenizers 0.13.3
sbpark/xlm-roberta-base-finetuned-panx-it
sbpark
2023-08-16T08:02:30Z
103
0
transformers
[ "transformers", "pytorch", "xlm-roberta", "token-classification", "generated_from_trainer", "dataset:xtreme", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2023-08-16T08:00:55Z
--- license: mit tags: - generated_from_trainer datasets: - xtreme metrics: - f1 model-index: - name: xlm-roberta-base-finetuned-panx-it results: - task: name: Token Classification type: token-classification dataset: name: xtreme type: xtreme config: PAN-X.it split: validation args: PAN-X.it metrics: - name: F1 type: f1 value: 0.8172920065252856 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-panx-it This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the xtreme dataset. It achieves the following results on the evaluation set: - Loss: 0.2486 - F1: 0.8173 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 24 - eval_batch_size: 24 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.813 | 1.0 | 70 | 0.3253 | 0.7414 | | 0.288 | 2.0 | 140 | 0.2487 | 0.7994 | | 0.188 | 3.0 | 210 | 0.2486 | 0.8173 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.0 - Datasets 2.1.0 - Tokenizers 0.13.3
sbpark/xlm-roberta-base-finetuned-panx-de-fr
sbpark
2023-08-16T07:58:24Z
109
0
transformers
[ "transformers", "pytorch", "xlm-roberta", "token-classification", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2023-08-16T07:50:57Z
--- license: mit tags: - generated_from_trainer metrics: - f1 model-index: - name: xlm-roberta-base-finetuned-panx-de-fr results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-panx-de-fr This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.1587 - F1: 0.8595 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 24 - eval_batch_size: 24 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.288 | 1.0 | 715 | 0.1785 | 0.8194 | | 0.15 | 2.0 | 1430 | 0.1593 | 0.8475 | | 0.0965 | 3.0 | 2145 | 0.1587 | 0.8595 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.0 - Datasets 2.1.0 - Tokenizers 0.13.3