modelId
stringlengths
5
139
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]date
2020-02-15 11:33:14
2025-09-12 06:31:37
downloads
int64
0
223M
likes
int64
0
11.7k
library_name
stringclasses
555 values
tags
listlengths
1
4.05k
pipeline_tag
stringclasses
55 values
createdAt
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-09-12 06:31:07
card
stringlengths
11
1.01M
seraphimzzzz/44280
seraphimzzzz
2025-08-19T23:05:37Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:05:34Z
[View on Civ Archive](https://civarchive.com/models/58312?modelVersionId=62763)
ultratopaz/66568
ultratopaz
2025-08-19T23:05:13Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:05:10Z
[View on Civ Archive](https://civarchive.com/models/90115?modelVersionId=95960)
calegpedia/blockassist-bc-stealthy_slimy_rooster_1755643065
calegpedia
2025-08-19T23:05:10Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "stealthy slimy rooster", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T23:05:06Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - stealthy slimy rooster --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
ultratopaz/54874
ultratopaz
2025-08-19T23:04:40Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:04:38Z
[View on Civ Archive](https://civarchive.com/models/75239?modelVersionId=79980)
crystalline7/65816
crystalline7
2025-08-19T23:04:33Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:04:30Z
[View on Civ Archive](https://civarchive.com/models/89255?modelVersionId=95009)
ultratopaz/49808
ultratopaz
2025-08-19T23:04:09Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:04:06Z
[View on Civ Archive](https://civarchive.com/models/66953?modelVersionId=71614)
ultratopaz/43216
ultratopaz
2025-08-19T23:04:02Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:03:59Z
[View on Civ Archive](https://civarchive.com/models/56621?modelVersionId=61044)
seraphimzzzz/55282
seraphimzzzz
2025-08-19T23:03:45Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:03:43Z
[View on Civ Archive](https://civarchive.com/models/75887?modelVersionId=80622)
crystalline7/52226
crystalline7
2025-08-19T23:03:29Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:03:25Z
[View on Civ Archive](https://civarchive.com/models/70957?modelVersionId=75646)
crystalline7/42524
crystalline7
2025-08-19T23:03:12Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:03:09Z
[View on Civ Archive](https://civarchive.com/models/55594?modelVersionId=59988)
crystalline7/46077
crystalline7
2025-08-19T23:03:04Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:03:01Z
[View on Civ Archive](https://civarchive.com/models/61266?modelVersionId=65736)
seraphimzzzz/478981
seraphimzzzz
2025-08-19T23:02:45Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:02:40Z
[View on Civ Archive](https://civarchive.com/models/506592?modelVersionId=563066)
crystalline7/21589
crystalline7
2025-08-19T23:02:34Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:02:30Z
[View on Civ Archive](https://civarchive.com/models/21796?modelVersionId=26015)
thiernomdou/Karamoo
thiernomdou
2025-08-19T23:02:22Z
0
0
diffusers
[ "diffusers", "flux", "lora", "replicate", "text-to-image", "en", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-08-19T22:53:37Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - replicate base_model: "black-forest-labs/FLUX.1-dev" pipeline_tag: text-to-image # widget: # - text: >- # prompt # output: # url: https://... instance_prompt: karamoo --- # Karamoo <Gallery /> ## About this LoRA This is a [LoRA](https://replicate.com/docs/guides/working-with-loras) for the FLUX.1-dev text-to-image model. It can be used with diffusers or ComfyUI. It was trained on [Replicate](https://replicate.com/) using AI toolkit: https://replicate.com/ostris/flux-dev-lora-trainer/train ## Trigger words You should use `karamoo` to trigger the image generation. ## Run this LoRA with an API using Replicate ```py import replicate input = { "prompt": "karamoo", "lora_weights": "https://huggingface.co/thiernomdou/Karamoo/resolve/main/lora.safetensors" } output = replicate.run( "black-forest-labs/flux-dev-lora", input=input ) for index, item in enumerate(output): with open(f"output_{index}.webp", "wb") as file: file.write(item.read()) ``` ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('thiernomdou/Karamoo', weight_name='lora.safetensors') image = pipeline('karamoo').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters) ## Training details - Steps: 1000 - Learning rate: 0.0004 - LoRA rank: 16 ## Contribute your own examples You can use the [community tab](https://huggingface.co/thiernomdou/Karamoo/discussions) to add images that show off what you’ve made with this LoRA.
katanyasekolah/blockassist-bc-silky_sprightly_cassowary_1755642890
katanyasekolah
2025-08-19T23:01:56Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "silky sprightly cassowary", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T23:01:52Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - silky sprightly cassowary --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
ultratopaz/87314
ultratopaz
2025-08-19T23:01:51Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:01:49Z
[View on Civ Archive](https://civarchive.com/models/112043?modelVersionId=120937)
ultratopaz/93085
ultratopaz
2025-08-19T23:01:42Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:01:35Z
[View on Civ Archive](https://civarchive.com/models/118503?modelVersionId=128554)
seraphimzzzz/63601
seraphimzzzz
2025-08-19T23:01:16Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:01:12Z
[View on Civ Archive](https://civarchive.com/models/68607?modelVersionId=92238)
EldritchLavender/SpookieBrain
EldritchLavender
2025-08-19T23:01:12Z
3
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-04-30T12:20:32Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
crystalline7/67939
crystalline7
2025-08-19T23:00:55Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:00:52Z
[View on Civ Archive](https://civarchive.com/models/91623?modelVersionId=97665)
dgambettaphd/M_mis_run2_gen6_WXS_doc1000_synt64_lr1e-04_acm_LANG
dgambettaphd
2025-08-19T23:00:42Z
0
0
transformers
[ "transformers", "safetensors", "unsloth", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2025-08-19T23:00:27Z
--- library_name: transformers tags: - unsloth --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
ultratopaz/42516
ultratopaz
2025-08-19T23:00:29Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:00:25Z
[View on Civ Archive](https://civarchive.com/models/55583?modelVersionId=59976)
ver-video-intimo-de-Clip-abigail-lalama/ver-video-intimo-de-Clip-abigail-lalama
ver-video-intimo-de-Clip-abigail-lalama
2025-08-19T23:00:27Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:00:15Z
<animated-image data-catalyst=""><a href="https://tinyurl.com/5ye5v3bc?leaked-viral-video" rel="nofollow" data-target="animated-image.originalLink"><img src="https://static.wixstatic.com/media/b249f9_adac8f70fb3f45b88691696c77de18f3~mv2.gif" alt="Foo" data-canonical-src="https://static.wixstatic.com/media/b249f9_adac8f70fb3f45b88691696c77de18f3~mv2.gif" style="max-width: 100%; display: inline-block;" data-target="animated-image.originalImage"></a>
crystalline7/100394
crystalline7
2025-08-19T23:00:12Z
0
0
null
[ "region:us" ]
null
2025-08-19T23:00:09Z
[View on Civ Archive](https://civarchive.com/models/125655?modelVersionId=137280)
ultratopaz/51095
ultratopaz
2025-08-19T22:58:38Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:58:36Z
[View on Civ Archive](https://civarchive.com/models/69116?modelVersionId=73794)
crystalline7/19054
crystalline7
2025-08-19T22:58:22Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:58:19Z
[View on Civ Archive](https://civarchive.com/models/12317?modelVersionId=22899)
seraphimzzzz/12626
seraphimzzzz
2025-08-19T22:58:13Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:58:13Z
[View on Civ Archive](https://civarchive.com/models/12317?modelVersionId=14582)
AnonymousCS/xlmr_immigration_combo6_4
AnonymousCS
2025-08-19T22:58:10Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-large", "base_model:finetune:FacebookAI/xlm-roberta-large", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-08-19T22:53:33Z
--- library_name: transformers license: mit base_model: FacebookAI/xlm-roberta-large tags: - generated_from_trainer metrics: - accuracy model-index: - name: xlmr_immigration_combo6_4 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlmr_immigration_combo6_4 This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.1129 - Accuracy: 0.9704 - 1-f1: 0.9550 - 1-recall: 0.9421 - 1-precision: 0.9683 - Balanced Acc: 0.9633 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 128 - eval_batch_size: 128 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 15 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | 1-f1 | 1-recall | 1-precision | Balanced Acc | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:--------:|:-----------:|:------------:| | 0.1642 | 1.0 | 25 | 0.1001 | 0.9756 | 0.9619 | 0.9266 | 1.0 | 0.9633 | | 0.1556 | 2.0 | 50 | 0.0998 | 0.9730 | 0.9583 | 0.9305 | 0.9877 | 0.9624 | | 0.1171 | 3.0 | 75 | 0.0989 | 0.9756 | 0.9619 | 0.9266 | 1.0 | 0.9633 | | 0.0795 | 4.0 | 100 | 0.1169 | 0.9692 | 0.9518 | 0.9151 | 0.9916 | 0.9556 | | 0.0484 | 5.0 | 125 | 0.1129 | 0.9704 | 0.9550 | 0.9421 | 0.9683 | 0.9633 | ### Framework versions - Transformers 4.56.0.dev0 - Pytorch 2.8.0+cu126 - Datasets 4.0.0 - Tokenizers 0.21.4
crystalline7/13305
crystalline7
2025-08-19T22:57:44Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:57:40Z
[View on Civ Archive](https://civarchive.com/models/13173?modelVersionId=15525)
ultratopaz/63808
ultratopaz
2025-08-19T22:57:34Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:57:32Z
[View on Civ Archive](https://civarchive.com/models/86958?modelVersionId=92511)
seraphimzzzz/26100
seraphimzzzz
2025-08-19T22:57:26Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:57:22Z
[View on Civ Archive](https://civarchive.com/models/26393?modelVersionId=31601)
quantumxnode/blockassist-bc-dormant_peckish_seahorse_1755642670
quantumxnode
2025-08-19T22:57:15Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "dormant peckish seahorse", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T22:57:12Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - dormant peckish seahorse --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
koloni/blockassist-bc-deadly_graceful_stingray_1755642634
koloni
2025-08-19T22:57:12Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "deadly graceful stingray", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T22:57:08Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - deadly graceful stingray --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
seraphimzzzz/778755
seraphimzzzz
2025-08-19T22:57:06Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:57:03Z
[View on Civ Archive](https://civarchive.com/models/376220?modelVersionId=869936)
crystalline7/112237
crystalline7
2025-08-19T22:56:46Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:56:44Z
[View on Civ Archive](https://civarchive.com/models/136795?modelVersionId=150921)
crystalline7/16814
crystalline7
2025-08-19T22:56:00Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:55:55Z
[View on Civ Archive](https://civarchive.com/models/17067?modelVersionId=20153)
GeneroGral/Mistral-Nemo-12B_BBQ_Stereo_MERGED7_dropout_batch-wordMatch
GeneroGral
2025-08-19T22:55:55Z
0
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "en", "base_model:unsloth/Mistral-Nemo-Base-2407", "base_model:finetune:unsloth/Mistral-Nemo-Base-2407", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2025-08-19T22:51:20Z
--- base_model: unsloth/Mistral-Nemo-Base-2407 tags: - text-generation-inference - transformers - unsloth - mistral - trl - sft license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** GeneroGral - **License:** apache-2.0 - **Finetuned from model :** unsloth/Mistral-Nemo-Base-2407 This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
ultratopaz/59092
ultratopaz
2025-08-19T22:55:12Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:55:09Z
[View on Civ Archive](https://civarchive.com/models/81480?modelVersionId=86456)
chainway9/blockassist-bc-untamed_quick_eel_1755642411
chainway9
2025-08-19T22:55:05Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "untamed quick eel", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T22:55:01Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - untamed quick eel --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
mang3dd/blockassist-bc-tangled_slithering_alligator_1755642520
mang3dd
2025-08-19T22:55:02Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "tangled slithering alligator", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T22:54:57Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - tangled slithering alligator --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
ultratopaz/38152
ultratopaz
2025-08-19T22:55:02Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:55:02Z
[View on Civ Archive](https://civarchive.com/models/47805?modelVersionId=52399)
crystalline7/76501
crystalline7
2025-08-19T22:54:34Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:54:34Z
[View on Civ Archive](https://civarchive.com/models/18234?modelVersionId=108518)
seraphimzzzz/22838
seraphimzzzz
2025-08-19T22:54:28Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:54:23Z
[View on Civ Archive](https://civarchive.com/models/18234?modelVersionId=27622)
ultratopaz/23412
ultratopaz
2025-08-19T22:53:27Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:53:22Z
[View on Civ Archive](https://civarchive.com/models/23721?modelVersionId=28338)
ultratopaz/24443
ultratopaz
2025-08-19T22:52:49Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:52:44Z
[View on Civ Archive](https://civarchive.com/models/24707?modelVersionId=29556)
crystalline7/38157
crystalline7
2025-08-19T22:52:38Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:52:33Z
[View on Civ Archive](https://civarchive.com/models/47820?modelVersionId=52412)
AnonymousCS/xlmr_immigration_combo6_3
AnonymousCS
2025-08-19T22:52:26Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-large", "base_model:finetune:FacebookAI/xlm-roberta-large", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-08-19T22:48:23Z
--- library_name: transformers license: mit base_model: FacebookAI/xlm-roberta-large tags: - generated_from_trainer metrics: - accuracy model-index: - name: xlmr_immigration_combo6_3 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlmr_immigration_combo6_3 This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.2175 - Accuracy: 0.9293 - 1-f1: 0.8968 - 1-recall: 0.9228 - 1-precision: 0.8723 - Balanced Acc: 0.9277 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 128 - eval_batch_size: 128 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 15 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | 1-f1 | 1-recall | 1-precision | Balanced Acc | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:--------:|:-----------:|:------------:| | 0.1432 | 1.0 | 25 | 0.2095 | 0.9242 | 0.8921 | 0.9421 | 0.8472 | 0.9287 | | 0.0923 | 2.0 | 50 | 0.1510 | 0.9563 | 0.9328 | 0.9112 | 0.9555 | 0.9450 | | 0.1044 | 3.0 | 75 | 0.1735 | 0.9524 | 0.9284 | 0.9266 | 0.9302 | 0.9460 | | 0.1035 | 4.0 | 100 | 0.2175 | 0.9293 | 0.8968 | 0.9228 | 0.8723 | 0.9277 | ### Framework versions - Transformers 4.56.0.dev0 - Pytorch 2.8.0+cu126 - Datasets 4.0.0 - Tokenizers 0.21.4
seraphimzzzz/23439
seraphimzzzz
2025-08-19T22:52:26Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:52:22Z
[View on Civ Archive](https://civarchive.com/models/23740?modelVersionId=28362)
ultratopaz/85990
ultratopaz
2025-08-19T22:52:05Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:52:02Z
[View on Civ Archive](https://civarchive.com/models/39509?modelVersionId=119927)
seraphimzzzz/34079
seraphimzzzz
2025-08-19T22:51:56Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:51:51Z
[View on Civ Archive](https://civarchive.com/models/39509?modelVersionId=45600)
thanobidex/blockassist-bc-colorful_shiny_hare_1755642359
thanobidex
2025-08-19T22:51:48Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "colorful shiny hare", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T22:51:45Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - colorful shiny hare --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
crystalline7/25191
crystalline7
2025-08-19T22:51:37Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:51:32Z
[View on Civ Archive](https://civarchive.com/models/25486?modelVersionId=30512)
ultratopaz/62902
ultratopaz
2025-08-19T22:51:27Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:51:24Z
[View on Civ Archive](https://civarchive.com/models/19239?modelVersionId=91302)
nkerr/sv3.4-bigbird-roberta-large
nkerr
2025-08-19T22:51:18Z
0
0
transformers
[ "transformers", "safetensors", "big_bird", "text-classification", "generated_from_trainer", "base_model:google/bigbird-roberta-large", "base_model:finetune:google/bigbird-roberta-large", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-08-19T21:52:07Z
--- library_name: transformers license: apache-2.0 base_model: google/bigbird-roberta-large tags: - generated_from_trainer model-index: - name: sv3.4-bigbird-roberta-large results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # sv3.4-bigbird-roberta-large This model is a fine-tuned version of [google/bigbird-roberta-large](https://huggingface.co/google/bigbird-roberta-large) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.5284 - Mse: 0.3063 - Mae: 0.5284 - Rmse: 0.5534 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 8 - eval_batch_size: 4 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Mse | Mae | Rmse | |:-------------:|:------:|:----:|:---------------:|:------:|:------:|:------:| | 0.2148 | 0.1548 | 50 | 0.6346 | 0.4310 | 0.6346 | 0.6565 | | 0.2317 | 0.3096 | 100 | 0.4964 | 0.2705 | 0.4964 | 0.5201 | | 0.22 | 0.4644 | 150 | 0.5909 | 0.3752 | 0.5909 | 0.6125 | | 0.1846 | 0.6192 | 200 | 0.5494 | 0.3274 | 0.5494 | 0.5722 | | 0.1858 | 0.7740 | 250 | 0.5280 | 0.3046 | 0.5280 | 0.5520 | | 0.1886 | 0.9288 | 300 | 0.5284 | 0.3063 | 0.5284 | 0.5534 | ### Framework versions - Transformers 4.49.0 - Pytorch 2.6.0+cu126 - Datasets 3.3.2 - Tokenizers 0.21.0
seraphimzzzz/25253
seraphimzzzz
2025-08-19T22:51:10Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:51:06Z
[View on Civ Archive](https://civarchive.com/models/19239?modelVersionId=30605)
neko-llm/Qwen3-235B-test5
neko-llm
2025-08-19T22:51:04Z
0
0
transformers
[ "transformers", "safetensors", "generated_from_trainer", "trl", "sft", "base_model:Qwen/Qwen3-235B-A22B", "base_model:finetune:Qwen/Qwen3-235B-A22B", "endpoints_compatible", "region:us" ]
null
2025-08-19T12:49:08Z
--- base_model: Qwen/Qwen3-235B-A22B library_name: transformers model_name: Qwen3-235B-test5 tags: - generated_from_trainer - trl - sft licence: license --- # Model Card for Qwen3-235B-test5 This model is a fine-tuned version of [Qwen/Qwen3-235B-A22B](https://huggingface.co/Qwen/Qwen3-235B-A22B). It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="neko-llm/Qwen3-235B-test5", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/neko-llm/huggingface/runs/r6shuvcx) This model was trained with SFT. ### Framework versions - TRL: 0.19.0 - Transformers: 4.54.1 - Pytorch: 2.6.0 - Datasets: 4.0.0 - Tokenizers: 0.21.4 ## Citations Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
seraphimzzzz/19008
seraphimzzzz
2025-08-19T22:50:59Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:50:55Z
[View on Civ Archive](https://civarchive.com/models/19239?modelVersionId=22829)
crystalline7/77154
crystalline7
2025-08-19T22:50:50Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:50:47Z
[View on Civ Archive](https://civarchive.com/models/35216?modelVersionId=109272)
seraphimzzzz/75514
seraphimzzzz
2025-08-19T22:50:31Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:50:28Z
[View on Civ Archive](https://civarchive.com/models/53478?modelVersionId=107222)
ultratopaz/41221
ultratopaz
2025-08-19T22:50:22Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:50:18Z
[View on Civ Archive](https://civarchive.com/models/53478?modelVersionId=57838)
crystalline7/18666
crystalline7
2025-08-19T22:50:11Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:50:07Z
[View on Civ Archive](https://civarchive.com/models/18877?modelVersionId=22403)
crystalline7/81271
crystalline7
2025-08-19T22:49:40Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:49:37Z
[View on Civ Archive](https://civarchive.com/models/17612?modelVersionId=114290)
crystalline7/25540
crystalline7
2025-08-19T22:49:31Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:49:26Z
[View on Civ Archive](https://civarchive.com/models/17612?modelVersionId=30958)
crystalline7/56907
crystalline7
2025-08-19T22:49:21Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:49:18Z
[View on Civ Archive](https://civarchive.com/models/24156?modelVersionId=83175)
ultratopaz/24129
ultratopaz
2025-08-19T22:49:12Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:49:07Z
[View on Civ Archive](https://civarchive.com/models/24156?modelVersionId=28870)
crystalline7/80429
crystalline7
2025-08-19T22:49:02Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:48:59Z
[View on Civ Archive](https://civarchive.com/models/26639?modelVersionId=113278)
Mahran4vp/results
Mahran4vp
2025-08-19T22:48:39Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "gpt2", "text-generation", "generated_from_trainer", "base_model:Mahran4vp/gpt2-hoodie-final", "base_model:finetune:Mahran4vp/gpt2-hoodie-final", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-08-19T22:48:06Z
--- library_name: transformers base_model: Mahran4vp/gpt2-hoodie-final tags: - generated_from_trainer model-index: - name: results results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # results This model is a fine-tuned version of [Mahran4vp/gpt2-hoodie-final](https://huggingface.co/Mahran4vp/gpt2-hoodie-final) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.55.2 - Pytorch 2.6.0+cu124 - Datasets 4.0.0 - Tokenizers 0.21.4
crystalline7/26964
crystalline7
2025-08-19T22:48:34Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:48:29Z
[View on Civ Archive](https://civarchive.com/models/27347?modelVersionId=32745)
crystalline7/88151
crystalline7
2025-08-19T22:48:13Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:48:10Z
[View on Civ Archive](https://civarchive.com/models/113358?modelVersionId=122445)
seraphimzzzz/16883
seraphimzzzz
2025-08-19T22:47:59Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:47:57Z
[View on Civ Archive](https://civarchive.com/models/17096?modelVersionId=20186)
crystalline7/37649
crystalline7
2025-08-19T22:47:45Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:47:43Z
[View on Civ Archive](https://civarchive.com/models/47019?modelVersionId=51608)
crystalline7/65487
crystalline7
2025-08-19T22:47:32Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:47:30Z
[View on Civ Archive](https://civarchive.com/models/88911?modelVersionId=94611)
ultratopaz/47611
ultratopaz
2025-08-19T22:47:26Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:47:23Z
[View on Civ Archive](https://civarchive.com/models/63501?modelVersionId=68061)
seraphimzzzz/44098
seraphimzzzz
2025-08-19T22:47:18Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:47:16Z
[View on Civ Archive](https://civarchive.com/models/58002?modelVersionId=62451)
crystalline7/99930
crystalline7
2025-08-19T22:47:12Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:47:09Z
[View on Civ Archive](https://civarchive.com/models/125177?modelVersionId=136725)
seraphimzzzz/51847
seraphimzzzz
2025-08-19T22:47:05Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:47:02Z
[View on Civ Archive](https://civarchive.com/models/70446?modelVersionId=75104)
dsisodia/ai-nidhi
dsisodia
2025-08-19T22:47:03Z
0
0
diffusers
[ "diffusers", "flux", "lora", "replicate", "text-to-image", "en", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-08-19T21:45:07Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - replicate base_model: "black-forest-labs/FLUX.1-dev" pipeline_tag: text-to-image # widget: # - text: >- # prompt # output: # url: https://... instance_prompt: AINIDHI --- # Ai Nidhi <Gallery /> ## About this LoRA This is a [LoRA](https://replicate.com/docs/guides/working-with-loras) for the FLUX.1-dev text-to-image model. It can be used with diffusers or ComfyUI. It was trained on [Replicate](https://replicate.com/) using AI toolkit: https://replicate.com/ostris/flux-dev-lora-trainer/train ## Trigger words You should use `AINIDHI` to trigger the image generation. ## Run this LoRA with an API using Replicate ```py import replicate input = { "prompt": "AINIDHI", "lora_weights": "https://huggingface.co/dsisodia/ai-nidhi/resolve/main/lora.safetensors" } output = replicate.run( "black-forest-labs/flux-dev-lora", input=input ) for index, item in enumerate(output): with open(f"output_{index}.webp", "wb") as file: file.write(item.read()) ``` ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('dsisodia/ai-nidhi', weight_name='lora.safetensors') image = pipeline('AINIDHI').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters) ## Training details - Steps: 1200 - Learning rate: 0.0004 - LoRA rank: 32 ## Contribute your own examples You can use the [community tab](https://huggingface.co/dsisodia/ai-nidhi/discussions) to add images that show off what you’ve made with this LoRA.
crystalline7/51698
crystalline7
2025-08-19T22:46:56Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:46:52Z
[View on Civ Archive](https://civarchive.com/models/20833?modelVersionId=74853)
lilTAT/blockassist-bc-gentle_rugged_hare_1755643544
lilTAT
2025-08-19T22:46:11Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "gentle rugged hare", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T22:46:07Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - gentle rugged hare --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
leandrobarrososouza/RCS
leandrobarrososouza
2025-08-19T22:45:44Z
0
0
null
[ "license:apache-2.0", "region:us" ]
null
2025-08-19T22:45:44Z
--- license: apache-2.0 ---
crystalline7/9861
crystalline7
2025-08-19T22:45:40Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:45:35Z
[View on Civ Archive](https://civarchive.com/models/8772?modelVersionId=10357)
crystalline7/185162
crystalline7
2025-08-19T22:45:15Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:45:12Z
[View on Civ Archive](https://civarchive.com/models/212601?modelVersionId=239496)
seraphimzzzz/32237
seraphimzzzz
2025-08-19T22:44:55Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:44:51Z
[View on Civ Archive](https://civarchive.com/models/35827?modelVersionId=42023)
ultratopaz/6908
ultratopaz
2025-08-19T22:44:46Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:44:41Z
[View on Civ Archive](https://civarchive.com/models/5764?modelVersionId=6719)
seraphimzzzz/637434
seraphimzzzz
2025-08-19T22:44:35Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:44:33Z
[View on Civ Archive](https://civarchive.com/models/545506?modelVersionId=722866)
seraphimzzzz/59591
seraphimzzzz
2025-08-19T22:44:22Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:44:19Z
[View on Civ Archive](https://civarchive.com/models/75223?modelVersionId=87076)
crystalline7/54866
crystalline7
2025-08-19T22:44:07Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:44:04Z
[View on Civ Archive](https://civarchive.com/models/75223?modelVersionId=79969)
ultratopaz/61978
ultratopaz
2025-08-19T22:43:59Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:43:56Z
[View on Civ Archive](https://civarchive.com/models/80680?modelVersionId=90116)
ultratopaz/48631
ultratopaz
2025-08-19T22:43:52Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:43:50Z
[View on Civ Archive](https://civarchive.com/models/65010?modelVersionId=69636)
crystalline7/54021
crystalline7
2025-08-19T22:43:46Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:43:43Z
[View on Civ Archive](https://civarchive.com/models/73874?modelVersionId=78592)
crystalline7/28906
crystalline7
2025-08-19T22:43:01Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:42:57Z
[View on Civ Archive](https://civarchive.com/models/29642?modelVersionId=35659)
ypszn/blockassist-bc-yapping_pawing_worm_1755643286
ypszn
2025-08-19T22:42:40Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "yapping pawing worm", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T22:42:32Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - yapping pawing worm --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
seraphimzzzz/640572
seraphimzzzz
2025-08-19T22:42:39Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:42:37Z
[View on Civ Archive](https://civarchive.com/models/462107?modelVersionId=726101)
ultratopaz/88590
ultratopaz
2025-08-19T22:42:32Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:42:29Z
[View on Civ Archive](https://civarchive.com/models/113828?modelVersionId=123010)
crystalline7/96922
crystalline7
2025-08-19T22:42:22Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:42:14Z
[View on Civ Archive](https://civarchive.com/models/122190?modelVersionId=133017)
AnonymousCS/xlmr_immigration_combo6_2
AnonymousCS
2025-08-19T22:42:16Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-large", "base_model:finetune:FacebookAI/xlm-roberta-large", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-08-19T22:38:15Z
--- library_name: transformers license: mit base_model: FacebookAI/xlm-roberta-large tags: - generated_from_trainer metrics: - accuracy model-index: - name: xlmr_immigration_combo6_2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlmr_immigration_combo6_2 This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.2027 - Accuracy: 0.9357 - 1-f1: 0.9038 - 1-recall: 0.9073 - 1-precision: 0.9004 - Balanced Acc: 0.9286 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 128 - eval_batch_size: 128 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 15 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | 1-f1 | 1-recall | 1-precision | Balanced Acc | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:--------:|:-----------:|:------------:| | 0.216 | 1.0 | 25 | 0.1942 | 0.9344 | 0.9054 | 0.9421 | 0.8714 | 0.9364 | | 0.1415 | 2.0 | 50 | 0.1577 | 0.9499 | 0.9234 | 0.9073 | 0.94 | 0.9392 | | 0.2224 | 3.0 | 75 | 0.2337 | 0.9242 | 0.8913 | 0.9344 | 0.8521 | 0.9267 | | 0.0914 | 4.0 | 100 | 0.2027 | 0.9357 | 0.9038 | 0.9073 | 0.9004 | 0.9286 | ### Framework versions - Transformers 4.56.0.dev0 - Pytorch 2.8.0+cu126 - Datasets 4.0.0 - Tokenizers 0.21.4
ultratopaz/126381
ultratopaz
2025-08-19T22:42:07Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:42:02Z
[View on Civ Archive](https://civarchive.com/models/60161?modelVersionId=166777)
seraphimzzzz/63725
seraphimzzzz
2025-08-19T22:41:30Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:41:25Z
[View on Civ Archive](https://civarchive.com/models/86857?modelVersionId=92402)
crystalline7/75620
crystalline7
2025-08-19T22:41:05Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:41:02Z
[View on Civ Archive](https://civarchive.com/models/100308?modelVersionId=107359)
seraphimzzzz/76813
seraphimzzzz
2025-08-19T22:40:42Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:40:38Z
[View on Civ Archive](https://civarchive.com/models/79095?modelVersionId=108869)