modelId
stringlengths 5
139
| author
stringlengths 2
42
| last_modified
timestamp[us, tz=UTC]date 2020-02-15 11:33:14
2025-09-02 06:30:45
| downloads
int64 0
223M
| likes
int64 0
11.7k
| library_name
stringclasses 533
values | tags
listlengths 1
4.05k
| pipeline_tag
stringclasses 55
values | createdAt
timestamp[us, tz=UTC]date 2022-03-02 23:29:04
2025-09-02 06:30:39
| card
stringlengths 11
1.01M
|
---|---|---|---|---|---|---|---|---|---|
omerbkts/blockassist-bc-keen_fast_giraffe_1756794107
|
omerbkts
| 2025-09-02T06:22:12Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"keen fast giraffe",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:22:06Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- keen fast giraffe
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
kcfabulosa/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-gentle_jumping_termite
|
kcfabulosa
| 2025-09-02T06:21:41Z | 10 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"rl-swarm",
"grpo",
"gensyn",
"I am gentle jumping termite",
"trl",
"genrl-swarm",
"I am gentle_jumping_termite",
"conversational",
"arxiv:2402.03300",
"base_model:unsloth/Qwen2.5-0.5B-Instruct",
"base_model:finetune:unsloth/Qwen2.5-0.5B-Instruct",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-05-13T22:28:29Z |
---
base_model: unsloth/Qwen2.5-0.5B-Instruct
library_name: transformers
model_name: Qwen2.5-0.5B-Instruct-Gensyn-Swarm-gentle_jumping_termite
tags:
- generated_from_trainer
- rl-swarm
- grpo
- gensyn
- I am gentle jumping termite
- trl
- genrl-swarm
- I am gentle_jumping_termite
licence: license
---
# Model Card for Qwen2.5-0.5B-Instruct-Gensyn-Swarm-gentle_jumping_termite
This model is a fine-tuned version of [unsloth/Qwen2.5-0.5B-Instruct](https://huggingface.co/unsloth/Qwen2.5-0.5B-Instruct).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="kcfabulosa/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-gentle_jumping_termite", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
### Framework versions
- TRL: 0.17.0
- Transformers: 4.51.3
- Pytorch: 2.7.0
- Datasets: 3.6.0
- Tokenizers: 0.21.1
## Citations
Cite GRPO as:
```bibtex
@article{zhihong2024deepseekmath,
title = {{DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models}},
author = {Zhihong Shao and Peiyi Wang and Qihao Zhu and Runxin Xu and Junxiao Song and Mingchuan Zhang and Y. K. Li and Y. Wu and Daya Guo},
year = 2024,
eprint = {arXiv:2402.03300},
}
```
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
duohuang/Affine-copycat
|
duohuang
| 2025-09-02T06:21:19Z | 0 | 0 | null |
[
"safetensors",
"gpt_oss",
"8-bit",
"mxfp4",
"region:us"
] | null | 2025-09-02T06:17:59Z |
# Affine
Mine open reasoning.
[Affine Discord](https://discord.com/invite/3T9X4Yn23e)
## Introduction
Affine is an incentivized RL environment which pays miners which make incremental improvements on a set of tasks (for instance, program abduction or coding). The mechanism is sybil-proof (you can't cheat by deploying multiple miners), decoy-proof (you can't cheat by packing models into certain environments), copy-proof (you can't cheat by stealing models), overfitting-proof (you can't cheat by overfitting to a single env).
How does Affine work? Affine validators incentivize miners to submit models to Subnet 64 on Bittensor (a.k.a Chutes) where they are inference load balanced and publicly available. These models are evaluated on a set of RL-environments with validators looking for the model which dominates the pareto frontier -- namely the model which outcompetes all other models on all envs (see `af validator`) The network is winners-take-all where miners are forced to copy, download and improve the pareto frontier model.
Why affine? Directed incentives for RL have never been achieved. The ability to direct intelligence and aggregate the work-effort of a large non-permissioned group of individuals on RL tasks will unlock fast advancement in intelligence, we intend to commoditize reasoning (intelligence's highest form) and break the intelligence sound barrier.
## Installation
```bash
# Install uv Astral
curl -LsSf https://astral.sh/uv/install.sh | sh
# Clone and install Affine
git clone https://github.com/AffineFoundation/affine.git
cd affine
uv venv && source .venv/bin/activate && uv pip install -e .
# Verify installation
af
```
## Validating
Set env vars, chutes api key.
```bash
# Copy .env and fill out validator items
cp .env.example .env
```
(Recommended): Run the validator with docker and watchtower autoupdate.
```bash
# Run the validator with watchtower.
docker-compose down && docker-compose pull && docker-compose up -d && docker-compose logs -f
```
Run the validator using the local override (build local image) + base compose
```bash
docker compose -f docker-compose.yml -f docker-compose.local.yml down --remove-orphans
docker compose -f docker-compose.yml -f docker-compose.local.yml up -d --build --remove-orphans
docker compose -f docker-compose.yml -f docker-compose.local.yml logs -f
```
Run the validator locally
```bash
# Start the validator with debug.
af -vv validate
```
# Mining
IMPORTANT: you require a ***developer enabled account*** on Chutes to mine. Normal API keys cannot deploy chutes right now.
1. Set env vars.
```bash
# Copy .env and fill out validator items
cp .env.example .env
```
2. Miners need a chutes developer account ( `chutes.ai` )
```bash
chutes register
```
3. Register your miner to Affine (S120).
```bash
btcli subnet register --wallet.name <your cold> --wallet.hotkey <your hot>
```
4. Pull a model off the network.
```bash
af -vvv pull <uid to pull> --model_path <i.e. ./my_model>
```
5. Improve the model
```bash
... magic RL stuff ...
```
6. Push the model to your miner.
```bash
af -vvv push --coldkey <your cold> --hotkey <your hot> --model_path <i.e. ./my_model>
```
# SDK
Affine is also an SDK you can use to generate and evaluate models envs.
```python
import affine as af
# Optionally turn on logging
af.trace(); af.debug(); af.info()
# Get all miner info or only for UID =5
miners = await af.get_miners()
miner = await af.get_miners( 5 )
# Generate a SAT challenge
chal = await af.SAT.generate()
# Generate a bunch.
chals = await af.ABDUCTION().many( 10 )
chals = await af.DEDUCTION().many( 10 )
# Query the model directly.
# NOTE: A CHUTES_API_KEY .env value is required for this command.
response = await af.query( chal.prompt, model = miner.model )
# Evaluate the response
evaluation = chal.evaluate( response )
print( evaluation.score )
# Async generator of results from last 100 blocks.
async for res in af.rollouts(100):
print (res) # Result objects
```
|
matherchodhuuu/blockassist-bc-lightfooted_skilled_chameleon_1756793969
|
matherchodhuuu
| 2025-09-02T06:20:40Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"lightfooted skilled chameleon",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:20:35Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- lightfooted skilled chameleon
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
xinnn32/blockassist-bc-meek_winged_caterpillar_1756793915
|
xinnn32
| 2025-09-02T06:20:05Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"meek winged caterpillar",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:19:37Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- meek winged caterpillar
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Hodiee/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-jagged_insectivorous_wasp
|
Hodiee
| 2025-09-02T06:19:21Z | 26 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"rl-swarm",
"genrl-swarm",
"grpo",
"gensyn",
"I am jagged_insectivorous_wasp",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-08-31T04:09:50Z |
---
library_name: transformers
tags:
- rl-swarm
- genrl-swarm
- grpo
- gensyn
- I am jagged_insectivorous_wasp
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
Maarij-Aqeel/lunar_lander_RL
|
Maarij-Aqeel
| 2025-09-02T06:18:24Z | 0 | 0 |
stable-baselines3
|
[
"stable-baselines3",
"LunarLander-v3",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] |
reinforcement-learning
| 2025-09-02T05:49:19Z |
---
library_name: stable-baselines3
tags:
- LunarLander-v3
- deep-reinforcement-learning
- reinforcement-learning
- stable-baselines3
model-index:
- name: MLP_Policy
results:
- task:
type: reinforcement-learning
name: reinforcement-learning
dataset:
name: LunarLander-v3
type: LunarLander-v3
metrics:
- type: mean_reward
value: 245.48 +/- 38.84
name: mean_reward
verified: false
---
# **MLP_Policy** Agent playing **LunarLander-v3**
This is a trained model of a **MLP_Policy** agent playing **LunarLander-v3**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
## Demo 🎥
Here’s the agent in action:

```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
|
AnonymousCS/populism_classifier_391
|
AnonymousCS
| 2025-09-02T06:17:57Z | 2 | 0 |
transformers
|
[
"transformers",
"safetensors",
"rembert",
"text-classification",
"generated_from_trainer",
"base_model:google/rembert",
"base_model:finetune:google/rembert",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-classification
| 2025-08-31T16:53:45Z |
---
library_name: transformers
license: apache-2.0
base_model: google/rembert
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: populism_classifier_391
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# populism_classifier_391
This model is a fine-tuned version of [google/rembert](https://huggingface.co/google/rembert) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.5634
- Accuracy: 0.9523
- 1-f1: 0.0
- 1-recall: 0.0
- 1-precision: 0.0
- Balanced Acc: 0.5
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 20
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | 1-f1 | 1-recall | 1-precision | Balanced Acc |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:----:|:--------:|:-----------:|:------------:|
| 0.0854 | 1.0 | 3484 | 0.6875 | 0.9523 | 0.0 | 0.0 | 0.0 | 0.5 |
| 0.6854 | 2.0 | 6968 | 0.5804 | 0.9523 | 0.0 | 0.0 | 0.0 | 0.5 |
| 0.4985 | 3.0 | 10452 | 0.5638 | 0.9523 | 0.0 | 0.0 | 0.0 | 0.5 |
| 0.2701 | 4.0 | 13936 | 0.5563 | 0.9523 | 0.0 | 0.0 | 0.0 | 0.5 |
| 0.5109 | 5.0 | 17420 | 0.6045 | 0.9523 | 0.0 | 0.0 | 0.0 | 0.5 |
| 0.3593 | 6.0 | 20904 | 0.5634 | 0.9523 | 0.0 | 0.0 | 0.0 | 0.5 |
### Framework versions
- Transformers 4.46.3
- Pytorch 2.4.1+cu121
- Datasets 3.1.0
- Tokenizers 0.20.3
|
hZzy/mistral-7b-expo-7b-L2EXPO-25-08-try-2
|
hZzy
| 2025-09-02T06:17:55Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"generated_from_trainer",
"expo",
"trl",
"arxiv:2305.18290",
"base_model:hZzy/mistral-7b-sft-25-1",
"base_model:finetune:hZzy/mistral-7b-sft-25-1",
"endpoints_compatible",
"region:us"
] | null | 2025-08-04T13:49:27Z |
---
base_model: hZzy/mistral-7b-sft-25-1
library_name: transformers
model_name: mistral-7b-expo-7b-L2EXPO-25-08-try-2
tags:
- generated_from_trainer
- expo
- trl
licence: license
---
# Model Card for mistral-7b-expo-7b-L2EXPO-25-08-try-2
This model is a fine-tuned version of [hZzy/mistral-7b-sft-25-1](https://huggingface.co/hZzy/mistral-7b-sft-25-1).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="hZzy/mistral-7b-expo-7b-L2EXPO-25-08-try-2", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/zhiyuzha-university-of-florida/huggingface/runs/gfjbzq02)
This model was trained with DPO, a method introduced in [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://huggingface.co/papers/2305.18290).
### Framework versions
- TRL: 0.20.0
- Transformers: 4.54.1
- Pytorch: 2.7.0+cu128
- Datasets: 4.0.0
- Tokenizers: 0.21.4
## Citations
Cite DPO as:
```bibtex
@inproceedings{rafailov2023direct,
title = {{Direct Preference Optimization: Your Language Model is Secretly a Reward Model}},
author = {Rafael Rafailov and Archit Sharma and Eric Mitchell and Christopher D. Manning and Stefano Ermon and Chelsea Finn},
year = 2023,
booktitle = {Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, December 10 - 16, 2023},
url = {http://papers.nips.cc/paper_files/paper/2023/hash/a85b405ed65c6477a4fe8302b5e06ce7-Abstract-Conference.html},
editor = {Alice Oh and Tristan Naumann and Amir Globerson and Kate Saenko and Moritz Hardt and Sergey Levine},
}
```
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
zveroboyua/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-leaping_unseen_barracuda
|
zveroboyua
| 2025-09-02T06:17:27Z | 99 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"rl-swarm",
"genrl-swarm",
"grpo",
"gensyn",
"I am leaping_unseen_barracuda",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-08-30T14:06:03Z |
---
library_name: transformers
tags:
- rl-swarm
- genrl-swarm
- grpo
- gensyn
- I am leaping_unseen_barracuda
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
vendi11/blockassist-bc-placid_placid_llama_1756793803
|
vendi11
| 2025-09-02T06:17:26Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"placid placid llama",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:17:22Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- placid placid llama
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
chalbus2/FTA
|
chalbus2
| 2025-09-02T06:16:56Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"conversational",
"en",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-09-02T06:16:35Z |
---
base_model: unsloth/deepseek-r1-distill-llama-8b-unsloth-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- llama
license: apache-2.0
language:
- en
---
# Uploaded finetuned model
- **Developed by:** chalbus2
- **License:** apache-2.0
- **Finetuned from model :** unsloth/deepseek-r1-distill-llama-8b-unsloth-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
klmdr22/blockassist-bc-wild_loud_newt_1756793742
|
klmdr22
| 2025-09-02T06:16:26Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"wild loud newt",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:16:21Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- wild loud newt
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
luckeciano/Qwen-2.5-7B-GRPO-NoBaseline-FisherMaskToken-0.1-FisherMaskSentence-1e-4-v2_9634
|
luckeciano
| 2025-09-02T06:14:46Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"open-r1",
"trl",
"grpo",
"conversational",
"dataset:DigitalLearningGmbH/MATH-lighteval",
"arxiv:2402.03300",
"base_model:Qwen/Qwen2.5-Math-7B",
"base_model:finetune:Qwen/Qwen2.5-Math-7B",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-09-02T04:54:26Z |
---
base_model: Qwen/Qwen2.5-Math-7B
datasets: DigitalLearningGmbH/MATH-lighteval
library_name: transformers
model_name: Qwen-2.5-7B-GRPO-NoBaseline-FisherMaskToken-0.1-FisherMaskSentence-1e-4-v2_8146
tags:
- generated_from_trainer
- open-r1
- trl
- grpo
licence: license
---
# Model Card for Qwen-2.5-7B-GRPO-NoBaseline-FisherMaskToken-0.1-FisherMaskSentence-1e-4-v2_8146
This model is a fine-tuned version of [Qwen/Qwen2.5-Math-7B](https://huggingface.co/Qwen/Qwen2.5-Math-7B) on the [DigitalLearningGmbH/MATH-lighteval](https://huggingface.co/datasets/DigitalLearningGmbH/MATH-lighteval) dataset.
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="luckeciano/Qwen-2.5-7B-GRPO-NoBaseline-FisherMaskToken-0.1-FisherMaskSentence-1e-4-v2_8146", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/max-ent-llms/PolicyGradientStability/runs/wq85gqp4)
This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
### Framework versions
- TRL: 0.16.0.dev0
- Transformers: 4.49.0
- Pytorch: 2.5.1
- Datasets: 3.4.1
- Tokenizers: 0.21.2
## Citations
Cite GRPO as:
```bibtex
@article{zhihong2024deepseekmath,
title = {{DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models}},
author = {Zhihong Shao and Peiyi Wang and Qihao Zhu and Runxin Xu and Junxiao Song and Mingchuan Zhang and Y. K. Li and Y. Wu and Daya Guo},
year = 2024,
eprint = {arXiv:2402.03300},
}
```
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
matherchodhuuu/blockassist-bc-lightfooted_skilled_chameleon_1756793589
|
matherchodhuuu
| 2025-09-02T06:14:24Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"lightfooted skilled chameleon",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:14:20Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- lightfooted skilled chameleon
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
shivangi596/medgemma-4b-it-sft-lora-crc100k
|
shivangi596
| 2025-09-02T06:14:10Z | 0 | 0 |
transformers
|
[
"transformers",
"tensorboard",
"safetensors",
"generated_from_trainer",
"trl",
"sft",
"base_model:google/medgemma-4b-it",
"base_model:finetune:google/medgemma-4b-it",
"endpoints_compatible",
"region:us"
] | null | 2025-08-06T05:56:26Z |
---
base_model: google/medgemma-4b-it
library_name: transformers
model_name: medgemma-4b-it-sft-lora-crc100k
tags:
- generated_from_trainer
- trl
- sft
licence: license
---
# Model Card for medgemma-4b-it-sft-lora-crc100k
This model is a fine-tuned version of [google/medgemma-4b-it](https://huggingface.co/google/medgemma-4b-it).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="shivangi596/medgemma-4b-it-sft-lora-crc100k", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
This model was trained with SFT.
### Framework versions
- TRL: 0.22.1
- Transformers: 4.56.0
- Pytorch: 2.8.0+cu126
- Datasets: 4.0.0
- Tokenizers: 0.22.0
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
aXsalll/blockassist-bc-chattering_galloping_ape_1756793535
|
aXsalll
| 2025-09-02T06:13:38Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"chattering galloping ape",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:12:43Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- chattering galloping ape
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Rudra-madlads/blockassist-bc-jumping_swift_gazelle_1756793524
|
Rudra-madlads
| 2025-09-02T06:12:53Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"jumping swift gazelle",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:12:37Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- jumping swift gazelle
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
jephiqx/blockassist-bc-roaring_tawny_dolphin_1756791736
|
jephiqx
| 2025-09-02T06:12:37Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"roaring tawny dolphin",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:08:49Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- roaring tawny dolphin
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Novachrono93/NMDreamcore
|
Novachrono93
| 2025-09-02T06:11:42Z | 0 | 0 |
diffusers
|
[
"diffusers",
"text-to-image",
"lora",
"template:diffusion-lora",
"base_model:dhead/wai-nsfw-illustrious-sdxl-v140-sdxl",
"base_model:adapter:dhead/wai-nsfw-illustrious-sdxl-v140-sdxl",
"region:us"
] |
text-to-image
| 2025-09-02T06:09:14Z |
---
tags:
- text-to-image
- lora
- diffusers
- template:diffusion-lora
widget:
- output:
url: images/1000041480.jpg
text: "UNICODE\0\0{\0\"\02\03\0\"\0:\0{\0\"\0c\0l\0a\0s\0s\0_\0t\0y\0p\0e\0\"\0:\0\"\0U\0p\0s\0c\0a\0l\0e\0M\0o\0d\0e\0l\0L\0o\0a\0d\0e\0r\0\"\0,\0\"\0i\0n\0p\0u\0t\0s\0\"\0:\0{\0\"\0m\0o\0d\0e\0l\0_\0n\0a\0m\0e\0\"\0:\0\"\0u\0r\0n\0:\0a\0i\0r\0:\0o\0t\0h\0e\0r\0:\0u\0p\0s\0c\0a\0l\0e\0r\0:\0c\0i\0v\0i\0t\0a\0i\0:\01\04\07\07\05\09\0@\01\06\04\08\02\01\0\"\0}\0,\0\"\0_\0m\0e\0t\0a\0\"\0:\0{\0\"\0t\0i\0t\0l\0e\0\"\0:\0\"\0L\0o\0a\0d\0 \0U\0p\0s\0c\0a\0l\0e\0 \0M\0o\0d\0e\0l\0\"\0}\0}\0,\0\"\02\06\0\"\0:\0{\0\"\0c\0l\0a\0s\0s\0_\0t\0y\0p\0e\0\"\0:\0\"\0L\0o\0a\0d\0I\0m\0a\0g\0e\0\"\0,\0\"\0i\0n\0p\0u\0t\0s\0\"\0:\0{\0\"\0i\0m\0a\0g\0e\0\"\0:\0\"\0h\0t\0t\0p\0s\0:\0/\0/\0o\0r\0c\0h\0e\0s\0t\0r\0a\0t\0i\0o\0n\0.\0c\0i\0v\0i\0t\0a\0i\0.\0c\0o\0m\0/\0v\02\0/\0c\0o\0n\0s\0u\0m\0e\0r\0/\0b\0l\0o\0b\0s\0/\0K\0G\0K\06\0Z\0J\00\0F\00\0Y\0C\0Q\0W\0T\0G\0M\0E\0K\0M\0B\0H\0Q\0K\0K\0R\00\0\"\0,\0\"\0u\0p\0l\0o\0a\0d\0\"\0:\0\"\0i\0m\0a\0g\0e\0\"\0}\0,\0\"\0_\0m\0e\0t\0a\0\"\0:\0{\0\"\0t\0i\0t\0l\0e\0\"\0:\0\"\0L\0o\0a\0d\0 \0I\0m\0a\0g\0e\0\"\0}\0}\0,\0\"\02\02\0\"\0:\0{\0\"\0c\0l\0a\0s\0s\0_\0t\0y\0p\0e\0\"\0:\0\"\0I\0m\0a\0g\0e\0U\0p\0s\0c\0a\0l\0e\0W\0i\0t\0h\0M\0o\0d\0e\0l\0\"\0,\0\"\0i\0n\0p\0u\0t\0s\0\"\0:\0{\0\"\0u\0p\0s\0c\0a\0l\0e\0_\0m\0o\0d\0e\0l\0\"\0:\0[\0\"\02\03\0\"\0,\00\0]\0,\0\"\0i\0m\0a\0g\0e\0\"\0:\0[\0\"\02\06\0\"\0,\00\0]\0}\0,\0\"\0_\0m\0e\0t\0a\0\"\0:\0{\0\"\0t\0i\0t\0l\0e\0\"\0:\0\"\0U\0p\0s\0c\0a\0l\0e\0 \0I\0m\0a\0g\0e\0 \0(\0u\0s\0i\0n\0g\0 \0M\0o\0d\0e\0l\0)\0\"\0}\0}\0,\0\"\02\04\0\"\0:\0{\0\"\0c\0l\0a\0s\0s\0_\0t\0y\0p\0e\0\"\0:\0\"\0I\0m\0a\0g\0e\0S\0c\0a\0l\0e\0\"\0,\0\"\0i\0n\0p\0u\0t\0s\0\"\0:\0{\0\"\0u\0p\0s\0c\0a\0l\0e\0_\0m\0e\0t\0h\0o\0d\0\"\0:\0\"\0b\0i\0l\0i\0n\0e\0a\0r\0\"\0,\0\"\0c\0r\0o\0p\0\"\0:\0\"\0d\0i\0s\0a\0b\0l\0e\0d\0\"\0,\0\"\0w\0i\0d\0t\0h\0\"\0:\02\02\04\00\0,\0\"\0h\0e\0i\0g\0h\0t\0\"\0:\03\08\04\00\0,\0\"\0i\0m\0a\0g\0e\0\"\0:\0[\0\"\02\02\0\"\0,\00\0]\0}\0,\0\"\0_\0m\0e\0t\0a\0\"\0:\0{\0\"\0t\0i\0t\0l\0e\0\"\0:\0\"\0U\0p\0s\0c\0a\0l\0e\0 \0I\0m\0a\0g\0e\0\"\0}\0}\0,\0\"\01\02\0\"\0:\0{\0\"\0c\0l\0a\0s\0s\0_\0t\0y\0p\0e\0\"\0:\0\"\0S\0a\0v\0e\0I\0m\0a\0g\0e\0\"\0,\0\"\0i\0n\0p\0u\0t\0s\0\"\0:\0{\0\"\0f\0i\0l\0e\0n\0a\0m\0e\0_\0p\0r\0e\0f\0i\0x\0\"\0:\0\"\0C\0o\0m\0f\0y\0U\0I\0\"\0,\0\"\0i\0m\0a\0g\0e\0s\0\"\0:\0[\0\"\02\04\0\"\0,\00\0]\0}\0,\0\"\0_\0m\0e\0t\0a\0\"\0:\0{\0\"\0t\0i\0t\0l\0e\0\"\0:\0\"\0S\0a\0v\0e\0 \0I\0m\0a\0g\0e\0\"\0}\0}\0,\0\"\0e\0x\0t\0r\0a\0\"\0:\0{\0\"\0a\0i\0r\0s\0\"\0:\0[\0\"\0u\0r\0n\0:\0a\0i\0r\0:\0o\0t\0h\0e\0r\0:\0u\0p\0s\0c\0a\0l\0e\0r\0:\0c\0i\0v\0i\0t\0a\0i\0:\01\04\07\07\05\09\0@\01\06\04\08\02\01\0\"\0]\0}\0,\0\"\0e\0x\0t\0r\0a\0M\0e\0t\0a\0d\0a\0t\0a\0\"\0:\0\"\0{\0\\\0u\00\00\02\02\0p\0r\0o\0m\0p\0t\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0l\0a\0z\0y\0p\0o\0s\0,\0 \0b\0e\0s\0t\0 \0q\0u\0a\0l\0i\0t\0y\0,\0 \0r\0e\0f\0i\0n\0e\0d\0 \0d\0e\0t\0a\0i\0l\0e\0d\0,\0 \0h\0i\0g\0h\0 \0q\0u\0a\0l\0i\0t\0y\0,\0 \0h\0i\0g\0h\0r\0e\0s\0,\0 \0c\0i\0n\0e\0m\0a\0t\0i\0c\0 \0l\0i\0g\0h\0t\0i\0n\0g\0,\0 \0a\0b\0s\0u\0r\0d\0r\0e\0s\0,\0 \08\0k\0,\0 \0d\0e\0t\0a\0i\0l\0e\0d\0,\0 \0h\0i\0g\0h\0e\0s\0t\0 \0d\0e\0f\0i\0n\0i\0t\0i\0o\0n\0,\0 \0h\0i\0g\0h\0 \0r\0e\0s\0o\0l\0u\0t\0i\0o\0n\0,\0 \0s\0c\0e\0n\0e\0r\0y\0,\0 \0s\0o\0u\0r\0c\0e\0_\0a\0n\0i\0m\0e\0,\0 \0a\0b\0s\0u\0r\0d\0_\0r\0e\0s\0,\0 \0m\0a\0s\0t\0e\0r\0p\0i\0e\0c\0e\0,\0 \0b\0e\0s\0t\0 \0q\0u\0a\0l\0i\0t\0y\0,\0h\0i\0g\0h\0 \0q\0u\0a\0l\0i\0t\0y\0,\0 \0n\0e\0w\0e\0s\0t\0,\0 \0h\0i\0g\0h\0r\0e\0s\0,\08\0K\0,\0H\0D\0R\0,\0a\0b\0s\0u\0r\0d\0r\0e\0s\0,\04\0u\0r\04\0_\0i\0l\0l\0u\0,\0s\0h\0i\0n\0y\0,\0 \04\0u\0r\04\0_\0i\0l\0l\0u\0,\01\0g\0i\0r\0l\0,\0 \0h\0a\0i\0r\0 \0o\0r\0n\0a\0m\0e\0n\0t\0,\0 \0w\0e\0a\0p\0o\0n\0,\0 \0h\0a\0i\0r\0 \0f\0l\0o\0w\0e\0r\0,\0 \0s\0w\0o\0r\0d\0,\0 \0h\0o\0l\0d\0i\0n\0g\0,\0 \0f\0l\0o\0w\0e\0r\0,\0 \0j\0a\0p\0a\0n\0e\0s\0e\0 \0c\0l\0o\0t\0h\0e\0s\0,\0 \0s\0o\0l\0o\0,\0 \0k\0i\0m\0o\0n\0o\0,\0 \0b\0l\0a\0c\0k\0 \0h\0a\0i\0r\0,\0 \0l\0o\0o\0k\0i\0n\0g\0 \0a\0t\0 \0v\0i\0e\0w\0e\0r\0,\0 \0h\0o\0l\0d\0i\0n\0g\0 \0w\0e\0a\0p\0o\0n\0,\0 \0h\0o\0l\0d\0i\0n\0g\0 \0s\0w\0o\0r\0d\0,\0 \0f\0l\0o\0r\0a\0l\0 \0p\0r\0i\0n\0t\0,\0 \0k\0a\0t\0a\0n\0a\0,\0 \0r\0i\0n\0g\0,\0 \0b\0a\0n\0g\0s\0,\0 \0j\0e\0w\0e\0l\0r\0y\0,\0 \0b\0l\0u\0e\0 \0e\0y\0e\0s\0,\0 \0l\0o\0n\0g\0 \0h\0a\0i\0r\0,\0 \0c\0l\0o\0s\0e\0d\0 \0m\0o\0u\0t\0h\0,\0 \0b\0l\0u\0s\0h\0,\0 \0g\0l\0i\0n\0t\0,\0 \0p\0e\0t\0a\0l\0s\0,\0 \0l\0o\0n\0g\0 \0s\0l\0e\0e\0v\0e\0s\0,\0 \0r\0e\0d\0 \0k\0i\0m\0o\0n\0o\0,\0 \0b\0l\0a\0c\0k\0 \0b\0a\0c\0k\0g\0r\0o\0u\0n\0d\0,\0 \0w\0i\0d\0e\0 \0s\0l\0e\0e\0v\0e\0s\0,\0 \0b\0o\0o\0t\0s\0,\0 \0o\0b\0i\0,\0m\0o\0t\0i\0o\0n\0 \0b\0l\0u\0r\0,\0 \0d\0y\0n\0a\0m\0i\0c\0 \0p\0o\0s\0e\0,\0 \0d\0y\0n\0a\0m\0i\0c\0 \0c\0o\0m\0p\0o\0s\0i\0t\0i\0o\0n\0,\0d\0u\0t\0c\0h\0 \0a\0n\0g\0l\0e\0,\0 \0d\0e\0p\0t\0h\0 \0o\0f\0 \0f\0i\0e\0l\0d\0,\0w\0h\0i\0t\0e\0_\0t\0h\0e\0m\0e\0,\0d\0e\0t\0a\0i\0l\0e\0d\0 \0b\0a\0c\0k\0g\0r\0o\0u\0d\0,\0f\0o\0r\0e\0s\0h\0o\0r\0t\0e\0n\0i\0n\0g\0,\0b\0l\0u\0r\0r\0y\0 \0e\0d\0g\0e\0s\0,\0(\0(\04\0k\0,\08\0k\0,\0U\0l\0t\0r\0a\0 \0H\0D\0)\0)\0,\0 \0(\0(\0M\0a\0s\0t\0e\0r\0p\0i\0e\0c\0e\0 \0:\01\0.\02\0)\0)\0,\0(\0(\0B\0e\0s\0t\0 \0q\0u\0a\0l\0i\0t\0y\0 \0:\01\0.\02\0)\0)\0,\0(\0(\0D\0e\0t\0a\0i\0l\0e\0d\0 \0:\01\0.\04\0)\0)\0,\0(\0(\0d\0y\0n\0a\0m\0i\0c\0 \0p\0o\0s\0e\0:\01\0.\04\0)\0)\0,\0(\0(\0D\0e\0t\0a\0i\0l\0e\0d\0 \0b\0a\0c\0k\0g\0r\0o\0u\0n\0d\0 \0:\01\0.\04\0)\0)\0,\0 \0d\0i\0g\0i\0t\0a\0l\0 \0a\0r\0t\0,\0 \0s\0u\0r\0r\0e\0a\0l\0,\0 \0d\0y\0n\0a\0m\0i\0c\0 \0c\0o\0m\0p\0o\0s\0i\0t\0i\0o\0n\0,\0 \0b\0e\0a\0u\0t\0i\0f\0u\0l\0 \0a\0t\0m\0o\0s\0p\0h\0e\0r\0e\0,\0 \0F\0l\0a\0t\0l\0i\0n\0e\0,\0 \0F\0l\0a\0t\0 \0v\0e\0c\0t\0o\0r\0 \0i\0l\0l\0u\0s\0t\0r\0a\0t\0i\0o\0n\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0n\0e\0g\0a\0t\0i\0v\0e\0P\0r\0o\0m\0p\0t\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0l\0o\0w\0 \0q\0u\0a\0l\0i\0t\0y\0,\0 \0w\0o\0r\0s\0t\0 \0q\0u\0a\0l\0i\0t\0y\0,\0 \0n\0o\0r\0m\0a\0l\0 \0q\0u\0a\0l\0i\0t\0y\0,\0 \0j\0p\0e\0g\0 \0a\0r\0t\0i\0f\0a\0c\0t\0s\0,\0 \0s\0i\0g\0n\0a\0t\0u\0r\0e\0,\0 \0w\0a\0t\0e\0r\0m\0a\0r\0k\0,\0 \0t\0e\0x\0t\0,\0 \0e\0r\0r\0o\0r\0,\0 \0b\0l\0u\0r\0r\0y\0,\0 \0u\0g\0l\0y\0,\0 \0p\0o\0o\0r\0l\0y\0 \0d\0r\0a\0w\0n\0,\0 \0s\0k\0e\0t\0c\0h\0,\0 \0m\0o\0n\0o\0c\0h\0r\0o\0m\0e\0,\0 \0m\0u\0t\0a\0t\0e\0d\0,\0 \0e\0x\0t\0r\0a\0 \0l\0i\0m\0b\0s\0,\0 \0d\0e\0f\0o\0r\0m\0e\0d\0,\0 \0d\0i\0s\0t\0o\0r\0t\0e\0d\0 \0f\0a\0c\0e\0,\0 \0b\0a\0d\0 \0p\0r\0o\0p\0o\0r\0t\0i\0o\0n\0s\0,\0 \0o\0u\0t\0 \0o\0f\0 \0f\0o\0c\0u\0s\0,\0 \0c\0l\0o\0n\0e\0d\0 \0f\0a\0c\0e\0,\0 \0m\0u\0t\0a\0t\0e\0d\0 \0h\0a\0n\0d\0s\0,\0 \0d\0i\0s\0f\0i\0g\0u\0r\0e\0d\0,\0 \0e\0x\0t\0r\0a\0 \0a\0r\0m\0s\0,\0 \0e\0x\0t\0r\0a\0 \0l\0e\0g\0s\0,\0 \0e\0x\0t\0r\0a\0 \0f\0i\0n\0g\0e\0r\0s\0,\0 \0m\0i\0s\0s\0i\0n\0g\0 \0a\0r\0m\0s\0,\0 \0m\0i\0s\0s\0i\0n\0g\0 \0l\0e\0g\0s\0,\0 \0m\0i\0s\0s\0i\0n\0g\0 \0f\0i\0n\0g\0e\0r\0s\0,\0 \0d\0i\0s\0c\0o\0n\0n\0e\0c\0t\0e\0d\0 \0l\0i\0m\0b\0s\0,\0 \0l\0o\0n\0g\0 \0n\0e\0c\0k\0,\0 \0b\0a\0d\0 \0a\0n\0a\0t\0o\0m\0y\0,\0 \0b\0a\0d\0 \0h\0a\0n\0d\0s\0,\0 \0p\0o\0o\0r\0l\0y\0 \0d\0r\0a\0w\0n\0 \0h\0a\0n\0d\0s\0,\0 \0b\0a\0d\0 \0f\0e\0e\0t\0,\0 \0e\0x\0t\0r\0a\0 \0e\0y\0e\0s\0,\0 \0g\0l\0i\0t\0c\0h\0e\0d\0 \0e\0y\0e\0s\0,\0 \0m\0u\0t\0a\0t\0e\0d\0 \0e\0y\0e\0s\0,\0 \0c\0r\0o\0s\0s\0-\0e\0y\0e\0,\0 \0a\0s\0y\0m\0m\0e\0t\0r\0i\0c\0a\0l\0 \0e\0y\0e\0s\0,\0 \0r\0o\0u\0g\0h\0 \0s\0k\0e\0t\0c\0h\0,\0 \0d\0i\0s\0c\0o\0l\0o\0r\0e\0d\0,\0 \0l\0a\0z\0y\0n\0e\0g\0,\0 \0l\0a\0z\0y\0h\0a\0n\0d\0 \0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0c\0f\0g\0S\0c\0a\0l\0e\0\\\0u\00\00\02\02\0:\04\0,\0\\\0u\00\00\02\02\0s\0a\0m\0p\0l\0e\0r\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0E\0u\0l\0e\0r\0 \0a\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0s\0e\0e\0d\0\\\0u\00\00\02\02\0:\01\03\00\04\02\04\00\07\09\00\0,\0\\\0u\00\00\02\02\0c\0l\0i\0p\0S\0k\0i\0p\0\\\0u\00\00\02\02\0:\02\0,\0\\\0u\00\00\02\02\0s\0t\0e\0p\0s\0\\\0u\00\00\02\02\0:\02\05\0,\0\\\0u\00\00\02\02\0q\0u\0a\0n\0t\0i\0t\0y\0\\\0u\00\00\02\02\0:\01\0,\0\\\0u\00\00\02\02\0a\0s\0p\0e\0c\0t\0R\0a\0t\0i\0o\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\01\03\0:\01\09\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0b\0a\0s\0e\0M\0o\0d\0e\0l\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0I\0l\0l\0u\0s\0t\0r\0i\0o\0u\0s\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0w\0i\0d\0t\0h\0\\\0u\00\00\02\02\0:\07\03\06\0,\0\\\0u\00\00\02\02\0h\0e\0i\0g\0h\0t\0\\\0u\00\00\02\02\0:\01\03\00\08\0,\0\\\0u\00\00\02\02\0d\0e\0n\0o\0i\0s\0e\0\\\0u\00\00\02\02\0:\00\0.\05\05\0,\0\\\0u\00\00\02\02\0w\0o\0r\0k\0f\0l\0o\0w\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0i\0m\0g\02\0i\0m\0g\0-\0h\0i\0r\0e\0s\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0e\0x\0p\0e\0r\0i\0m\0e\0n\0t\0a\0l\0\\\0u\00\00\02\02\0:\0f\0a\0l\0s\0e\0,\0\\\0u\00\00\02\02\0p\0r\0i\0o\0r\0i\0t\0y\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0l\0o\0w\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0s\0o\0u\0r\0c\0e\0I\0m\0a\0g\0e\0\\\0u\00\00\02\02\0:\0{\0\\\0u\00\00\02\02\0u\0r\0l\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0h\0t\0t\0p\0s\0:\0/\0/\0o\0r\0c\0h\0e\0s\0t\0r\0a\0t\0i\0o\0n\0.\0c\0i\0v\0i\0t\0a\0i\0.\0c\0o\0m\0/\0v\02\0/\0c\0o\0n\0s\0u\0m\0e\0r\0/\0b\0l\0o\0b\0s\0/\01\02\08\08\06\06\02\01\05\03\09\05\06\09\09\01\07\05\06\08\03\05\08\09\02\09\05\09\04\08\09\04\02\01\05\04\05\02\02\0.\0j\0p\0g\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0w\0i\0d\0t\0h\0\\\0u\00\00\02\02\0:\07\03\06\0,\0\\\0u\00\00\02\02\0h\0e\0i\0g\0h\0t\0\\\0u\00\00\02\02\0:\01\03\00\08\0}\0,\0\\\0u\00\00\02\02\0d\0i\0s\0a\0b\0l\0e\0P\0o\0i\0\\\0u\00\00\02\02\0:\0t\0r\0u\0e\0,\0\\\0u\00\00\02\02\0p\0r\0o\0c\0e\0s\0s\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0i\0m\0g\02\0i\0m\0g\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0r\0e\0s\0o\0u\0r\0c\0e\0s\0\\\0u\00\00\02\02\0:\0[\0{\0\\\0u\00\00\02\02\0m\0o\0d\0e\0l\0V\0e\0r\0s\0i\0o\0n\0I\0d\0\\\0u\00\00\02\02\0:\01\07\07\02\06\04\05\0,\0\\\0u\00\00\02\02\0s\0t\0r\0e\0n\0g\0t\0h\0\\\0u\00\00\02\02\0:\01\0}\0,\0{\0\\\0u\00\00\02\02\0m\0o\0d\0e\0l\0V\0e\0r\0s\0i\0o\0n\0I\0d\0\\\0u\00\00\02\02\0:\01\08\00\09\05\07\05\0,\0\\\0u\00\00\02\02\0s\0t\0r\0e\0n\0g\0t\0h\0\\\0u\00\00\02\02\0:\00\0.\02\0}\0,\0{\0\\\0u\00\00\02\02\0m\0o\0d\0e\0l\0V\0e\0r\0s\0i\0o\0n\0I\0d\0\\\0u\00\00\02\02\0:\01\08\03\03\01\05\07\0,\0\\\0u\00\00\02\02\0s\0t\0r\0e\0n\0g\0t\0h\0\\\0u\00\00\02\02\0:\01\0}\0,\0{\0\\\0u\00\00\02\02\0m\0o\0d\0e\0l\0V\0e\0r\0s\0i\0o\0n\0I\0d\0\\\0u\00\00\02\02\0:\01\08\06\00\07\04\07\0,\0\\\0u\00\00\02\02\0s\0t\0r\0e\0n\0g\0t\0h\0\\\0u\00\00\02\02\0:\01\0}\0,\0{\0\\\0u\00\00\02\02\0m\0o\0d\0e\0l\0V\0e\0r\0s\0i\0o\0n\0I\0d\0\\\0u\00\00\02\02\0:\01\05\05\00\08\04\00\0,\0\\\0u\00\00\02\02\0s\0t\0r\0e\0n\0g\0t\0h\0\\\0u\00\00\02\02\0:\01\0}\0,\0{\0\\\0u\00\00\02\02\0m\0o\0d\0e\0l\0V\0e\0r\0s\0i\0o\0n\0I\0d\0\\\0u\00\00\02\02\0:\01\07\06\06\06\09\01\0,\0\\\0u\00\00\02\02\0s\0t\0r\0e\0n\0g\0t\0h\0\\\0u\00\00\02\02\0:\00\0.\08\0}\0]\0}\0\"\0}"
base_model: dhead/wai-nsfw-illustrious-sdxl-v140-sdxl
instance_prompt: null
---
# NM Dreamcore
<Gallery />
## Download model
[Download](/Novachrono93/NMDreamcore/tree/main) them in the Files & versions tab.
|
AhunInteligence/w2v-bert-2.0-amharic-finetuned-ctc
|
AhunInteligence
| 2025-09-02T06:11:22Z | 0 | 0 |
transformers
|
[
"transformers",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | 2025-09-02T06:11:19Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
klmdr22/blockassist-bc-wild_loud_newt_1756793381
|
klmdr22
| 2025-09-02T06:10:24Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"wild loud newt",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:10:20Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- wild loud newt
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
EJ024/autotrain-wwea6-a16mi
|
EJ024
| 2025-09-02T06:09:05Z | 0 | 0 |
transformers
|
[
"transformers",
"tensorboard",
"safetensors",
"xlm-roberta",
"text-classification",
"autotrain",
"base_model:Davlan/xlm-roberta-base-ner-hrl",
"base_model:finetune:Davlan/xlm-roberta-base-ner-hrl",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-classification
| 2025-09-02T06:06:34Z |
---
library_name: transformers
tags:
- autotrain
- text-classification
base_model: Davlan/xlm-roberta-base-ner-hrl
widget:
- text: "I love AutoTrain"
---
# Model Trained Using AutoTrain
- Problem type: Text Classification
## Validation Metrics
loss: 2.372678518295288
f1_macro: 0.03305785123966942
f1_micro: 0.14285714285714285
f1_weighted: 0.05194805194805195
precision_macro: 0.0202020202020202
precision_micro: 0.14285714285714285
precision_weighted: 0.031746031746031744
recall_macro: 0.09090909090909091
recall_micro: 0.14285714285714285
recall_weighted: 0.14285714285714285
accuracy: 0.14285714285714285
|
liukevin666/blockassist-bc-yawning_striped_cassowary_1756793247
|
liukevin666
| 2025-09-02T06:08:29Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"yawning striped cassowary",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:08:22Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- yawning striped cassowary
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
zhiqing/Hunyuan-MT-Chimera-7B-INT8
|
zhiqing
| 2025-09-02T06:08:21Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"hunyuan_v1_dense",
"text-generation",
"translation",
"base_model:tencent/Hunyuan-MT-Chimera-7B",
"base_model:quantized:tencent/Hunyuan-MT-Chimera-7B",
"autotrain_compatible",
"endpoints_compatible",
"compressed-tensors",
"region:us"
] |
translation
| 2025-09-02T04:31:24Z |
---
library_name: transformers
tags:
- translation
base_model:
- tencent/Hunyuan-MT-Chimera-7B
---
<p align="center">
<img src="https://dscache.tencent-cloud.cn/upload/uploader/hunyuan-64b418fd052c033b228e04bc77bbc4b54fd7f5bc.png" width="400"/> <br>
</p><p></p>
## Model Introduction
The Hunyuan Translation Model comprises a translation model, Hunyuan-MT-7B, and an ensemble model, Hunyuan-MT-Chimera. The translation model is used to translate source text into the target language, while the ensemble model integrates multiple translation outputs to produce a higher-quality result. It primarily supports mutual translation among 33 languages, including five ethnic minority languages in China.
### Key Features and Advantages
- In the WMT25 competition, the model achieved first place in 30 out of the 31 language categories it participated in.
- Hunyuan-MT-7B achieves industry-leading performance among models of comparable scale
- Hunyuan-MT-Chimera-7B is the industry’s first open-source translation ensemble model, elevating translation quality to a new level
- A comprehensive training framework for translation models has been proposed, spanning from pretrain → cross-lingual pretraining (CPT) → supervised fine-tuning (SFT) → translation enhancement → ensemble refinement, achieving state-of-the-art (SOTA) results for models of similar size
## Related News
* 2025.9.1 We have open-sourced **Hunyuan-MT-7B** , **Hunyuan-MT-Chimera-7B** on Hugging Face.
<br>
## Prompts
### Prompt Template for ZH<=>XX Translation.
```
把下面的文本翻译成<target_language>,不要额外解释。
<source_text>
```
### Prompt Template for XX<=>XX Translation, excluding ZH<=>XX.
```
Translate the following segment into <target_language>, without additional explanation.
<source_text>
```
### Prompt Template for Hunyuan-MT-Chimera-7B-INT8
```
Analyze the following multiple <target_language> translations of the <source_language> segment surrounded in triple backticks and generate a single refined <target_language> translation. Only output the refined translation, do not explain.
The <source_language> segment:
```<source_text>```
The multiple <target_language> translations:
1. ```<translated_text1>```
2. ```<translated_text2>```
3. ```<translated_text3>```
4. ```<translated_text4>```
5. ```<translated_text5>```
6. ```<translated_text6>```
```
### Use with transformers
First, please install transformers, recommends v4.56.0
```SHELL
pip install transformers==4.56.0
```
The following code snippet shows how to use the transformers library to load and apply the model.
*!!! If you want to load fp8 model with transformers, you need to change the name"ignored_layers" in config.json to "ignore" and upgrade the compressed-tensors to compressed-tensors-0.11.0.*
we use zhiqing/Hunyuan-MT-Chimera-7B-INT8 for example
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import os
model_name_or_path = "zhiqing/Hunyuan-MT-Chimera-7B-INT8"
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path)
model = AutoModelForCausalLM.from_pretrained(model_name_or_path, device_map="auto") # You may want to use bfloat16 and/or move to GPU here
messages = [
{"role": "user", "content": "Translate the following segment into Chinese, without additional explanation.\n\nIt’s on the house."},
]
tokenized_chat = tokenizer.apply_chat_template(
messages,
tokenize=True
add_generation_prompt=False,
return_tensors="pt"
)
outputs = model.generate(tokenized_chat.to(model.device), max_new_tokens=2048)
output_text = tokenizer.decode(outputs[0])
```
### Use with vLLM
```SHELL
pip install vllm --upgrade
```
```SHELL
vllm serve zhiqing/Hunyuan-MT-Chimera-7B-INT8
```
We recommend using the following set of parameters for inference. Note that our model does not have the default system_prompt.
```json
{
"top_k": 20,
"top_p": 0.6,
"repetition_penalty": 1.05,
"temperature": 0.7
}
```
Citing Hunyuan-MT:
```bibtex
@misc{hunyuanmt2025,
title={Hunyuan-MT Technical Report},
author={Mao Zheng, Zheng Li, Bingxin Qu, Mingyang Song, Yang Du, Mingrui Sun, Di Wang, Tao Chen, Jiaqi Zhu, Xingwu Sun, Yufei Wang, Can Xu, Chen Li, Kai Wang, Decheng Wu},
howpublished={\url{https://github.com/Tencent-Hunyuan/Hunyuan-MT}},
year={2025}
}
```
|
Novachrono93/TrendCraft
|
Novachrono93
| 2025-09-02T06:07:45Z | 0 | 0 |
diffusers
|
[
"diffusers",
"text-to-image",
"lora",
"template:diffusion-lora",
"base_model:dhead/wai-nsfw-illustrious-sdxl-v140-sdxl",
"base_model:adapter:dhead/wai-nsfw-illustrious-sdxl-v140-sdxl",
"region:us"
] |
text-to-image
| 2025-09-02T06:06:39Z |
---
tags:
- text-to-image
- lora
- diffusers
- template:diffusion-lora
widget:
- output:
url: images/1000041480.jpg
text: "UNICODE\0\0{\0\"\02\03\0\"\0:\0{\0\"\0c\0l\0a\0s\0s\0_\0t\0y\0p\0e\0\"\0:\0\"\0U\0p\0s\0c\0a\0l\0e\0M\0o\0d\0e\0l\0L\0o\0a\0d\0e\0r\0\"\0,\0\"\0i\0n\0p\0u\0t\0s\0\"\0:\0{\0\"\0m\0o\0d\0e\0l\0_\0n\0a\0m\0e\0\"\0:\0\"\0u\0r\0n\0:\0a\0i\0r\0:\0o\0t\0h\0e\0r\0:\0u\0p\0s\0c\0a\0l\0e\0r\0:\0c\0i\0v\0i\0t\0a\0i\0:\01\04\07\07\05\09\0@\01\06\04\08\02\01\0\"\0}\0,\0\"\0_\0m\0e\0t\0a\0\"\0:\0{\0\"\0t\0i\0t\0l\0e\0\"\0:\0\"\0L\0o\0a\0d\0 \0U\0p\0s\0c\0a\0l\0e\0 \0M\0o\0d\0e\0l\0\"\0}\0}\0,\0\"\02\06\0\"\0:\0{\0\"\0c\0l\0a\0s\0s\0_\0t\0y\0p\0e\0\"\0:\0\"\0L\0o\0a\0d\0I\0m\0a\0g\0e\0\"\0,\0\"\0i\0n\0p\0u\0t\0s\0\"\0:\0{\0\"\0i\0m\0a\0g\0e\0\"\0:\0\"\0h\0t\0t\0p\0s\0:\0/\0/\0o\0r\0c\0h\0e\0s\0t\0r\0a\0t\0i\0o\0n\0.\0c\0i\0v\0i\0t\0a\0i\0.\0c\0o\0m\0/\0v\02\0/\0c\0o\0n\0s\0u\0m\0e\0r\0/\0b\0l\0o\0b\0s\0/\0K\0G\0K\06\0Z\0J\00\0F\00\0Y\0C\0Q\0W\0T\0G\0M\0E\0K\0M\0B\0H\0Q\0K\0K\0R\00\0\"\0,\0\"\0u\0p\0l\0o\0a\0d\0\"\0:\0\"\0i\0m\0a\0g\0e\0\"\0}\0,\0\"\0_\0m\0e\0t\0a\0\"\0:\0{\0\"\0t\0i\0t\0l\0e\0\"\0:\0\"\0L\0o\0a\0d\0 \0I\0m\0a\0g\0e\0\"\0}\0}\0,\0\"\02\02\0\"\0:\0{\0\"\0c\0l\0a\0s\0s\0_\0t\0y\0p\0e\0\"\0:\0\"\0I\0m\0a\0g\0e\0U\0p\0s\0c\0a\0l\0e\0W\0i\0t\0h\0M\0o\0d\0e\0l\0\"\0,\0\"\0i\0n\0p\0u\0t\0s\0\"\0:\0{\0\"\0u\0p\0s\0c\0a\0l\0e\0_\0m\0o\0d\0e\0l\0\"\0:\0[\0\"\02\03\0\"\0,\00\0]\0,\0\"\0i\0m\0a\0g\0e\0\"\0:\0[\0\"\02\06\0\"\0,\00\0]\0}\0,\0\"\0_\0m\0e\0t\0a\0\"\0:\0{\0\"\0t\0i\0t\0l\0e\0\"\0:\0\"\0U\0p\0s\0c\0a\0l\0e\0 \0I\0m\0a\0g\0e\0 \0(\0u\0s\0i\0n\0g\0 \0M\0o\0d\0e\0l\0)\0\"\0}\0}\0,\0\"\02\04\0\"\0:\0{\0\"\0c\0l\0a\0s\0s\0_\0t\0y\0p\0e\0\"\0:\0\"\0I\0m\0a\0g\0e\0S\0c\0a\0l\0e\0\"\0,\0\"\0i\0n\0p\0u\0t\0s\0\"\0:\0{\0\"\0u\0p\0s\0c\0a\0l\0e\0_\0m\0e\0t\0h\0o\0d\0\"\0:\0\"\0b\0i\0l\0i\0n\0e\0a\0r\0\"\0,\0\"\0c\0r\0o\0p\0\"\0:\0\"\0d\0i\0s\0a\0b\0l\0e\0d\0\"\0,\0\"\0w\0i\0d\0t\0h\0\"\0:\02\02\04\00\0,\0\"\0h\0e\0i\0g\0h\0t\0\"\0:\03\08\04\00\0,\0\"\0i\0m\0a\0g\0e\0\"\0:\0[\0\"\02\02\0\"\0,\00\0]\0}\0,\0\"\0_\0m\0e\0t\0a\0\"\0:\0{\0\"\0t\0i\0t\0l\0e\0\"\0:\0\"\0U\0p\0s\0c\0a\0l\0e\0 \0I\0m\0a\0g\0e\0\"\0}\0}\0,\0\"\01\02\0\"\0:\0{\0\"\0c\0l\0a\0s\0s\0_\0t\0y\0p\0e\0\"\0:\0\"\0S\0a\0v\0e\0I\0m\0a\0g\0e\0\"\0,\0\"\0i\0n\0p\0u\0t\0s\0\"\0:\0{\0\"\0f\0i\0l\0e\0n\0a\0m\0e\0_\0p\0r\0e\0f\0i\0x\0\"\0:\0\"\0C\0o\0m\0f\0y\0U\0I\0\"\0,\0\"\0i\0m\0a\0g\0e\0s\0\"\0:\0[\0\"\02\04\0\"\0,\00\0]\0}\0,\0\"\0_\0m\0e\0t\0a\0\"\0:\0{\0\"\0t\0i\0t\0l\0e\0\"\0:\0\"\0S\0a\0v\0e\0 \0I\0m\0a\0g\0e\0\"\0}\0}\0,\0\"\0e\0x\0t\0r\0a\0\"\0:\0{\0\"\0a\0i\0r\0s\0\"\0:\0[\0\"\0u\0r\0n\0:\0a\0i\0r\0:\0o\0t\0h\0e\0r\0:\0u\0p\0s\0c\0a\0l\0e\0r\0:\0c\0i\0v\0i\0t\0a\0i\0:\01\04\07\07\05\09\0@\01\06\04\08\02\01\0\"\0]\0}\0,\0\"\0e\0x\0t\0r\0a\0M\0e\0t\0a\0d\0a\0t\0a\0\"\0:\0\"\0{\0\\\0u\00\00\02\02\0p\0r\0o\0m\0p\0t\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0l\0a\0z\0y\0p\0o\0s\0,\0 \0b\0e\0s\0t\0 \0q\0u\0a\0l\0i\0t\0y\0,\0 \0r\0e\0f\0i\0n\0e\0d\0 \0d\0e\0t\0a\0i\0l\0e\0d\0,\0 \0h\0i\0g\0h\0 \0q\0u\0a\0l\0i\0t\0y\0,\0 \0h\0i\0g\0h\0r\0e\0s\0,\0 \0c\0i\0n\0e\0m\0a\0t\0i\0c\0 \0l\0i\0g\0h\0t\0i\0n\0g\0,\0 \0a\0b\0s\0u\0r\0d\0r\0e\0s\0,\0 \08\0k\0,\0 \0d\0e\0t\0a\0i\0l\0e\0d\0,\0 \0h\0i\0g\0h\0e\0s\0t\0 \0d\0e\0f\0i\0n\0i\0t\0i\0o\0n\0,\0 \0h\0i\0g\0h\0 \0r\0e\0s\0o\0l\0u\0t\0i\0o\0n\0,\0 \0s\0c\0e\0n\0e\0r\0y\0,\0 \0s\0o\0u\0r\0c\0e\0_\0a\0n\0i\0m\0e\0,\0 \0a\0b\0s\0u\0r\0d\0_\0r\0e\0s\0,\0 \0m\0a\0s\0t\0e\0r\0p\0i\0e\0c\0e\0,\0 \0b\0e\0s\0t\0 \0q\0u\0a\0l\0i\0t\0y\0,\0h\0i\0g\0h\0 \0q\0u\0a\0l\0i\0t\0y\0,\0 \0n\0e\0w\0e\0s\0t\0,\0 \0h\0i\0g\0h\0r\0e\0s\0,\08\0K\0,\0H\0D\0R\0,\0a\0b\0s\0u\0r\0d\0r\0e\0s\0,\04\0u\0r\04\0_\0i\0l\0l\0u\0,\0s\0h\0i\0n\0y\0,\0 \04\0u\0r\04\0_\0i\0l\0l\0u\0,\01\0g\0i\0r\0l\0,\0 \0h\0a\0i\0r\0 \0o\0r\0n\0a\0m\0e\0n\0t\0,\0 \0w\0e\0a\0p\0o\0n\0,\0 \0h\0a\0i\0r\0 \0f\0l\0o\0w\0e\0r\0,\0 \0s\0w\0o\0r\0d\0,\0 \0h\0o\0l\0d\0i\0n\0g\0,\0 \0f\0l\0o\0w\0e\0r\0,\0 \0j\0a\0p\0a\0n\0e\0s\0e\0 \0c\0l\0o\0t\0h\0e\0s\0,\0 \0s\0o\0l\0o\0,\0 \0k\0i\0m\0o\0n\0o\0,\0 \0b\0l\0a\0c\0k\0 \0h\0a\0i\0r\0,\0 \0l\0o\0o\0k\0i\0n\0g\0 \0a\0t\0 \0v\0i\0e\0w\0e\0r\0,\0 \0h\0o\0l\0d\0i\0n\0g\0 \0w\0e\0a\0p\0o\0n\0,\0 \0h\0o\0l\0d\0i\0n\0g\0 \0s\0w\0o\0r\0d\0,\0 \0f\0l\0o\0r\0a\0l\0 \0p\0r\0i\0n\0t\0,\0 \0k\0a\0t\0a\0n\0a\0,\0 \0r\0i\0n\0g\0,\0 \0b\0a\0n\0g\0s\0,\0 \0j\0e\0w\0e\0l\0r\0y\0,\0 \0b\0l\0u\0e\0 \0e\0y\0e\0s\0,\0 \0l\0o\0n\0g\0 \0h\0a\0i\0r\0,\0 \0c\0l\0o\0s\0e\0d\0 \0m\0o\0u\0t\0h\0,\0 \0b\0l\0u\0s\0h\0,\0 \0g\0l\0i\0n\0t\0,\0 \0p\0e\0t\0a\0l\0s\0,\0 \0l\0o\0n\0g\0 \0s\0l\0e\0e\0v\0e\0s\0,\0 \0r\0e\0d\0 \0k\0i\0m\0o\0n\0o\0,\0 \0b\0l\0a\0c\0k\0 \0b\0a\0c\0k\0g\0r\0o\0u\0n\0d\0,\0 \0w\0i\0d\0e\0 \0s\0l\0e\0e\0v\0e\0s\0,\0 \0b\0o\0o\0t\0s\0,\0 \0o\0b\0i\0,\0m\0o\0t\0i\0o\0n\0 \0b\0l\0u\0r\0,\0 \0d\0y\0n\0a\0m\0i\0c\0 \0p\0o\0s\0e\0,\0 \0d\0y\0n\0a\0m\0i\0c\0 \0c\0o\0m\0p\0o\0s\0i\0t\0i\0o\0n\0,\0d\0u\0t\0c\0h\0 \0a\0n\0g\0l\0e\0,\0 \0d\0e\0p\0t\0h\0 \0o\0f\0 \0f\0i\0e\0l\0d\0,\0w\0h\0i\0t\0e\0_\0t\0h\0e\0m\0e\0,\0d\0e\0t\0a\0i\0l\0e\0d\0 \0b\0a\0c\0k\0g\0r\0o\0u\0d\0,\0f\0o\0r\0e\0s\0h\0o\0r\0t\0e\0n\0i\0n\0g\0,\0b\0l\0u\0r\0r\0y\0 \0e\0d\0g\0e\0s\0,\0(\0(\04\0k\0,\08\0k\0,\0U\0l\0t\0r\0a\0 \0H\0D\0)\0)\0,\0 \0(\0(\0M\0a\0s\0t\0e\0r\0p\0i\0e\0c\0e\0 \0:\01\0.\02\0)\0)\0,\0(\0(\0B\0e\0s\0t\0 \0q\0u\0a\0l\0i\0t\0y\0 \0:\01\0.\02\0)\0)\0,\0(\0(\0D\0e\0t\0a\0i\0l\0e\0d\0 \0:\01\0.\04\0)\0)\0,\0(\0(\0d\0y\0n\0a\0m\0i\0c\0 \0p\0o\0s\0e\0:\01\0.\04\0)\0)\0,\0(\0(\0D\0e\0t\0a\0i\0l\0e\0d\0 \0b\0a\0c\0k\0g\0r\0o\0u\0n\0d\0 \0:\01\0.\04\0)\0)\0,\0 \0d\0i\0g\0i\0t\0a\0l\0 \0a\0r\0t\0,\0 \0s\0u\0r\0r\0e\0a\0l\0,\0 \0d\0y\0n\0a\0m\0i\0c\0 \0c\0o\0m\0p\0o\0s\0i\0t\0i\0o\0n\0,\0 \0b\0e\0a\0u\0t\0i\0f\0u\0l\0 \0a\0t\0m\0o\0s\0p\0h\0e\0r\0e\0,\0 \0F\0l\0a\0t\0l\0i\0n\0e\0,\0 \0F\0l\0a\0t\0 \0v\0e\0c\0t\0o\0r\0 \0i\0l\0l\0u\0s\0t\0r\0a\0t\0i\0o\0n\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0n\0e\0g\0a\0t\0i\0v\0e\0P\0r\0o\0m\0p\0t\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0l\0o\0w\0 \0q\0u\0a\0l\0i\0t\0y\0,\0 \0w\0o\0r\0s\0t\0 \0q\0u\0a\0l\0i\0t\0y\0,\0 \0n\0o\0r\0m\0a\0l\0 \0q\0u\0a\0l\0i\0t\0y\0,\0 \0j\0p\0e\0g\0 \0a\0r\0t\0i\0f\0a\0c\0t\0s\0,\0 \0s\0i\0g\0n\0a\0t\0u\0r\0e\0,\0 \0w\0a\0t\0e\0r\0m\0a\0r\0k\0,\0 \0t\0e\0x\0t\0,\0 \0e\0r\0r\0o\0r\0,\0 \0b\0l\0u\0r\0r\0y\0,\0 \0u\0g\0l\0y\0,\0 \0p\0o\0o\0r\0l\0y\0 \0d\0r\0a\0w\0n\0,\0 \0s\0k\0e\0t\0c\0h\0,\0 \0m\0o\0n\0o\0c\0h\0r\0o\0m\0e\0,\0 \0m\0u\0t\0a\0t\0e\0d\0,\0 \0e\0x\0t\0r\0a\0 \0l\0i\0m\0b\0s\0,\0 \0d\0e\0f\0o\0r\0m\0e\0d\0,\0 \0d\0i\0s\0t\0o\0r\0t\0e\0d\0 \0f\0a\0c\0e\0,\0 \0b\0a\0d\0 \0p\0r\0o\0p\0o\0r\0t\0i\0o\0n\0s\0,\0 \0o\0u\0t\0 \0o\0f\0 \0f\0o\0c\0u\0s\0,\0 \0c\0l\0o\0n\0e\0d\0 \0f\0a\0c\0e\0,\0 \0m\0u\0t\0a\0t\0e\0d\0 \0h\0a\0n\0d\0s\0,\0 \0d\0i\0s\0f\0i\0g\0u\0r\0e\0d\0,\0 \0e\0x\0t\0r\0a\0 \0a\0r\0m\0s\0,\0 \0e\0x\0t\0r\0a\0 \0l\0e\0g\0s\0,\0 \0e\0x\0t\0r\0a\0 \0f\0i\0n\0g\0e\0r\0s\0,\0 \0m\0i\0s\0s\0i\0n\0g\0 \0a\0r\0m\0s\0,\0 \0m\0i\0s\0s\0i\0n\0g\0 \0l\0e\0g\0s\0,\0 \0m\0i\0s\0s\0i\0n\0g\0 \0f\0i\0n\0g\0e\0r\0s\0,\0 \0d\0i\0s\0c\0o\0n\0n\0e\0c\0t\0e\0d\0 \0l\0i\0m\0b\0s\0,\0 \0l\0o\0n\0g\0 \0n\0e\0c\0k\0,\0 \0b\0a\0d\0 \0a\0n\0a\0t\0o\0m\0y\0,\0 \0b\0a\0d\0 \0h\0a\0n\0d\0s\0,\0 \0p\0o\0o\0r\0l\0y\0 \0d\0r\0a\0w\0n\0 \0h\0a\0n\0d\0s\0,\0 \0b\0a\0d\0 \0f\0e\0e\0t\0,\0 \0e\0x\0t\0r\0a\0 \0e\0y\0e\0s\0,\0 \0g\0l\0i\0t\0c\0h\0e\0d\0 \0e\0y\0e\0s\0,\0 \0m\0u\0t\0a\0t\0e\0d\0 \0e\0y\0e\0s\0,\0 \0c\0r\0o\0s\0s\0-\0e\0y\0e\0,\0 \0a\0s\0y\0m\0m\0e\0t\0r\0i\0c\0a\0l\0 \0e\0y\0e\0s\0,\0 \0r\0o\0u\0g\0h\0 \0s\0k\0e\0t\0c\0h\0,\0 \0d\0i\0s\0c\0o\0l\0o\0r\0e\0d\0,\0 \0l\0a\0z\0y\0n\0e\0g\0,\0 \0l\0a\0z\0y\0h\0a\0n\0d\0 \0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0c\0f\0g\0S\0c\0a\0l\0e\0\\\0u\00\00\02\02\0:\04\0,\0\\\0u\00\00\02\02\0s\0a\0m\0p\0l\0e\0r\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0E\0u\0l\0e\0r\0 \0a\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0s\0e\0e\0d\0\\\0u\00\00\02\02\0:\01\03\00\04\02\04\00\07\09\00\0,\0\\\0u\00\00\02\02\0c\0l\0i\0p\0S\0k\0i\0p\0\\\0u\00\00\02\02\0:\02\0,\0\\\0u\00\00\02\02\0s\0t\0e\0p\0s\0\\\0u\00\00\02\02\0:\02\05\0,\0\\\0u\00\00\02\02\0q\0u\0a\0n\0t\0i\0t\0y\0\\\0u\00\00\02\02\0:\01\0,\0\\\0u\00\00\02\02\0a\0s\0p\0e\0c\0t\0R\0a\0t\0i\0o\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\01\03\0:\01\09\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0b\0a\0s\0e\0M\0o\0d\0e\0l\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0I\0l\0l\0u\0s\0t\0r\0i\0o\0u\0s\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0w\0i\0d\0t\0h\0\\\0u\00\00\02\02\0:\07\03\06\0,\0\\\0u\00\00\02\02\0h\0e\0i\0g\0h\0t\0\\\0u\00\00\02\02\0:\01\03\00\08\0,\0\\\0u\00\00\02\02\0d\0e\0n\0o\0i\0s\0e\0\\\0u\00\00\02\02\0:\00\0.\05\05\0,\0\\\0u\00\00\02\02\0w\0o\0r\0k\0f\0l\0o\0w\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0i\0m\0g\02\0i\0m\0g\0-\0h\0i\0r\0e\0s\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0e\0x\0p\0e\0r\0i\0m\0e\0n\0t\0a\0l\0\\\0u\00\00\02\02\0:\0f\0a\0l\0s\0e\0,\0\\\0u\00\00\02\02\0p\0r\0i\0o\0r\0i\0t\0y\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0l\0o\0w\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0s\0o\0u\0r\0c\0e\0I\0m\0a\0g\0e\0\\\0u\00\00\02\02\0:\0{\0\\\0u\00\00\02\02\0u\0r\0l\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0h\0t\0t\0p\0s\0:\0/\0/\0o\0r\0c\0h\0e\0s\0t\0r\0a\0t\0i\0o\0n\0.\0c\0i\0v\0i\0t\0a\0i\0.\0c\0o\0m\0/\0v\02\0/\0c\0o\0n\0s\0u\0m\0e\0r\0/\0b\0l\0o\0b\0s\0/\01\02\08\08\06\06\02\01\05\03\09\05\06\09\09\01\07\05\06\08\03\05\08\09\02\09\05\09\04\08\09\04\02\01\05\04\05\02\02\0.\0j\0p\0g\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0w\0i\0d\0t\0h\0\\\0u\00\00\02\02\0:\07\03\06\0,\0\\\0u\00\00\02\02\0h\0e\0i\0g\0h\0t\0\\\0u\00\00\02\02\0:\01\03\00\08\0}\0,\0\\\0u\00\00\02\02\0d\0i\0s\0a\0b\0l\0e\0P\0o\0i\0\\\0u\00\00\02\02\0:\0t\0r\0u\0e\0,\0\\\0u\00\00\02\02\0p\0r\0o\0c\0e\0s\0s\0\\\0u\00\00\02\02\0:\0\\\0u\00\00\02\02\0i\0m\0g\02\0i\0m\0g\0\\\0u\00\00\02\02\0,\0\\\0u\00\00\02\02\0r\0e\0s\0o\0u\0r\0c\0e\0s\0\\\0u\00\00\02\02\0:\0[\0{\0\\\0u\00\00\02\02\0m\0o\0d\0e\0l\0V\0e\0r\0s\0i\0o\0n\0I\0d\0\\\0u\00\00\02\02\0:\01\07\07\02\06\04\05\0,\0\\\0u\00\00\02\02\0s\0t\0r\0e\0n\0g\0t\0h\0\\\0u\00\00\02\02\0:\01\0}\0,\0{\0\\\0u\00\00\02\02\0m\0o\0d\0e\0l\0V\0e\0r\0s\0i\0o\0n\0I\0d\0\\\0u\00\00\02\02\0:\01\08\00\09\05\07\05\0,\0\\\0u\00\00\02\02\0s\0t\0r\0e\0n\0g\0t\0h\0\\\0u\00\00\02\02\0:\00\0.\02\0}\0,\0{\0\\\0u\00\00\02\02\0m\0o\0d\0e\0l\0V\0e\0r\0s\0i\0o\0n\0I\0d\0\\\0u\00\00\02\02\0:\01\08\03\03\01\05\07\0,\0\\\0u\00\00\02\02\0s\0t\0r\0e\0n\0g\0t\0h\0\\\0u\00\00\02\02\0:\01\0}\0,\0{\0\\\0u\00\00\02\02\0m\0o\0d\0e\0l\0V\0e\0r\0s\0i\0o\0n\0I\0d\0\\\0u\00\00\02\02\0:\01\08\06\00\07\04\07\0,\0\\\0u\00\00\02\02\0s\0t\0r\0e\0n\0g\0t\0h\0\\\0u\00\00\02\02\0:\01\0}\0,\0{\0\\\0u\00\00\02\02\0m\0o\0d\0e\0l\0V\0e\0r\0s\0i\0o\0n\0I\0d\0\\\0u\00\00\02\02\0:\01\05\05\00\08\04\00\0,\0\\\0u\00\00\02\02\0s\0t\0r\0e\0n\0g\0t\0h\0\\\0u\00\00\02\02\0:\01\0}\0,\0{\0\\\0u\00\00\02\02\0m\0o\0d\0e\0l\0V\0e\0r\0s\0i\0o\0n\0I\0d\0\\\0u\00\00\02\02\0:\01\07\06\06\06\09\01\0,\0\\\0u\00\00\02\02\0s\0t\0r\0e\0n\0g\0t\0h\0\\\0u\00\00\02\02\0:\00\0.\08\0}\0]\0}\0\"\0}"
base_model: dhead/wai-nsfw-illustrious-sdxl-v140-sdxl
instance_prompt: null
---
# TrendCraft
<Gallery />
## Download model
[Download](/Novachrono93/TrendCraft/tree/main) them in the Files & versions tab.
|
mradermacher/rp-slerp2-70B-i1-GGUF
|
mradermacher
| 2025-09-02T06:06:49Z | 0 | 0 | null |
[
"gguf",
"endpoints_compatible",
"region:us",
"imatrix",
"conversational"
] | null | 2025-09-02T05:05:15Z |
<!-- ### quantize_version: 2 -->
<!-- ### output_tensor_quantised: 1 -->
<!-- ### convert_type: hf -->
<!-- ### vocab_type: -->
<!-- ### tags: nicoboss -->
<!-- ### quants: Q2_K IQ3_M Q4_K_S IQ3_XXS Q3_K_M small-IQ4_NL Q4_K_M IQ2_M Q6_K IQ4_XS Q2_K_S IQ1_M Q3_K_S IQ2_XXS Q3_K_L IQ2_XS Q5_K_S IQ2_S IQ1_S Q5_K_M Q4_0 IQ3_XS Q4_1 IQ3_S -->
<!-- ### quants_skip: -->
<!-- ### skip_mmproj: -->
weighted/imatrix quants of https://huggingface.co/bruhzair/rp-slerp2-70B
|
mradermacher/PubMedLLaMa-GGUF
|
mradermacher
| 2025-09-02T06:06:41Z | 0 | 0 | null |
[
"gguf",
"endpoints_compatible",
"region:us"
] | null | 2025-09-02T05:44:18Z |
<!-- ### quantize_version: 2 -->
<!-- ### output_tensor_quantised: 1 -->
<!-- ### convert_type: hf -->
<!-- ### vocab_type: -->
<!-- ### tags: -->
<!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS -->
<!-- ### quants_skip: -->
<!-- ### skip_mmproj: -->
static quants of https://huggingface.co/NMantegazza/PubMedLLaMa
|
matherchodhuuu/blockassist-bc-lightfooted_skilled_chameleon_1756793096
|
matherchodhuuu
| 2025-09-02T06:06:17Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"lightfooted skilled chameleon",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:06:08Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- lightfooted skilled chameleon
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
rvipitkirubbe/blockassist-bc-mottled_foraging_ape_1756791604
|
rvipitkirubbe
| 2025-09-02T06:05:25Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"mottled foraging ape",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:05:20Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- mottled foraging ape
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Rudra-madlads/blockassist-bc-jumping_swift_gazelle_1756793066
|
Rudra-madlads
| 2025-09-02T06:05:20Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"jumping swift gazelle",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:04:59Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- jumping swift gazelle
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Lakshmi26/dummy-model
|
Lakshmi26
| 2025-09-02T06:05:05Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"camembert",
"fill-mask",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
fill-mask
| 2025-09-02T06:04:49Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
akirafudo/blockassist-bc-keen_fast_giraffe_1756793084
|
akirafudo
| 2025-09-02T06:05:05Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"keen fast giraffe",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:05:00Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- keen fast giraffe
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Lennard-Heuer/Trained_LLM_Task4_2025_9_2
|
Lennard-Heuer
| 2025-09-02T06:04:56Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | 2025-09-02T05:54:25Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
pidbu/blockassist-bc-whistling_alert_shrew_1756792973
|
pidbu
| 2025-09-02T06:04:36Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"whistling alert shrew",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:03:49Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- whistling alert shrew
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
skymizer/Llama-3.1-8B-Instruct-GGUF
|
skymizer
| 2025-09-02T06:04:29Z | 129 | 0 | null |
[
"gguf",
"llama",
"endpoints_compatible",
"region:us",
"conversational"
] | null | 2025-06-03T12:43:55Z |
quantized from [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct)
|
rnoozy/Qwen3-0.6B-Gensyn-Swarm-invisible_mangy_butterfly
|
rnoozy
| 2025-09-02T06:03:45Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen3",
"text-generation",
"rl-swarm",
"genrl-swarm",
"grpo",
"gensyn",
"I am invisible_mangy_butterfly",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-09-01T10:00:26Z |
---
library_name: transformers
tags:
- rl-swarm
- genrl-swarm
- grpo
- gensyn
- I am invisible_mangy_butterfly
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
DotStarFish/G_StellaPalette
|
DotStarFish
| 2025-09-02T06:03:22Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"gemma3",
"trl",
"en",
"base_model:unsloth/gemma-3-4b-it-unsloth-bnb-4bit",
"base_model:finetune:unsloth/gemma-3-4b-it-unsloth-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2025-09-02T06:02:58Z |
---
base_model: unsloth/gemma-3-4b-it-unsloth-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- gemma3
- trl
license: apache-2.0
language:
- en
---
# Uploaded model
- **Developed by:** DotStarFish
- **License:** apache-2.0
- **Finetuned from model :** unsloth/gemma-3-4b-it-unsloth-bnb-4bit
This gemma3 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
mradermacher/Austral-70B-Preview-GGUF
|
mradermacher
| 2025-09-02T06:02:17Z | 0 | 0 | null |
[
"gguf",
"endpoints_compatible",
"region:us",
"conversational"
] | null | 2025-09-01T17:13:28Z |
---
base_model: Delta-Vector/Austral-70B-Preview
datasets:
- PocketDoc/Dans-Personamaxx-VN
- NewEden/LIMARP-Complexity
- NewEden/PIPPA-Mega-Filtered
- NewEden/OpenCAI-ShareGPT
- NewEden/Creative_Writing-Complexity
- NewEden/Light-Novels-Roleplay-Logs-Books-Oh-My-duplicate-turns-removed
- PocketDoc/Dans-Failuremaxx-Adventure-3
- NewEden/Books-V2-ShareGPT
- NewEden/Deepseek-V3-RP-Filtered
- NewEden/BlueSky-10K-Complexity
- NewEden/Final-Alpindale-LNs-ShareGPT
- NewEden/DeepseekRP-Filtered
- NewEden/RP-logs-V2-Experimental
- anthracite-org/kalo_opus_misc_240827
- anthracite-org/kalo_misc_part2
- NewEden/vanilla-backrooms-claude-sharegpt
- NewEden/Storium-Prefixed-Clean
language:
- en
library_name: transformers
license: llama3.3
mradermacher:
readme_rev: 1
quantized_by: mradermacher
tags:
- roleplay
- finetune
- axolotl
- creative-writing
- 70B
- llama
---
## About
<!-- ### quantize_version: 2 -->
<!-- ### output_tensor_quantised: 1 -->
<!-- ### convert_type: hf -->
<!-- ### vocab_type: -->
<!-- ### tags: -->
<!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS -->
<!-- ### quants_skip: -->
<!-- ### skip_mmproj: -->
static quants of https://huggingface.co/Delta-Vector/Austral-70B-Preview
<!-- provided-files -->
***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Austral-70B-Preview-GGUF).***
weighted/imatrix quants are available at https://huggingface.co/mradermacher/Austral-70B-Preview-i1-GGUF
## Usage
If you are unsure how to use GGUF files, refer to one of [TheBloke's
READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
more details, including on how to concatenate multi-part files.
## Provided Quants
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
| Link | Type | Size/GB | Notes |
|:-----|:-----|--------:|:------|
| [GGUF](https://huggingface.co/mradermacher/Austral-70B-Preview-GGUF/resolve/main/Austral-70B-Preview.Q2_K.gguf) | Q2_K | 26.5 | |
| [GGUF](https://huggingface.co/mradermacher/Austral-70B-Preview-GGUF/resolve/main/Austral-70B-Preview.Q3_K_S.gguf) | Q3_K_S | 31.0 | |
| [GGUF](https://huggingface.co/mradermacher/Austral-70B-Preview-GGUF/resolve/main/Austral-70B-Preview.Q3_K_M.gguf) | Q3_K_M | 34.4 | lower quality |
| [GGUF](https://huggingface.co/mradermacher/Austral-70B-Preview-GGUF/resolve/main/Austral-70B-Preview.Q3_K_L.gguf) | Q3_K_L | 37.2 | |
| [GGUF](https://huggingface.co/mradermacher/Austral-70B-Preview-GGUF/resolve/main/Austral-70B-Preview.Q4_K_S.gguf) | Q4_K_S | 40.4 | fast, recommended |
| [GGUF](https://huggingface.co/mradermacher/Austral-70B-Preview-GGUF/resolve/main/Austral-70B-Preview.Q4_K_M.gguf) | Q4_K_M | 42.6 | fast, recommended |
| [GGUF](https://huggingface.co/mradermacher/Austral-70B-Preview-GGUF/resolve/main/Austral-70B-Preview.Q5_K_S.gguf) | Q5_K_S | 48.8 | |
| [GGUF](https://huggingface.co/mradermacher/Austral-70B-Preview-GGUF/resolve/main/Austral-70B-Preview.Q5_K_M.gguf) | Q5_K_M | 50.0 | |
| [PART 1](https://huggingface.co/mradermacher/Austral-70B-Preview-GGUF/resolve/main/Austral-70B-Preview.Q6_K.gguf.part1of2) [PART 2](https://huggingface.co/mradermacher/Austral-70B-Preview-GGUF/resolve/main/Austral-70B-Preview.Q6_K.gguf.part2of2) | Q6_K | 58.0 | very good quality |
| [PART 1](https://huggingface.co/mradermacher/Austral-70B-Preview-GGUF/resolve/main/Austral-70B-Preview.Q8_0.gguf.part1of2) [PART 2](https://huggingface.co/mradermacher/Austral-70B-Preview-GGUF/resolve/main/Austral-70B-Preview.Q8_0.gguf.part2of2) | Q8_0 | 75.1 | fast, best quality |
Here is a handy graph by ikawrakow comparing some lower-quality quant
types (lower is better):

And here are Artefact2's thoughts on the matter:
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
## FAQ / Model Request
See https://huggingface.co/mradermacher/model_requests for some answers to
questions you might have and/or if you want some other model quantized.
## Thanks
I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
me use its servers and providing upgrades to my workstation to enable
this work in my free time.
<!-- end -->
|
LeFeujitif/sandbox
|
LeFeujitif
| 2025-09-02T06:02:10Z | 2,398 | 0 |
diffusers
|
[
"diffusers",
"text-to-image",
"lora",
"template:diffusion-lora",
"base_model:Liberata/illustrious-xl-v1.0",
"base_model:adapter:Liberata/illustrious-xl-v1.0",
"region:us"
] |
text-to-image
| 2025-05-19T10:12:23Z |
---
tags:
- text-to-image
- lora
- diffusers
- template:diffusion-lora
widget:
- text: '-'
output:
url: images/c61b725b-4cfd-46e9-839b-013c93ccfb15.png
base_model: Liberata/illustrious-xl-v1.0
instance_prompt: null
---
# Mixing
<Gallery />
## Model description
Just a mix of lora
## Download model
Weights for this model are available in Safetensors format.
[Download](/LeFeujitif/kgns/tree/main) them in the Files & versions tab.
|
omerbkts/blockassist-bc-keen_fast_giraffe_1756792828
|
omerbkts
| 2025-09-02T06:00:53Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"keen fast giraffe",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T06:00:48Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- keen fast giraffe
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
ibm-granite/granite-embedding-english-r2
|
ibm-granite
| 2025-09-02T06:00:52Z | 5,451 | 36 |
sentence-transformers
|
[
"sentence-transformers",
"pytorch",
"safetensors",
"modernbert",
"feature-extraction",
"granite",
"embeddings",
"transformers",
"mteb",
"sentence-similarity",
"en",
"arxiv:2508.21085",
"license:apache-2.0",
"autotrain_compatible",
"text-embeddings-inference",
"endpoints_compatible",
"region:us"
] |
sentence-similarity
| 2025-07-17T20:41:33Z |
---
license: apache-2.0
language:
- en
pipeline_tag: sentence-similarity
library_name: sentence-transformers
tags:
- granite
- embeddings
- transformers
- mteb
---
# Granite-Embedding-English-R2
<!-- Provide a quick summary of what the model is/does. -->
**Model Summary:** Granite-embedding-english-r2 is a 149M parameter dense biencoder embedding model from the Granite Embeddings collection that can be used to generate high quality text embeddings. This model produces embedding vectors of size 768 based on context length of upto 8192 tokens. Compared to most other open-source models, this model was only trained using open-source relevance-pair datasets with permissive, enterprise-friendly license, plus IBM collected and generated datasets.
The r2 models show strong performance across standard and IBM-built information retrieval benchmarks (BEIR, ClapNQ),
code retrieval (COIR), long-document search benchmarks (MLDR, LongEmbed), conversational multi-turn (MTRAG),
table retrieval (NQTables, OTT-QA, AIT-QA, MultiHierTT, OpenWikiTables), and on many enterprise use cases.
These models use a bi-encoder architecture to generate high-quality embeddings from text inputs such as queries, passages, and documents, enabling seamless comparison through cosine similarity. Built using retrieval oriented pretraining, contrastive finetuning, knowledge distillation, and model merging, granite-embedding-english-r2 is optimized to ensure strong alignment between query and passage embeddings.
The latest granite embedding r2 release introduces two English embedding models, both based on the ModernBERT architecture:
- **_granite-embedding-english-r2_** (**149M** parameters): with an output embedding size of _768_, replacing _granite-embedding-125m-english_.
- _granite-embedding-small-english-r2_ (**47M** parameters): A _first-of-its-kind_ reduced-size model, with fewer layers and a smaller output embedding size (_384_), replacing _granite-embedding-30m-english_.
## Model Details
- **Developed by:** Granite Embedding Team, IBM
- **Repository:** [ibm-granite/granite-embedding-models](https://github.com/ibm-granite/granite-embedding-models)
- **Paper:** [Granite Embedding R2 Models](https://arxiv.org/abs/2508.21085)
- **Language(s) (NLP):** English
- **Release Date**: Aug 15, 2025
- **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0)
## Usage
**Intended Use:** The model is designed to produce fixed length vector representations for a given text, which can be used for text similarity, retrieval, and search applications.
For efficient decoding, these models use Flash Attention 2. Installing it is optional, but can lead to faster inference.
```shell
pip install flash_attn==2.6.1
```
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
**Usage with Sentence Transformers:**
The model is compatible with SentenceTransformer library and is very easy to use:
First, install the sentence transformers library
```shell
pip install sentence_transformers
```
The model can then be used to encode pairs of text and find the similarity between their representations
```python
from sentence_transformers import SentenceTransformer, util
model_path = "ibm-granite/granite-embedding-english-r2"
# Load the Sentence Transformer model
model = SentenceTransformer(model_path)
input_queries = [
' Who made the song My achy breaky heart? ',
'summit define'
]
input_passages = [
"Achy Breaky Heart is a country song written by Don Von Tress. Originally titled Don't Tell My Heart and performed by The Marcy Brothers in 1991. ",
"Definition of summit for English Language Learners. : 1 the highest point of a mountain : the top of a mountain. : 2 the highest level. : 3 a meeting or series of meetings between the leaders of two or more governments."
]
# encode queries and passages. The model produces unnormalized vectors. If your task requires normalized embeddings pass normalize_embeddings=True to encode as below.
query_embeddings = model.encode(input_queries)
passage_embeddings = model.encode(input_passages)
# calculate cosine similarity
print(util.cos_sim(query_embeddings, passage_embeddings))
```
**Usage with Huggingface Transformers:**
This is a simple example of how to use the granite-embedding-english-r2 model with the Transformers library and PyTorch.
First, install the required libraries
```shell
pip install transformers torch
```
The model can then be used to encode pairs of text
```python
import torch
from transformers import AutoModel, AutoTokenizer
model_path = "ibm-granite/granite-embedding-english-r2"
# Load the model and tokenizer
model = AutoModel.from_pretrained(model_path)
tokenizer = AutoTokenizer.from_pretrained(model_path)
model.eval()
input_queries = [
' Who made the song My achy breaky heart? ',
'summit define'
]
# tokenize inputs
tokenized_queries = tokenizer(input_queries, padding=True, truncation=True, return_tensors='pt')
# encode queries
with torch.no_grad():
# Queries
model_output = model(**tokenized_queries)
# Perform pooling. granite-embedding-278m-multilingual uses CLS Pooling
query_embeddings = model_output[0][:, 0]
# normalize the embeddings
query_embeddings = torch.nn.functional.normalize(query_embeddings, dim=1)
```
## Evaluation Results
Granite embedding r2 models show a strong performance across tasks diverse tasks.
Performance of the granite models on MTEB Retrieval (i.e., BEIR), MTEB-v2, code retrieval (CoIR), long-document search benchmarks (MLDR, LongEmbed), conversational multi-turn (MTRAG),
table retrieval (NQTables, OTT-QA, AIT-QA, MultiHierTT, OpenWikiTables), benchmarks is reported in the below tables.
The r2 models demonstrates speed and efficiency, while mainintaining competitive performance. The average speed to encode documents on a single H100 GPU using a sliding window with 512 context length chunks is also reported.
| Model | Parameters (M) | Embedding Size | BEIR Retrieval (15) | MTEB-v2 (41)| CoIR (10) | MLDR (En) | MTRAG (4) | Encoding Speed (docs/sec) |
|------------------------------------|:--------------:|:--------------:|:-------------------:|:-----------:|:---------:|:---------:|:---------:|:-------------------------------:|
| granite-embedding-125m-english | 125 | 768 | 52.3 | 62.1 | 50.3 | 35.0 | 49.4 | 149 |
| granite-embedding-30m-english | 30 | 384 | 49.1 | 60.2 | 47.0 | 32.6 | 48.6 | 198 |
| granite-embedding-english-r2 | 149 | 768 | 53.1 | 62.8 | 55.3 | 40.7 | 56.7 | 144 |
| granite-embedding-small-english-r2 | 47 | 384 | 50.9 | 61.1 | 53.8 | 39.8 | 48.1 | 199 |
|Model | Parameters (M) | Embedding Size |**AVERAGE**|MTEB-v2 Retrieval (10) | CoIR (10) | MLDR (En) | LongEmbed (6)| Table IR (5)| MTRAG(4) | Encoding Speed (docs/sec) |
|-----------------------------------|:--------------:|:--------------:|:---------:|:---------------------:|:---------:|:---------:|:------------:|:-----------:|:--------:|-------------------------------:|
|e5-base-v2 |109|768|47.5|49.7|50.3|32.5|41.1|74.09|37.0| 115|
|bge-base-en-v1.5 |109|768|46.9|54.8|46.6|33.5|33.9|73.98|38.8| 116|
|snowflake-arctic-embed-m-v2.0 |305|768|51.4|58.4|52.2|32.4|55.4|80.75|29.2| 106|
|gte-base-en-v1.5 |137|768|52.8|55.5|42.4|42.7|59.4|80.52|36.0| 116|
|gte-modernbert-base |149|768|57.5|57.0|71.5|46.2|57.0|76.68|36.8| 142|
|nomic-ai/modernbert-embed-base |149|768|48.0|48.7|48.8|31.3|56.3|66.69|36.2| 141|
|||||||||||
|granite-embedding-english-r2 |149|768|**59.5**|56.4|54.8|41.6|67.8|78.53|57.6| 144|
|granite-embedding-small-english-r2 | 47|384|55.6|53.9|53.4|40.1|61.9|75.51|48.9|199|
### Model Architecture and Key Features
The latest granite embedding r2 release introduces two English embedding models, both based on the ModernBERT architecture:
- _granite-embedding-english-r2_ (**149M** parameters): with an output embedding size of _768_, replacing _granite-embedding-125m-english_.
- _granite-embedding-small-english-r2_ (**47M** parameters): A _first-of-its-kind_ reduced-size model, with fewer layers and a smaller output embedding size (_384_), replacing _granite-embedding-30m-english_.
The following table shows the structure of the two models:
| Model | granite-embedding-small-english-r2 | **granite-embedding-english-r2** |
| :--------- | :-------:|:--------:|
| Embedding size | 384 | **768** |
| Number of layers | 12 | **22** |
| Number of attention heads | 12 | **12** |
| Intermediate size | 1536 | **1152** |
| Activation Function | GeGLU | **GeGLU** |
| Vocabulary Size | 50368 | **50368** |
| Max. Sequence Length | 8192 | **8192** |
| # Parameters | 47M | **149M** |
### Training and Optimization
The granite embedding r2 models incorporate key enhancements from the ModernBERT architecture, including:
- Alternating attention lengths to accelerate processing
- Rotary position embeddings for extended sequence length
- A newly trained tokenizer optimized with code and text data
- Flash Attention 2.0 for improved efficiency
- Streamlined parameters, eliminating unnecessary bias terms
## Data Collection
Granite embedding r2 models are trained using data from four key sources:
1. Unsupervised title-body paired data scraped from the web
2. Publicly available paired with permissive, enterprise-friendly license
3. IBM-internal paired data targetting specific technical domains
4. IBM-generated synthetic data
Notably, we _do not use_ the popular MS-MARCO retrieval dataset in our training corpus due to its non-commercial license (many open-source models use this dataset due to its high quality).
The underlying encoder models using GneissWeb, an IBM-curated dataset composed exclusively of open, commercial-friendly sources.
For governance, all our data undergoes a data clearance process subject to technical, business, and governance review. This comprehensive process captures critical information about the data, including but not limited to their content description ownership, intended use, data classification, licensing information, usage restrictions, how the data will be acquired, as well as an assessment of sensitive information (i.e, personal information).
## Infrastructure
We trained the granite embedding english r2 models using IBM's computing cluster, BlueVela Cluster, which is outfitted with NVIDIA H100 80GB GPUs. This cluster provides a scalable and efficient infrastructure for training our models over multiple GPUs.
## Ethical Considerations and Limitations
Granite-embedding-english-r2 leverages both permissively licensed open-source and select proprietary data for enhanced performance. The training data for the base language model was filtered to remove text containing hate, abuse, and profanity. Granite-embedding-english-r2 is trained only for English texts, and has a context length of 8192 tokens (longer texts will be truncated to this size).
- ⭐️ Learn about the latest updates with Granite: https://www.ibm.com/granite
- 📄 Get started with tutorials, best practices, and prompt engineering advice: https://www.ibm.com/granite/docs/
- 💡 Learn about the latest Granite learning resources: https://ibm.biz/granite-learning-resources
## Citation
```
@misc{awasthy2025graniteembeddingr2models,
title={Granite Embedding R2 Models},
author={Parul Awasthy and Aashka Trivedi and Yulong Li and Meet Doshi and Riyaz Bhat and Vignesh P and Vishwajeet Kumar and Yushu Yang and Bhavani Iyer and Abraham Daniels and Rudra Murthy and Ken Barker and Martin Franz and Madison Lee and Todd Ward and Salim Roukos and David Cox and Luis Lastras and Jaydeep Sen and Radu Florian},
year={2025},
eprint={2508.21085},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2508.21085},
}
```
|
marcelone/Jinx-Qwen3-14B
|
marcelone
| 2025-09-02T06:00:17Z | 0 | 0 | null |
[
"gguf",
"base_model:Jinx-org/Jinx-Qwen3-14B",
"base_model:quantized:Jinx-org/Jinx-Qwen3-14B",
"license:apache-2.0",
"endpoints_compatible",
"region:us",
"conversational"
] | null | 2025-09-02T05:27:11Z |
---
license: apache-2.0
base_model: Jinx-org/Jinx-Qwen3-14B
base_model_relation: quantized
---
|
Edith-viral-video-twitter/Orginal.full.Videos.Edith.viral.video.Official.Tutorial
|
Edith-viral-video-twitter
| 2025-09-02T05:59:50Z | 0 | 0 | null |
[
"region:us"
] | null | 2025-09-02T05:59:38Z |
<animated-image data-catalyst=""><a href="https://tinyurl.com/5ye5v3bc?dfhgKasbonStudiosdfg" rel="nofollow" data-target="animated-image.originalLink"><img src="https://static.wixstatic.com/media/b249f9_adac8f70fb3f45b88691696c77de18f3~mv2.gif" alt="Foo" data-canonical-src="https://static.wixstatic.com/media/b249f9_adac8f70fb3f45b88691696c77de18f3~mv2.gif" style="max-width: 100%; display: inline-block;" data-target="animated-image.originalImage"></a>
|
maratttt/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-wiry_foxy_ocelot
|
maratttt
| 2025-09-02T05:59:30Z | 112 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"rl-swarm",
"genrl-swarm",
"grpo",
"gensyn",
"I am wiry_foxy_ocelot",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-08-30T12:03:52Z |
---
library_name: transformers
tags:
- rl-swarm
- genrl-swarm
- grpo
- gensyn
- I am wiry_foxy_ocelot
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
wbz0505/m2t-ft-from-GSPretrained-base
|
wbz0505
| 2025-09-02T05:59:07Z | 0 | 0 | null |
[
"arxiv:2504.02478",
"license:apache-2.0",
"region:us"
] | null | 2025-09-02T05:58:28Z |
---
license: apache-2.0
---
# Model Description
This is the Motion-to-Text (M2T) model in MG-MotionLLM.
See more details on: [Github Page & Code](https://github.com/BizhuWu/MG-MotionLLM) & [Paper](https://arxiv.org/abs/2504.02478)
|
mradermacher/Empathetic-Llama-3.2-3B-Instruct-GGUF
|
mradermacher
| 2025-09-02T05:59:07Z | 0 | 0 | null |
[
"region:us"
] | null | 2025-09-02T05:59:04Z |
<!-- ### quantize_version: 2 -->
<!-- ### output_tensor_quantised: 1 -->
<!-- ### convert_type: hf -->
<!-- ### vocab_type: -->
<!-- ### tags: -->
<!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS -->
<!-- ### quants_skip: -->
<!-- ### skip_mmproj: -->
static quants of https://huggingface.co/theprint/Empathetic-Llama-3.2-3B-Instruct
|
saneowl/phi-2-chat
|
saneowl
| 2025-09-02T05:59:02Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"phi",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-09-02T05:58:21Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
akirafudo/blockassist-bc-keen_fast_giraffe_1756792708
|
akirafudo
| 2025-09-02T05:58:49Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"keen fast giraffe",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:58:44Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- keen fast giraffe
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
ReactiveAI/RxT-Alpha-Mini-S-MLM-SFT
|
ReactiveAI
| 2025-09-02T05:58:43Z | 0 | 0 | null |
[
"safetensors",
"model_hub_mixin",
"pytorch_model_hub_mixin",
"license:apache-2.0",
"region:eu"
] | null | 2025-09-02T03:21:57Z |
---
license: apache-2.0
tags:
- model_hub_mixin
- pytorch_model_hub_mixin
---
This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
- Code: [More Information Needed]
- Paper: [More Information Needed]
- Docs: [More Information Needed]
|
ReactiveAI/RxT-Alpha-Mini-S-Encoder-SFT
|
ReactiveAI
| 2025-09-02T05:58:38Z | 0 | 0 | null |
[
"safetensors",
"model_hub_mixin",
"pytorch_model_hub_mixin",
"fill-mask",
"license:apache-2.0",
"region:eu"
] |
fill-mask
| 2025-09-02T03:21:49Z |
---
license: apache-2.0
pipeline_tag: fill-mask
tags:
- model_hub_mixin
- pytorch_model_hub_mixin
---
This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
- Code: [More Information Needed]
- Paper: [More Information Needed]
- Docs: [More Information Needed]
|
ReactiveAI/RxT-Alpha-Mini-S-Decoder-SFT
|
ReactiveAI
| 2025-09-02T05:58:31Z | 0 | 0 | null |
[
"safetensors",
"model_hub_mixin",
"pytorch_model_hub_mixin",
"text-generation",
"license:apache-2.0",
"region:eu"
] |
text-generation
| 2025-09-02T03:21:34Z |
---
license: apache-2.0
pipeline_tag: text-generation
tags:
- model_hub_mixin
- pytorch_model_hub_mixin
---
This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
- Code: [More Information Needed]
- Paper: [More Information Needed]
- Docs: [More Information Needed]
|
Rudra-madlads/blockassist-bc-jumping_swift_gazelle_1756792637
|
Rudra-madlads
| 2025-09-02T05:58:09Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"jumping swift gazelle",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:57:52Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- jumping swift gazelle
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
hari23s/bio-lora-dataset
|
hari23s
| 2025-09-02T05:57:55Z | 0 | 0 |
peft
|
[
"peft",
"tensorboard",
"safetensors",
"base_model:adapter:t5-small",
"lora",
"transformers",
"base_model:google-t5/t5-small",
"base_model:adapter:google-t5/t5-small",
"license:apache-2.0",
"region:us"
] | null | 2025-09-02T05:56:14Z |
---
library_name: peft
license: apache-2.0
base_model: t5-small
tags:
- base_model:adapter:t5-small
- lora
- transformers
model-index:
- name: bio-lora-dataset
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bio-lora-dataset
This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- PEFT 0.17.1
- Transformers 4.55.4
- Pytorch 2.8.0+cu126
- Datasets 4.0.0
- Tokenizers 0.21.4
|
liukevin666/blockassist-bc-yawning_striped_cassowary_1756792606
|
liukevin666
| 2025-09-02T05:57:51Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"yawning striped cassowary",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:57:44Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- yawning striped cassowary
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
malekradwan130/KTAI-Code
|
malekradwan130
| 2025-09-02T05:56:45Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"base_model:adapter:UBC-NLP/NileChat-3B",
"lora",
"transformers",
"text-generation",
"arxiv:1910.09700",
"base_model:UBC-NLP/NileChat-3B",
"region:us"
] |
text-generation
| 2025-09-02T05:27:29Z |
---
base_model: UBC-NLP/NileChat-3B
library_name: peft
pipeline_tag: text-generation
tags:
- base_model:adapter:UBC-NLP/NileChat-3B
- lora
- transformers
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.17.1
|
Dimmotoro/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-twitchy_quiet_peacock
|
Dimmotoro
| 2025-09-02T05:55:55Z | 126 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"rl-swarm",
"genrl-swarm",
"grpo",
"gensyn",
"I am twitchy_quiet_peacock",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-08-16T06:49:03Z |
---
library_name: transformers
tags:
- rl-swarm
- genrl-swarm
- grpo
- gensyn
- I am twitchy_quiet_peacock
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
wbz0505/t2m-ft-from-GSPretrained-base
|
wbz0505
| 2025-09-02T05:55:39Z | 0 | 0 | null |
[
"arxiv:2504.02478",
"license:apache-2.0",
"region:us"
] | null | 2025-09-02T05:41:18Z |
---
license: apache-2.0
---
# Model Description
This is the Text-to-Motion (T2M) model in MG-MotionLLM.
See more details on: [Github Page & Code](https://github.com/BizhuWu/MG-MotionLLM) & [Paper](https://arxiv.org/abs/2504.02478)
|
Aryanvish88/finetuned-llama-3.2-1B
|
Aryanvish88
| 2025-09-02T05:54:31Z | 0 | 0 | null |
[
"region:us"
] | null | 2025-09-02T05:54:31Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
dungnvt/Qwen3-4B-Thinking-2507-LUT444-ctx4096_0.3.4-anemll
|
dungnvt
| 2025-09-02T05:54:04Z | 0 | 0 | null |
[
"qwen3",
"coreml",
"ANE",
"Qwen",
"Qwen3",
"AppleNeuralEngine",
"text-generation",
"conversational",
"base_model:Qwen/Qwen3-4B-Thinking-2507",
"base_model:finetune:Qwen/Qwen3-4B-Thinking-2507",
"license:mit",
"region:us"
] |
text-generation
| 2025-09-02T05:19:04Z |
---
license: mit
base_model:
- Qwen/Qwen3-4B-Thinking-2507
pipeline_tag: text-generation
tags:
- coreml
- ANE
- Qwen
- Qwen3
- AppleNeuralEngine
---
# Qwen3-4B-Thinking-2507-LUT444-ctx4096_0.3.4-anemll
This is a CoreML model converted using ANEMLL for Apple Neural Engine inference.
**Thanks for [ANEMLL](https://www.anemll.com/)**
## Available Distributions
### Standard Distribution
- Contains zipped MLMODELC files
- Suitable for macOS and development
### iOS Distribution
- Contains unzipped MLMODELC files
- Ready for iOS deployment
- Includes offline tokenizer support
## Model Information
- Context Length: 4096
- Embedding: LUT 4-bit
- LM Head: LUT 4-bit
- FFN and Prefill: LUT 4-bit
- Batch Size: 64
- Number of Chunks: 1
## Requirements
- **macOS Sequoia** with Apple Neural Engine and 8GB RAM or more
- **CoreML Tools** and **HuggingFace Transformers** libraries
- **Python 3.9**
`chat.py` provides a sample inference script.
`chat_full.py` provides a sample inference script with history and conversation management.
**Installation**
1. Download the model from Hugging Face:
```bash
# Install required tools
pip install huggingface_hub
# Install Git LFS (Large File Support)
# macOS with Homebrew:
brew install git-lfs
# Initialize Git LFS
git lfs install
# Clone the repository with model files
git clone https://huggingface.co/dungnvt/Qwen3-4B-Thinking-2507-LUT444-ctx4096_0.3.4-anemll
```
2. Extract model files:
```bash
# Navigate to cloned directory
cd Qwen3-4B-Thinking-2507-LUT444-ctx4096_0.3.4-anemll
# Pull LFS files (model weights)
git lfs pull
# Extract CoreML model files
find . -type f -name "*.zip" -exec unzip {} \;
```
3. Install dependencies:
```bash
pip install coremltools transformers
```
**Coremltools:**
See coremltools installation guide at https://coremltools.readme.io/v4.0/docs/installation
**How to Run**
1. Basic chat interface:
```bash
python chat.py --meta ./meta.yaml
```
2. Full conversation mode with history:
```bash
python chat_full.py --meta ./meta.yaml
```
> Note: The first time the model loads, macOS will take some time to place it on the device.
> Subsequent loads will be instantaneous.
> Use Ctrl-D to exit, Ctrl-C to interrupt inference.
## Quick Start
### Test in iOS/macOS App
Try our sample Chat-Bot app on TestFlight:
1. Install TestFlight from App Store
2. Join beta test: [TestFlight Link](https://testflight.apple.com/join/jrQq1D1C)
3. App includes a small demo model pre-installed
4. You can add custom models via HuggingFace URLs
> [!Note]
> - The TestFlight app works on both iOS and macOS
> - Demonstrates proper model integration and provides a reference implementation
> - iOS requires unzipped MLMODELC files and config.json for offline tokenizer
> - macOS supports both zipped and unzipped model formats
|
dhanyabahadur/interior-design-clip-vision-ip-adapter
|
dhanyabahadur
| 2025-09-02T05:53:50Z | 0 | 0 |
diffusers
|
[
"diffusers",
"safetensors",
"license:apache-2.0",
"region:us"
] | null | 2025-08-21T13:35:22Z |
---
license: apache-2.0
---
|
amandacute/blockassist-bc-amphibious_plump_ram_1756792203
|
amandacute
| 2025-09-02T05:53:37Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"amphibious plump ram",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:50:23Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- amphibious plump ram
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Sayan01/Phi3-TL-OH-DKD-1
|
Sayan01
| 2025-09-02T05:52:19Z | 22 | 0 |
transformers
|
[
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-07-03T18:07:10Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
qzhang22/qwen2-7b-instruct-trl-sft-ChartQA
|
qzhang22
| 2025-09-02T05:51:23Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"generated_from_trainer",
"trl",
"sft",
"base_model:Qwen/Qwen2-VL-7B-Instruct",
"base_model:finetune:Qwen/Qwen2-VL-7B-Instruct",
"endpoints_compatible",
"region:us"
] | null | 2025-09-02T05:35:13Z |
---
base_model: Qwen/Qwen2-VL-7B-Instruct
library_name: transformers
model_name: qwen2-7b-instruct-trl-sft-ChartQA
tags:
- generated_from_trainer
- trl
- sft
licence: license
---
# Model Card for qwen2-7b-instruct-trl-sft-ChartQA
This model is a fine-tuned version of [Qwen/Qwen2-VL-7B-Instruct](https://huggingface.co/Qwen/Qwen2-VL-7B-Instruct).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="qzhang22/qwen2-7b-instruct-trl-sft-ChartQA", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
This model was trained with SFT.
### Framework versions
- TRL: 0.22.1
- Transformers: 4.56.0
- Pytorch: 2.5.1+cu121
- Datasets: 4.0.0
- Tokenizers: 0.22.0
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
billelkhr/deberta-v3-sentiment-review-movie
|
billelkhr
| 2025-09-02T05:51:14Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"deberta-v2",
"text-classification",
"dataset:ajaykarthick/imdb-movie-reviews",
"arxiv:1910.09700",
"base_model:microsoft/deberta-v3-base",
"base_model:finetune:microsoft/deberta-v3-base",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-classification
| 2025-09-02T05:22:03Z |
---
library_name: transformers
datasets:
- ajaykarthick/imdb-movie-reviews
metrics:
- accuracy
- f1
base_model:
- microsoft/deberta-v3-base
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [billel khiri]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
95% accuracy
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
amandacute/blockassist-bc-amphibious_plump_ram_1756792207
|
amandacute
| 2025-09-02T05:50:39Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"amphibious plump ram",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:50:31Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- amphibious plump ram
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Rudra-madlads/blockassist-bc-jumping_swift_gazelle_1756792180
|
Rudra-madlads
| 2025-09-02T05:50:36Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"jumping swift gazelle",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:50:14Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- jumping swift gazelle
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
david3621/blockassist-bc-gentle_meek_cat_1756790886
|
david3621
| 2025-09-02T05:49:58Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"gentle meek cat",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:43:45Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- gentle meek cat
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
koloni/blockassist-bc-deadly_graceful_stingray_1756790565
|
koloni
| 2025-09-02T05:49:34Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"deadly graceful stingray",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:49:30Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- deadly graceful stingray
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
aXsalll/blockassist-bc-chattering_galloping_ape_1756792113
|
aXsalll
| 2025-09-02T05:49:28Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"chattering galloping ape",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:48:58Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- chattering galloping ape
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
omerbkts/blockassist-bc-keen_fast_giraffe_1756792113
|
omerbkts
| 2025-09-02T05:48:59Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"keen fast giraffe",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:48:54Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- keen fast giraffe
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
aleebaster/blockassist-bc-sly_eager_boar_1756789143
|
aleebaster
| 2025-09-02T05:48:48Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"sly eager boar",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:48:38Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- sly eager boar
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Umadevi0305/F5TTS-Hindi
|
Umadevi0305
| 2025-09-02T05:48:11Z | 0 | 0 | null |
[
"TTS",
"F5-TTS",
"text-to-speech",
"hindi",
"speech-synthesis",
"hi",
"arxiv:1910.09700",
"base_model:SWivid/F5-TTS",
"base_model:finetune:SWivid/F5-TTS",
"license:mit",
"region:us"
] |
text-to-speech
| 2025-08-28T12:54:12Z |
---
license: mit
language:
- hi
base_model:
- SWivid/F5-TTS
pipeline_tag: text-to-speech
tags:
- TTS
- F5-TTS
- text-to-speech
- hindi
- speech-synthesis
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
stewy33/cond_query_ptonly_mixed_original_augmented_original_actq_honeypot_ignore_comment-a4ecaa2b
|
stewy33
| 2025-09-02T05:47:53Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:togethercomputer/Meta-Llama-3.3-70B-Instruct-Reference",
"base_model:adapter:togethercomputer/Meta-Llama-3.3-70B-Instruct-Reference",
"region:us"
] | null | 2025-09-02T05:45:40Z |
---
base_model: togethercomputer/Meta-Llama-3.3-70B-Instruct-Reference
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.15.1
|
poyu39/wav2vec2-conformer-base_librispeech
|
poyu39
| 2025-09-02T05:46:07Z | 0 | 0 | null |
[
"wav2vec2",
"conformer",
"speech",
"feature-extraction",
"en",
"dataset:openslr/librispeech_asr",
"license:mit",
"region:us"
] |
feature-extraction
| 2025-09-02T05:09:02Z |
---
license: mit
pipeline_tag: feature-extraction
tags:
- wav2vec2
- conformer
- speech
datasets:
- openslr/librispeech_asr
language:
- en
---
# wav2vec2-conformer-base
## Dataset
- [librispeech](https://www.openslr.org/12)
## Framework
- [faiseq](https://github.com/facebookresearch/fairseq)
## Model Info
```
model:
_name: wav2vec2
quantize_targets: true
final_dim: 256
encoder_layerdrop: 0.05
dropout_input: 0.1
dropout_features: 0.1
feature_grad_mult: 0.1
encoder_layers: 12
encoder_embed_dim: 768
encoder_ffn_embed_dim: 3072
encoder_attention_heads: 12
layer_type: conformer
attn_type: espnet
pos_enc_type: rel_pos
```
This model is use in [emotion-conformer](https://github.com/poyu39/emotion-conformer)
|
amandacute/blockassist-bc-amphibious_plump_ram_1756791804
|
amandacute
| 2025-09-02T05:45:05Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"amphibious plump ram",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:44:58Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- amphibious plump ram
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
amandacute/blockassist-bc-amphibious_plump_ram_1756791813
|
amandacute
| 2025-09-02T05:45:00Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"amphibious plump ram",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:44:52Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- amphibious plump ram
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF
|
mradermacher
| 2025-09-02T05:44:35Z | 0 | 0 |
transformers
|
[
"transformers",
"gguf",
"llama",
"text-generation-inference",
"summarization",
"translation",
"question-answering",
"uz",
"en",
"dataset:yahma/alpaca-cleaned",
"dataset:behbudiy/alpaca-cleaned-uz",
"dataset:behbudiy/translation-instruction",
"base_model:behbudiy/Llama-3.1-8B-Instruct-Uz",
"base_model:quantized:behbudiy/Llama-3.1-8B-Instruct-Uz",
"license:llama3.1",
"endpoints_compatible",
"region:us",
"imatrix",
"conversational"
] |
question-answering
| 2025-09-02T04:57:01Z |
---
base_model: behbudiy/Llama-3.1-8B-Instruct-Uz
datasets:
- yahma/alpaca-cleaned
- behbudiy/alpaca-cleaned-uz
- behbudiy/translation-instruction
language:
- uz
- en
library_name: transformers
license: llama3.1
mradermacher:
readme_rev: 1
quantized_by: mradermacher
tags:
- llama
- text-generation-inference
- summarization
- translation
- question-answering
---
## About
<!-- ### quantize_version: 2 -->
<!-- ### output_tensor_quantised: 1 -->
<!-- ### convert_type: hf -->
<!-- ### vocab_type: -->
<!-- ### tags: nicoboss -->
<!-- ### quants: Q2_K IQ3_M Q4_K_S IQ3_XXS Q3_K_M small-IQ4_NL Q4_K_M IQ2_M Q6_K IQ4_XS Q2_K_S IQ1_M Q3_K_S IQ2_XXS Q3_K_L IQ2_XS Q5_K_S IQ2_S IQ1_S Q5_K_M Q4_0 IQ3_XS Q4_1 IQ3_S -->
<!-- ### quants_skip: -->
<!-- ### skip_mmproj: -->
weighted/imatrix quants of https://huggingface.co/behbudiy/Llama-3.1-8B-Instruct-Uz
<!-- provided-files -->
***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Llama-3.1-8B-Instruct-Uz-i1-GGUF).***
static quants are available at https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-GGUF
## Usage
If you are unsure how to use GGUF files, refer to one of [TheBloke's
READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
more details, including on how to concatenate multi-part files.
## Provided Quants
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
| Link | Type | Size/GB | Notes |
|:-----|:-----|--------:|:------|
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.imatrix.gguf) | imatrix | 0.1 | imatrix file (for creating your own qwuants) |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-IQ1_S.gguf) | i1-IQ1_S | 2.1 | for the desperate |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-IQ1_M.gguf) | i1-IQ1_M | 2.3 | mostly desperate |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 2.5 | |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-IQ2_XS.gguf) | i1-IQ2_XS | 2.7 | |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-IQ2_S.gguf) | i1-IQ2_S | 2.9 | |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-IQ2_M.gguf) | i1-IQ2_M | 3.0 | |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-Q2_K_S.gguf) | i1-Q2_K_S | 3.1 | very low quality |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-Q2_K.gguf) | i1-Q2_K | 3.3 | IQ3_XXS probably better |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 3.4 | lower quality |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-IQ3_XS.gguf) | i1-IQ3_XS | 3.6 | |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-Q3_K_S.gguf) | i1-Q3_K_S | 3.8 | IQ3_XS probably better |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-IQ3_S.gguf) | i1-IQ3_S | 3.8 | beats Q3_K* |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-IQ3_M.gguf) | i1-IQ3_M | 3.9 | |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-Q3_K_M.gguf) | i1-Q3_K_M | 4.1 | IQ3_S probably better |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-Q3_K_L.gguf) | i1-Q3_K_L | 4.4 | IQ3_M probably better |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-IQ4_XS.gguf) | i1-IQ4_XS | 4.5 | |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-Q4_0.gguf) | i1-Q4_0 | 4.8 | fast, low quality |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-IQ4_NL.gguf) | i1-IQ4_NL | 4.8 | prefer IQ4_XS |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-Q4_K_S.gguf) | i1-Q4_K_S | 4.8 | optimal size/speed/quality |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-Q4_K_M.gguf) | i1-Q4_K_M | 5.0 | fast, recommended |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-Q4_1.gguf) | i1-Q4_1 | 5.2 | |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-Q5_K_S.gguf) | i1-Q5_K_S | 5.7 | |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-Q5_K_M.gguf) | i1-Q5_K_M | 5.8 | |
| [GGUF](https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-Uz-i1-GGUF/resolve/main/Llama-3.1-8B-Instruct-Uz.i1-Q6_K.gguf) | i1-Q6_K | 6.7 | practically like static Q6_K |
Here is a handy graph by ikawrakow comparing some lower-quality quant
types (lower is better):

And here are Artefact2's thoughts on the matter:
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
## FAQ / Model Request
See https://huggingface.co/mradermacher/model_requests for some answers to
questions you might have and/or if you want some other model quantized.
## Thanks
I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
me use its servers and providing upgrades to my workstation to enable
this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.
<!-- end -->
|
mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF
|
mradermacher
| 2025-09-02T05:44:08Z | 0 | 0 |
transformers
|
[
"transformers",
"gguf",
"en",
"base_model:ArtusDev/L3.3-Electra-R1-70b-Elarablation",
"base_model:quantized:ArtusDev/L3.3-Electra-R1-70b-Elarablation",
"endpoints_compatible",
"region:us",
"imatrix",
"conversational"
] | null | 2025-09-01T06:49:18Z |
---
base_model: ArtusDev/L3.3-Electra-R1-70b-Elarablation
language:
- en
library_name: transformers
mradermacher:
readme_rev: 1
quantized_by: mradermacher
---
## About
<!-- ### quantize_version: 2 -->
<!-- ### output_tensor_quantised: 1 -->
<!-- ### convert_type: hf -->
<!-- ### vocab_type: -->
<!-- ### tags: nicoboss -->
<!-- ### quants: Q2_K IQ3_M Q4_K_S IQ3_XXS Q3_K_M small-IQ4_NL Q4_K_M IQ2_M Q6_K IQ4_XS Q2_K_S IQ1_M Q3_K_S IQ2_XXS Q3_K_L IQ2_XS Q5_K_S IQ2_S IQ1_S Q5_K_M Q4_0 IQ3_XS Q4_1 IQ3_S -->
<!-- ### quants_skip: -->
<!-- ### skip_mmproj: -->
weighted/imatrix quants of https://huggingface.co/ArtusDev/L3.3-Electra-R1-70b-Elarablation
<!-- provided-files -->
***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#L3.3-Electra-R1-70b-Elarablation-i1-GGUF).***
static quants are available at https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-GGUF
## Usage
If you are unsure how to use GGUF files, refer to one of [TheBloke's
READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
more details, including on how to concatenate multi-part files.
## Provided Quants
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
| Link | Type | Size/GB | Notes |
|:-----|:-----|--------:|:------|
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.imatrix.gguf) | imatrix | 0.1 | imatrix file (for creating your own qwuants) |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-IQ1_S.gguf) | i1-IQ1_S | 15.4 | for the desperate |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-IQ1_M.gguf) | i1-IQ1_M | 16.9 | mostly desperate |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 19.2 | |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-IQ2_XS.gguf) | i1-IQ2_XS | 21.2 | |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-IQ2_S.gguf) | i1-IQ2_S | 22.3 | |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-IQ2_M.gguf) | i1-IQ2_M | 24.2 | |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q2_K_S.gguf) | i1-Q2_K_S | 24.6 | very low quality |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q2_K.gguf) | i1-Q2_K | 26.5 | IQ3_XXS probably better |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 27.6 | lower quality |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-IQ3_XS.gguf) | i1-IQ3_XS | 29.4 | |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-IQ3_S.gguf) | i1-IQ3_S | 31.0 | beats Q3_K* |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q3_K_S.gguf) | i1-Q3_K_S | 31.0 | IQ3_XS probably better |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-IQ3_M.gguf) | i1-IQ3_M | 32.0 | |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q3_K_M.gguf) | i1-Q3_K_M | 34.4 | IQ3_S probably better |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q3_K_L.gguf) | i1-Q3_K_L | 37.2 | IQ3_M probably better |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-IQ4_XS.gguf) | i1-IQ4_XS | 38.0 | |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q4_0.gguf) | i1-Q4_0 | 40.2 | fast, low quality |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q4_K_S.gguf) | i1-Q4_K_S | 40.4 | optimal size/speed/quality |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q4_K_M.gguf) | i1-Q4_K_M | 42.6 | fast, recommended |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q4_1.gguf) | i1-Q4_1 | 44.4 | |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q5_K_S.gguf) | i1-Q5_K_S | 48.8 | |
| [GGUF](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q5_K_M.gguf) | i1-Q5_K_M | 50.0 | |
| [PART 1](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q6_K.gguf.part1of2) [PART 2](https://huggingface.co/mradermacher/L3.3-Electra-R1-70b-Elarablation-i1-GGUF/resolve/main/L3.3-Electra-R1-70b-Elarablation.i1-Q6_K.gguf.part2of2) | i1-Q6_K | 58.0 | practically like static Q6_K |
Here is a handy graph by ikawrakow comparing some lower-quality quant
types (lower is better):

And here are Artefact2's thoughts on the matter:
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
## FAQ / Model Request
See https://huggingface.co/mradermacher/model_requests for some answers to
questions you might have and/or if you want some other model quantized.
## Thanks
I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
me use its servers and providing upgrades to my workstation to enable
this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.
<!-- end -->
|
aXsalll/blockassist-bc-chattering_galloping_ape_1756791785
|
aXsalll
| 2025-09-02T05:43:54Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"chattering galloping ape",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:43:30Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- chattering galloping ape
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
yanTemp/qwen2.5-7b-instruct-trl-sft-ChartQA
|
yanTemp
| 2025-09-02T05:43:24Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"generated_from_trainer",
"trl",
"sft",
"base_model:Qwen/Qwen2.5-VL-7B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-VL-7B-Instruct",
"endpoints_compatible",
"region:us"
] | null | 2025-08-31T08:59:08Z |
---
base_model: Qwen/Qwen2.5-VL-7B-Instruct
library_name: transformers
model_name: qwen2.5-7b-instruct-trl-sft-ChartQA
tags:
- generated_from_trainer
- trl
- sft
licence: license
---
# Model Card for qwen2.5-7b-instruct-trl-sft-ChartQA
This model is a fine-tuned version of [Qwen/Qwen2.5-VL-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="yanTemp/qwen2.5-7b-instruct-trl-sft-ChartQA", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
This model was trained with SFT.
### Framework versions
- TRL: 0.22.1
- Transformers: 4.56.0
- Pytorch: 2.6.0+cu126
- Datasets: 4.0.0
- Tokenizers: 0.22.0
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
omerbkts/blockassist-bc-keen_fast_giraffe_1756791759
|
omerbkts
| 2025-09-02T05:43:03Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"keen fast giraffe",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:42:58Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- keen fast giraffe
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
crystal0112/air-purifier-function-call-merged
|
crystal0112
| 2025-09-02T05:43:00Z | 0 | 0 | null |
[
"safetensors",
"llama",
"region:us"
] | null | 2025-09-02T05:15:56Z |
# llama_function_call_merged_baseline
이 모델은 한국어 음성 명령을 Function Call로 변환하기 위해 fine-tuning된 Llama 3.2 1B 모델입니다.
## 모델 정보
- **Base Model**: Llama 3.2 1B Instruct
- **Fine-tuning**: LoRA (Low-Rank Adaptation)
- **Task**: Function Call Generation
- **Language**: Korean
## 사용법
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
# 토크나이저 로드
tokenizer = AutoTokenizer.from_pretrained("llama_function_call_merged_baseline")
# 베이스 모델 로드
base_model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.2-1B-Instruct")
# LoRA 모델 로드
model = PeftModel.from_pretrained(base_model, "llama_function_call_merged_baseline")
```
## 지원하는 함수들
- MR: 사용자와의 대화 응답
- MO: 날씨 정보 조회
- IO: 공기질 정보 조회
- PG: 공기청정기 제어
- FF: 특정 위치 공기청정
- NK: 풍량/팬 속도 제어
- WN: 밝기/조명 제어
- ZV: 음량/볼륨 제어
- 기타 35개 함수...
## 라이선스
이 모델은 학습 데이터와 동일한 라이선스를 따릅니다.
|
Rudra-madlads/blockassist-bc-jumping_swift_gazelle_1756791717
|
Rudra-madlads
| 2025-09-02T05:42:56Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"jumping swift gazelle",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:42:33Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- jumping swift gazelle
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Rusik11/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-sizable_hunting_cockroach
|
Rusik11
| 2025-09-02T05:42:49Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"rl-swarm",
"genrl-swarm",
"grpo",
"gensyn",
"I am sizable_hunting_cockroach",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-09-02T05:42:34Z |
---
library_name: transformers
tags:
- rl-swarm
- genrl-swarm
- grpo
- gensyn
- I am sizable_hunting_cockroach
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
hssnjfry/blockassist-bc-climbing_pouncing_dragonfly_1756791675
|
hssnjfry
| 2025-09-02T05:42:38Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"climbing pouncing dragonfly",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:42:13Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- climbing pouncing dragonfly
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
allenai/MolmoAct-7B-D-Pretrain-RT-1-0812
|
allenai
| 2025-09-02T05:41:54Z | 492 | 4 |
transformers
|
[
"transformers",
"safetensors",
"molmoact",
"image-text-to-text",
"molmo",
"olmo",
"reasoning",
"vla",
"robotics",
"manipulation",
"custom_code",
"en",
"arxiv:2508.07917",
"base_model:Qwen/Qwen2.5-7B",
"base_model:finetune:Qwen/Qwen2.5-7B",
"license:apache-2.0",
"region:us"
] |
robotics
| 2025-08-11T03:53:16Z |
---
base_model:
- Qwen/Qwen2.5-7B
- google/siglip2-so400m-patch14-384
language:
- en
library_name: transformers
license: apache-2.0
pipeline_tag: robotics
tags:
- molmoact
- molmo
- olmo
- reasoning
- vla
- robotics
- manipulation
paper: 2508.07917
---
<img src="molmoact_logo.svg" alt="MolmoAct Logo" style="width: auto; height: 50px;">
# MolmoAct 7B-D Pretrain RT-1
MolmoAct is a fully open-source action reasoning model for robotic manipulation developed by the Allen Institute for AI, as described in their paper [MolmoAct: Action Reasoning Models that can Reason in Space](https://huggingface.co/papers/2508.07917).
MolmoAct is trained on a subset of OXE and MolmoAct Dataset, a dataset with 10k high-quality trajectories of a single-arm Franka robot performing 93 unique manipulation tasks in both home and tabletop environments. It has state-of-the-art performance among vision-language-action models on multiple benchmarks while being fully open-source. You can find all models in the MolmoAct family [here](https://huggingface.co/collections/allenai/molmoact-689697591a3936fba38174d7).
**Learn more about MolmoAct** in our announcement [blog post](https://allenai.org/blog/molmoact) or the [paper](https://arxiv.org/abs/2508.07917).
**MolmoAct 7B-D Pretrain RT-1** is based on [Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) and uses [SigLip2](https://huggingface.co/google/siglip2-so400m-patch14-384) as the vision backbone, which is initialized using Molmo's pre-training approach. It is first pre-trained on MolmoAct's [Pre-training Mixture](https://huggingface.co/datasets/allenai/MolmoAct-Pretraining-Mixture), and then fine-tuned on RT-1 data using the same configuration of mid-training. This model is intended to be used for replicating our fine-tuned results on SimplerEnv (Google Robot).
This checkpoint is a **preview** of the MolmoAct release. All artifacts used in creating MolmoAct (data, training code, evaluations, intermediate checkpoints) will be made available at a later date, furthering our commitment to open-source AI development and reproducibility.
**Update:** Checkpoints are now stored in FP32 (previously BF16). The model was trained in FP32, so publishing FP32 weights aligns with training and enables fine-tuning or continued training directly from this repo. For inference, you can still run BF16 by casting at load, which is what we did for evaluations. See more in the [instructions](#quick-start) below.
Quick links:
- 📂 [All Models](https://huggingface.co/collections/allenai/molmoact-689697591a3936fba38174d7)
- 📂 [All Data](https://huggingface.co/collections/allenai/molmoact-data-mixture-6897e583e13b6c2cf3ea2b80)
- 📃 [Paper](https://arxiv.org/abs/2508.07917)
- 💻 [Code](https://github.com/allenai/MolmoAct)
- 🎥 [Blog Post](https://allenai.org/blog/molmoact)
- 🎥 [Video](https://youtu.be/-_wag1X25OE?si=Xi_kUaJTmcQBx1f6)
## Quick Start
To run MolmoAct, first install dependencies:
```bash
pip install einops torchvision accelerate
pip install transformers==4.52
```
Then, follow these steps:
```python
from transformers import AutoProcessor, AutoModelForImageTextToText
import torch
from PIL import Image
import requests
from io import BytesIO
ckpt = "allenai/MolmoAct-7B-D-Pretrain-RT-1-0812"
# load the processor
processor = AutoProcessor.from_pretrained(
ckpt,
trust_remote_code=True,
torch_dtype="bfloat16",
device_map="auto",
padding_side="left",
)
# load the model
model = AutoModelForImageTextToText.from_pretrained(
ckpt,
trust_remote_code=True,
torch_dtype="bfloat16",
device_map="auto",
)
# task instruction
instruction = "pick orange can"
# strictly follow this reasoning prompt
prompt = (
f"The task is {instruction}. "
"What is the action that the robot should take. "
f"To figure out the action that the robot should take to {instruction}, "
"let's think through it step by step. "
"First, what is the depth map for this image? "
"Second, what is the trajectory of the end effector? "
"Based on the depth map of the image and the trajectory of the end effector, "
"what is the action that the robot should take?"
)
# apply chat template
text = processor.apply_chat_template(
[
{
"role": "user",
"content": [dict(type="text", text=prompt)]
}
],
tokenize=False,
add_generation_prompt=True,
)
# image observation
url = "https://huggingface.co/allenai/MolmoAct-7B-D-Pretrain-0812/resolve/main/example.png"
r = requests.get(url, headers={"User-Agent": "python-requests"}, timeout=30)
r.raise_for_status()
img = Image.open(BytesIO(r.content)).convert("RGB")
imgs = [img]
# process the image and text
inputs = processor(
images=[imgs],
text=text,
padding=True,
return_tensors="pt",
)
# move inputs to the correct device
inputs = {k: v.to(model.device) for k, v in inputs.items()}
# generate output
with torch.inference_mode():
with torch.autocast("cuda", enabled=True, dtype=torch.bfloat16):
generated_ids = model.generate(**inputs, max_new_tokens=256)
# only get generated tokens; decode them to text
generated_tokens = generated_ids[:, inputs['input_ids'].size(1):]
generated_text = processor.batch_decode(generated_tokens, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0]
# print the generated text
print(f"generated text: {generated_text}")
# >>> The depth map of the image is ... The trajectory of the end effector is ...
# Based on these information, the action that the robot should take is ...
# parse out all depth perception tokens
depth = model.parse_depth(generated_text)
print(f"generated depth perception tokens: {depth}")
# >>> [ "<DEPTH_START><DEPTH_1><DEPTH_2>...<DEPTH_END>" ]
# parse out all visual reasoning traces
trace = model.parse_trace(generated_text)
print(f"generated visual reasoning trace: {trace}")
# >>> [ [[242, 115], [140, 77], [94, 58], [140, 44], [153, 26]]] ]
# parse out all actions, unnormalizing with key of fractal20220817_data
action = model.parse_action(generated_text, unnorm_key="fractal20220817_data")
print(f"generated action: {action}")
# >>> [ [0.0732076061122558, 0.08228153779226191, -0.027760173818644346,
# 0.15932856272248652, -0.09686601126895233, 0.043916773912953344,
# 0.996078431372549] ]
```
## License and Use
This model is licensed under Apache 2.0. It is intended for research and educational use.
For more information, please see our [Responsible Use Guidelines](https://allenai.org/responsible-use).
## Model and Hardware Safety
MolmoAct offers the ability to inspect a visual trace of its intended actions in space before they occur, allowing users to ensure safe behavior by proactively auditing and adjusting the actions of any hardware acting under the model’s instructions. MolmoAct’s action space is bounded within the data provided, and compliance is built into the model to prevent excessive force when resistance is detected. Please follow the hardware manufacturer’s guidelines when using this model with a robot and perform all operations in a safely configured environment.
## Citation
```bibtex
@misc{molmoact2025,
title={MolmoAct: Action Reasoning Models that can Reason in Space},
author={Jason Lee and Jiafei Duan and Haoquan Fang and Yuquan Deng and Shuo Liu and Boyang Li and Bohan Fang and Jieyu Zhang and Yi Ru Wang and Sangho Lee and Winson Han and Wilbert Pumacay and Angelica Wu and Rose Hendrix and Karen Farley and Eli VanderBilt and Ali Farhadi and Dieter Fox and Ranjay Krishna},
year={2025},
eprint={2508.07917},
archivePrefix={arXiv},
primaryClass={cs.RO},
url={https://arxiv.org/abs/2508.07917}
}
```
|
the-usan/urdu-crime-dehshatgardi-v1
|
the-usan
| 2025-09-02T05:41:26Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"roberta",
"text-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-classification
| 2025-09-02T05:41:15Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
akirafudo/blockassist-bc-keen_fast_giraffe_1756791644
|
akirafudo
| 2025-09-02T05:41:03Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"keen fast giraffe",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:41:00Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- keen fast giraffe
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
PranayO98/Smoothie-Qwen3-1.7B-Gensyn-Swarm-sprightly_frisky_goose
|
PranayO98
| 2025-09-02T05:40:20Z | 110 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen3",
"text-generation",
"rl-swarm",
"genrl-swarm",
"grpo",
"gensyn",
"I am sprightly_frisky_goose",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-08-29T18:40:57Z |
---
library_name: transformers
tags:
- rl-swarm
- genrl-swarm
- grpo
- gensyn
- I am sprightly_frisky_goose
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
pidbu/blockassist-bc-whistling_alert_shrew_1756791510
|
pidbu
| 2025-09-02T05:39:50Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"whistling alert shrew",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:39:13Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- whistling alert shrew
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
vwzyrraz7l/blockassist-bc-tall_hunting_vulture_1756790077
|
vwzyrraz7l
| 2025-09-02T05:39:04Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"tall hunting vulture",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:39:01Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- tall hunting vulture
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Abhisawkar/ship_classifier
|
Abhisawkar
| 2025-09-02T05:38:36Z | 0 | 0 | null |
[
"license:apache-2.0",
"region:us"
] | null | 2025-09-02T05:38:36Z |
---
license: apache-2.0
---
|
EnriqueSolarte/qwen2.5-VL-7B-instruct-amazon-description
|
EnriqueSolarte
| 2025-09-02T05:38:07Z | 0 | 0 |
transformers
|
[
"transformers",
"tensorboard",
"safetensors",
"generated_from_trainer",
"trl",
"sft",
"base_model:Qwen/Qwen2.5-VL-7B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-VL-7B-Instruct",
"endpoints_compatible",
"region:us"
] | null | 2025-09-02T05:19:13Z |
---
base_model: Qwen/Qwen2.5-VL-7B-Instruct
library_name: transformers
model_name: qwen2.5-VL-7B-instruct-amazon-description
tags:
- generated_from_trainer
- trl
- sft
licence: license
---
# Model Card for qwen2.5-VL-7B-instruct-amazon-description
This model is a fine-tuned version of [Qwen/Qwen2.5-VL-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="EnriqueSolarte/qwen2.5-VL-7B-instruct-amazon-description", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
This model was trained with SFT.
### Framework versions
- TRL: 0.22.1
- Transformers: 4.56.0
- Pytorch: 2.8.0
- Datasets: 4.0.0
- Tokenizers: 0.22.0
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
yanTemp/qwen2-7b-instruct-trl-sft-ChartQA
|
yanTemp
| 2025-09-02T05:37:15Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"generated_from_trainer",
"trl",
"sft",
"base_model:Qwen/Qwen2.5-VL-7B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-VL-7B-Instruct",
"endpoints_compatible",
"region:us"
] | null | 2025-09-02T05:33:10Z |
---
base_model: Qwen/Qwen2.5-VL-7B-Instruct
library_name: transformers
model_name: qwen2-7b-instruct-trl-sft-ChartQA
tags:
- generated_from_trainer
- trl
- sft
licence: license
---
# Model Card for qwen2-7b-instruct-trl-sft-ChartQA
This model is a fine-tuned version of [Qwen/Qwen2.5-VL-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="yanTemp/qwen2-7b-instruct-trl-sft-ChartQA", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
This model was trained with SFT.
### Framework versions
- TRL: 0.22.1
- Transformers: 4.56.0
- Pytorch: 2.6.0+cu126
- Datasets: 4.0.0
- Tokenizers: 0.22.0
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
DavidAU/Qwen3-21B-QiMing-V1.0-TOTAL-RECALL
|
DavidAU
| 2025-09-02T05:36:38Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen3",
"text-generation",
"programming",
"code generation",
"code",
"codeqwen",
"moe",
"coding",
"coder",
"qwen2",
"chat",
"qwen",
"qwen-coder",
"finetune",
"brainstorm 20x",
"brainstorm",
"optional thinking",
"creative",
"all use cases",
"QiMing",
"QiMing-holos",
"bagua",
"decision-making",
"strategic-analysis",
"cognitive-architecture",
"philosophy-driven-ai",
"conversational",
"en",
"fr",
"zh",
"de",
"base_model:aifeifei798/QiMing-v1.0-14B",
"base_model:finetune:aifeifei798/QiMing-v1.0-14B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-09-02T01:27:11Z |
---
license: apache-2.0
library_name: transformers
language:
- en
- fr
- zh
- de
tags:
- programming
- code generation
- code
- codeqwen
- programming
- code generation
- code
- codeqwen
- moe
- coding
- coder
- qwen2
- chat
- qwen
- qwen-coder
- chat
- qwen
- qwen-coder
- qwen3
- finetune
- brainstorm 20x
- brainstorm
- optional thinking
- creative
- all use cases
- QiMing
- QiMing-holos
- bagua
- decision-making
- strategic-analysis
- cognitive-architecture
- chat
- philosophy-driven-ai
base_model:
- aifeifei798/QiMing-v1.0-14B
pipeline_tag: text-generation
---
<h2>Qwen3-21B-QiMing-V1.0-TOTAL-RECALL</h2>
QiMing-v1.0-14B with Brainstorm 20x (by DavidAU) applied.
Part of project to benchmark Brainstorm versions.
[ more to come ]
|
liukevin666/blockassist-bc-yawning_striped_cassowary_1756791311
|
liukevin666
| 2025-09-02T05:36:32Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"yawning striped cassowary",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:36:09Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- yawning striped cassowary
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
maxibillion1975/blockassist-bc-iridescent_squeaky_sandpiper_1756789807
|
maxibillion1975
| 2025-09-02T05:35:43Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"iridescent squeaky sandpiper",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:35:40Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- iridescent squeaky sandpiper
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Rudra-madlads/blockassist-bc-jumping_swift_gazelle_1756791254
|
Rudra-madlads
| 2025-09-02T05:35:11Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"jumping swift gazelle",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:34:50Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- jumping swift gazelle
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
akirafudo/blockassist-bc-keen_fast_giraffe_1756791258
|
akirafudo
| 2025-09-02T05:34:40Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"keen fast giraffe",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-02T05:34:35Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- keen fast giraffe
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Subsets and Splits
Filtered Qwen2.5 Distill Models
Identifies specific configurations of models by filtering cards that contain 'distill', 'qwen2.5', '7b' while excluding certain base models and incorrect model ID patterns, uncovering unique model variants.
Filtered Model Cards Count
Finds the count of entries with specific card details that include 'distill', 'qwen2.5', '7b' but exclude certain base models, revealing valuable insights about the dataset's content distribution.
Filtered Distill Qwen 7B Models
Filters for specific card entries containing 'distill', 'qwen', and '7b', excluding certain strings and patterns, to identify relevant model configurations.
Filtered Qwen-7b Model Cards
The query performs a detailed filtering based on specific keywords and excludes certain entries, which could be useful for identifying a specific subset of cards but does not provide deeper insights or trends.
Filtered Qwen 7B Model Cards
The query filters for specific terms related to "distilled" or "distill", "qwen", and "7b" in the 'card' column but excludes certain base models, providing a limited set of entries for further inspection.
Qwen 7B Distilled Models
The query provides a basic filtering of records to find specific card names that include keywords related to distilled Qwen 7b models, excluding a particular base model, which gives limited insight but helps in focusing on relevant entries.
Qwen 7B Distilled Model Cards
The query filters data based on specific keywords in the modelId and card fields, providing limited insight primarily useful for locating specific entries rather than revealing broad patterns or trends.
Qwen 7B Distilled Models
Finds all entries containing the terms 'distilled', 'qwen', and '7b' in a case-insensitive manner, providing a filtered set of records but without deeper analysis.
Distilled Qwen 7B Models
The query filters for specific model IDs containing 'distilled', 'qwen', and '7b', providing a basic retrieval of relevant entries but without deeper analysis or insight.
Filtered Model Cards with Distill Qwen2.
Filters and retrieves records containing specific keywords in the card description while excluding certain phrases, providing a basic count of relevant entries.
Filtered Model Cards with Distill Qwen 7
The query filters specific variations of card descriptions containing 'distill', 'qwen', and '7b' while excluding a particular base model, providing limited but specific data retrieval.
Distill Qwen 7B Model Cards
The query filters and retrieves rows where the 'card' column contains specific keywords ('distill', 'qwen', and '7b'), providing a basic filter result that can help in identifying specific entries.