modelId
stringlengths 5
139
| author
stringlengths 2
42
| last_modified
timestamp[us, tz=UTC]date 2020-02-15 11:33:14
2025-08-30 00:39:23
| downloads
int64 0
223M
| likes
int64 0
11.7k
| library_name
stringclasses 526
values | tags
listlengths 1
4.05k
| pipeline_tag
stringclasses 55
values | createdAt
timestamp[us, tz=UTC]date 2022-03-02 23:29:04
2025-08-30 00:39:08
| card
stringlengths 11
1.01M
|
---|---|---|---|---|---|---|---|---|---|
kowndinya23/ultrafeedback_binarized-alpaca-llama-3-1b-2-epochs-alpha-0.4-beta-1-2-epochs
|
kowndinya23
| 2025-06-08T06:17:06Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"trl",
"dpo",
"conversational",
"dataset:trl-lib/ultrafeedback_binarized",
"arxiv:2305.18290",
"base_model:kowndinya23/alpaca-cleaned-llama-3-1b-2-epochs-alpha-0.4-beta-1",
"base_model:finetune:kowndinya23/alpaca-cleaned-llama-3-1b-2-epochs-alpha-0.4-beta-1",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T05:21:41Z |
---
base_model: kowndinya23/alpaca-cleaned-llama-3-1b-2-epochs-alpha-0.4-beta-1
datasets: trl-lib/ultrafeedback_binarized
library_name: transformers
model_name: ultrafeedback_binarized-alpaca-llama-3-1b-2-epochs-alpha-0.4-beta-1-2-epochs
tags:
- generated_from_trainer
- trl
- dpo
licence: license
---
# Model Card for ultrafeedback_binarized-alpaca-llama-3-1b-2-epochs-alpha-0.4-beta-1-2-epochs
This model is a fine-tuned version of [kowndinya23/alpaca-cleaned-llama-3-1b-2-epochs-alpha-0.4-beta-1](https://huggingface.co/kowndinya23/alpaca-cleaned-llama-3-1b-2-epochs-alpha-0.4-beta-1) on the [trl-lib/ultrafeedback_binarized](https://huggingface.co/datasets/trl-lib/ultrafeedback_binarized) dataset.
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="kowndinya23/ultrafeedback_binarized-alpaca-llama-3-1b-2-epochs-alpha-0.4-beta-1-2-epochs", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://adobesensei.wandb.io/hrenduchinta/huggingface/runs/a9ncbcla)
This model was trained with DPO, a method introduced in [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://huggingface.co/papers/2305.18290).
### Framework versions
- TRL: 0.18.1
- Transformers: 4.52.4
- Pytorch: 2.7.0
- Datasets: 3.6.0
- Tokenizers: 0.21.1
## Citations
Cite DPO as:
```bibtex
@inproceedings{rafailov2023direct,
title = {{Direct Preference Optimization: Your Language Model is Secretly a Reward Model}},
author = {Rafael Rafailov and Archit Sharma and Eric Mitchell and Christopher D. Manning and Stefano Ermon and Chelsea Finn},
year = 2023,
booktitle = {Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, December 10 - 16, 2023},
url = {http://papers.nips.cc/paper_files/paper/2023/hash/a85b405ed65c6477a4fe8302b5e06ce7-Abstract-Conference.html},
editor = {Alice Oh and Tristan Naumann and Amir Globerson and Kate Saenko and Moritz Hardt and Sergey Levine},
}
```
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
AdityaMayukhSom/Qwen3-1.7B-HyperMixSub
|
AdityaMayukhSom
| 2025-06-08T06:12:37Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"generated_from_trainer",
"unsloth",
"trl",
"sft",
"base_model:unsloth/Qwen3-1.7B-unsloth-bnb-4bit",
"base_model:finetune:unsloth/Qwen3-1.7B-unsloth-bnb-4bit",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T06:12:23Z |
---
base_model: unsloth/Qwen3-1.7B-unsloth-bnb-4bit
library_name: transformers
model_name: Qwen3-1.7B-HyperMixSub
tags:
- generated_from_trainer
- unsloth
- trl
- sft
licence: license
---
# Model Card for Qwen3-1.7B-HyperMixSub
This model is a fine-tuned version of [unsloth/Qwen3-1.7B-unsloth-bnb-4bit](https://huggingface.co/unsloth/Qwen3-1.7B-unsloth-bnb-4bit).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="AdityaMayukhSom/Qwen3-1.7B-HyperMixSub", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
This model was trained with SFT.
### Framework versions
- TRL: 0.18.0
- Transformers: 4.51.3
- Pytorch: 2.7.0+cu128
- Datasets: 3.6.0
- Tokenizers: 0.21.1
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
appellaai/gemma-3-british-2
|
appellaai
| 2025-06-08T06:11:52Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"gemma3",
"image-text-to-text",
"text-generation-inference",
"unsloth",
"conversational",
"en",
"base_model:unsloth/gemma-3-12b-it-unsloth-bnb-4bit",
"base_model:finetune:unsloth/gemma-3-12b-it-unsloth-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
image-text-to-text
| 2025-06-08T06:00:05Z |
---
base_model: unsloth/gemma-3-12b-it-unsloth-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- gemma3
license: apache-2.0
language:
- en
---
# Uploaded finetuned model
- **Developed by:** appellaai
- **License:** apache-2.0
- **Finetuned from model :** unsloth/gemma-3-12b-it-unsloth-bnb-4bit
This gemma3 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
pch11/waterfrontpark
|
pch11
| 2025-06-08T06:10:34Z | 0 | 0 | null |
[
"license:apache-2.0",
"region:us"
] | null | 2025-06-08T06:07:56Z |
---
license: apache-2.0
---
|
tumuyan2/realsr-models
|
tumuyan2
| 2025-06-08T06:10:31Z | 0 | 13 | null |
[
"Super-Resolution",
"ESRGAN",
"Waifu2x",
"region:us"
] | null | 2024-03-11T12:21:57Z |
---
tags:
- Super-Resolution
- ESRGAN
- Waifu2x
---
# Extra Models for RealSR-NCNN-Android
[Available models](https://huggingface.co/tumuyan2/realsr-models/tree/main)
These are some models prepared for [RealSR-NCNN-Android](https://github.com/tumuyan2/RealSR-NCNN-Android).
You can download whichever directory you need.
## Waifu2x Models
https://github.com/nihui/waifu2x-ncnn-vulkan
| Model Folder | Remark |
| ----------------------------------- | ------ |
| models-upconv_7_photo | |
| models-upconv_7_anime_style_art_rgb | |
| models-cunet | |
## SRMD Models
https://github.com/nihui/srmd-ncnn-vulkan
| Model Folder | Remark |
| ------------ | ------ |
| models-srmd | |
## RealSR Models
https://github.com/nihui/realsr-ncnn-vulkan
| Model Folder | Remark |
| ---------------- | ------ |
| models-DF2K | |
| models-DF2K_JPEG | |
## ESRGAN Models
https://upscale.wiki/wiki/Model_Database
| Model Folder | Scale | Remark | Author | Source |
| ---------------------------------------------------- |:-----:|:-----------------------------:|:-------------------------------------:| ----------------------------------------------------------------------------------------------------------- |
| models-ESRGAN-1x_Fatality_NoiseToner-sharpen_denoise | 1x | sharpen & denoise | DinJerr | https://1drv.ms/u/s!Aip-EMByJHY2gYQUcbSTFgrdwtMjQA?e=A5p6lH |
| models-ESRGAN-1x_NMKD-YandereInpaint_375000_G | 1x | Inpainting | Nmkd | https://icedrive.net/1/43GNBihZyi |
| models-ESRGAN-1x_sudo_inpaint_PartialConv2D_424000_G | 1x | Inpainting | sudo rm -rf / --no-preserve-root#8353 | https://e.pcloud.link/publink/show?code=kZQOu7ZldzmFyMPUcFNGkEvwqOxQ8Bl3CeX |
| models-ESRGAN-8x_NMKD-Typescale_175k | 8X | Text | NMKD | https://icedrive.net/s/43GNBihZyi |
| models-ESRGAN-BSTexty_86000G | 2X | Text | BlackScout | https://drive.google.com/file/d/15ovbadCoYs7q8nSd5Mq02PqBOpwiBkoS/view |
| models-ESRGAN-AnimeSharp | 4x | Anime or Text | Kim2091 | https://mega.nz/folder/rdpkjZzC#eUXPed_vntJKLrB0wpeJ-w |
| models-ESRGAN-AnimeSharpLite | 4x | Anime | Kim2091 | https://mega.nz/folder/bEoRQIRR#kEsaVHtwRL9vwfa5k2osyQ |
| models-ESRGAN-FatePlusLite | 4x | Anime PSP games | Kim2091 | https://mega.nz/folder/zRYh3SII#QIm6T-rzhxjBLeYF1zSDpg |
| models-ESRGAN-Dropout | 2x | Anime | sudo | https://e1.pcloud.link/publink/show?code=kZ7rGRZW2IcOpNMQeXDTTRQ4aPVBFyyJV5X |
| models-ESRGAN-Remacri | 4x | General | Foolhardy | https://u.pcloud.link/publink/show?code=kZgSLsXZ0M1fT3kFGfRXg2tNtoUgbSI4kcSy |
| models-ESRGAN-UltraMix_Balanced | 4x | | Kim2091 | https://mega.nz/folder/3Jo2AAAa#4CGEwUM0dKu3kkaJa-qUIA |
| models-ESRGAN-SourceBook_v1 | 4x | Book | tumuyan | https://github.com/tumuyan/SourceBook-Dataset |
| models-ESRGAN-4xHFA2k | 4x | Anime | Phhofm | https://github.com/Phhofm/models/tree/main/4xHFA2k |
| models-ESRGAN-Nomos8kSC | 4x | Photo | Phhofm | https://github.com/Phhofm/models/tree/main/4xNomos8kSC |
| models-ESRGAN-UltraSharp-fp16 | 4x | Anime | Kim2091 | https://mega.nz/folder/qZRBmaIY#nIG8KyWFcGNTuMX_XNbJ_g |
| models-RealeSR-general-v3 | 4x | General | Xinntao | https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.5.0/realesr-general-x4v3.pth |
| models-ESRGAN-TGHQFace | 8x | Face | TorrentGuy | https://drive.google.com/uc?export=download&confirm=1&id=1OyOJIW224hBhb-aTCbuUQb0qzKmE4oH6 |
| models-ESRGAN-WTP-ColorDS-fp16 | 4x | remove screentone / halftones | umzi.x.dead | https://objectstorage.us-phoenix-1.oraclecloud.com/n/ax6ygfvpvzka/b/open-modeldb-files/o/4x-WTP-ColorDS.pth |
## MNN-SR models
in folder `models-MNNSR`
| Model Name | Scale | Remark | Author | Source |
| ------------------------------------------ |:-----:|:-----------------------------:|:-----------:| ----------------------------------------------------------------------------------------------------------- |
| 4x-WTP-ColorDS_fp16 | 4x | remove screentone / halftones | umzi.x.dead | https://objectstorage.us-phoenix-1.oraclecloud.com/n/ax6ygfvpvzka/b/open-modeldb-files/o/4x-WTP-ColorDS.pth |
| MoeSR-ESRGAN-jp_Illustration-fix2-x4.mnn | 4x | jp Illustration | luoyily | https://github.com/TeamMoeAI/MoeSR/releases/download/v1.0.0/MoeSR.models.RealESRGAN.7z |
| MoeSR-ESRGAN-jp_Illustration-fix1-d-x4.mnn | 4x | jp Illustration | luoyily | https://github.com/TeamMoeAI/MoeSR/releases/download/v1.0.0/MoeSR.models.RealESRGAN.7z |
|
looklook123/wenyanwen
|
looklook123
| 2025-06-08T06:08:31Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"llama",
"trl",
"en",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T06:08:15Z |
---
base_model: unsloth/meta-llama-3.1-8b-instruct-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- trl
license: apache-2.0
language:
- en
---
# Uploaded model
- **Developed by:** looklook123
- **License:** apache-2.0
- **Finetuned from model :** unsloth/meta-llama-3.1-8b-instruct-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF
|
mradermacher
| 2025-06-08T06:00:12Z | 0 | 0 |
transformers
|
[
"transformers",
"gguf",
"en",
"dataset:jondurbin/gutenberg-dpo-v0.1",
"dataset:nbeerbower/gutenberg2-dpo",
"dataset:nbeerbower/gutenberg-moderne-dpo",
"dataset:nbeerbower/synthetic-fiction-dpo",
"dataset:nbeerbower/Arkhaios-DPO",
"dataset:nbeerbower/Purpura-DPO",
"dataset:nbeerbower/Schule-DPO",
"base_model:nbeerbower/Qwen3-Gutenberg-Encore-14B",
"base_model:quantized:nbeerbower/Qwen3-Gutenberg-Encore-14B",
"license:apache-2.0",
"endpoints_compatible",
"region:us",
"imatrix",
"conversational"
] | null | 2025-06-07T14:25:27Z |
---
base_model: nbeerbower/Qwen3-Gutenberg-Encore-14B
datasets:
- jondurbin/gutenberg-dpo-v0.1
- nbeerbower/gutenberg2-dpo
- nbeerbower/gutenberg-moderne-dpo
- nbeerbower/synthetic-fiction-dpo
- nbeerbower/Arkhaios-DPO
- nbeerbower/Purpura-DPO
- nbeerbower/Schule-DPO
language:
- en
library_name: transformers
license: apache-2.0
quantized_by: mradermacher
---
## About
<!-- ### quantize_version: 2 -->
<!-- ### output_tensor_quantised: 1 -->
<!-- ### convert_type: hf -->
<!-- ### vocab_type: -->
<!-- ### tags: nicoboss -->
weighted/imatrix quants of https://huggingface.co/nbeerbower/Qwen3-Gutenberg-Encore-14B
<!-- provided-files -->
static quants are available at https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF
## Usage
If you are unsure how to use GGUF files, refer to one of [TheBloke's
READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
more details, including on how to concatenate multi-part files.
## Provided Quants
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
| Link | Type | Size/GB | Notes |
|:-----|:-----|--------:|:------|
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-IQ1_S.gguf) | i1-IQ1_S | 3.7 | for the desperate |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-IQ1_M.gguf) | i1-IQ1_M | 3.9 | mostly desperate |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 4.4 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-IQ2_XS.gguf) | i1-IQ2_XS | 4.8 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-IQ2_S.gguf) | i1-IQ2_S | 5.1 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-IQ2_M.gguf) | i1-IQ2_M | 5.4 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-Q2_K_S.gguf) | i1-Q2_K_S | 5.5 | very low quality |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-Q2_K.gguf) | i1-Q2_K | 5.9 | IQ3_XXS probably better |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 6.0 | lower quality |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-IQ3_XS.gguf) | i1-IQ3_XS | 6.5 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-Q3_K_S.gguf) | i1-Q3_K_S | 6.8 | IQ3_XS probably better |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-IQ3_S.gguf) | i1-IQ3_S | 6.8 | beats Q3_K* |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-IQ3_M.gguf) | i1-IQ3_M | 7.0 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-Q3_K_M.gguf) | i1-Q3_K_M | 7.4 | IQ3_S probably better |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-Q3_K_L.gguf) | i1-Q3_K_L | 8.0 | IQ3_M probably better |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-IQ4_XS.gguf) | i1-IQ4_XS | 8.2 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-IQ4_NL.gguf) | i1-IQ4_NL | 8.6 | prefer IQ4_XS |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-Q4_0.gguf) | i1-Q4_0 | 8.6 | fast, low quality |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-Q4_K_S.gguf) | i1-Q4_K_S | 8.7 | optimal size/speed/quality |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-Q4_K_M.gguf) | i1-Q4_K_M | 9.1 | fast, recommended |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-Q4_1.gguf) | i1-Q4_1 | 9.5 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-Q5_K_S.gguf) | i1-Q5_K_S | 10.4 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-Q5_K_M.gguf) | i1-Q5_K_M | 10.6 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.i1-Q6_K.gguf) | i1-Q6_K | 12.2 | practically like static Q6_K |
Here is a handy graph by ikawrakow comparing some lower-quality quant
types (lower is better):

And here are Artefact2's thoughts on the matter:
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
## FAQ / Model Request
See https://huggingface.co/mradermacher/model_requests for some answers to
questions you might have and/or if you want some other model quantized.
## Thanks
I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
me use its servers and providing upgrades to my workstation to enable
this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.
<!-- end -->
|
mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF
|
mradermacher
| 2025-06-08T06:00:07Z | 50 | 0 |
transformers
|
[
"transformers",
"gguf",
"en",
"dataset:jondurbin/gutenberg-dpo-v0.1",
"dataset:nbeerbower/gutenberg2-dpo",
"dataset:nbeerbower/gutenberg-moderne-dpo",
"dataset:nbeerbower/synthetic-fiction-dpo",
"dataset:nbeerbower/Arkhaios-DPO",
"dataset:nbeerbower/Purpura-DPO",
"dataset:nbeerbower/Schule-DPO",
"base_model:nbeerbower/Qwen3-Gutenberg-Encore-14B",
"base_model:quantized:nbeerbower/Qwen3-Gutenberg-Encore-14B",
"license:apache-2.0",
"endpoints_compatible",
"region:us",
"conversational"
] | null | 2025-06-06T14:55:42Z |
---
base_model: nbeerbower/Qwen3-Gutenberg-Encore-14B
datasets:
- jondurbin/gutenberg-dpo-v0.1
- nbeerbower/gutenberg2-dpo
- nbeerbower/gutenberg-moderne-dpo
- nbeerbower/synthetic-fiction-dpo
- nbeerbower/Arkhaios-DPO
- nbeerbower/Purpura-DPO
- nbeerbower/Schule-DPO
language:
- en
library_name: transformers
license: apache-2.0
quantized_by: mradermacher
---
## About
<!-- ### quantize_version: 2 -->
<!-- ### output_tensor_quantised: 1 -->
<!-- ### convert_type: hf -->
<!-- ### vocab_type: -->
<!-- ### tags: -->
static quants of https://huggingface.co/nbeerbower/Qwen3-Gutenberg-Encore-14B
<!-- provided-files -->
weighted/imatrix quants are available at https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-i1-GGUF
## Usage
If you are unsure how to use GGUF files, refer to one of [TheBloke's
READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
more details, including on how to concatenate multi-part files.
## Provided Quants
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
| Link | Type | Size/GB | Notes |
|:-----|:-----|--------:|:------|
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.Q2_K.gguf) | Q2_K | 5.9 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.Q3_K_S.gguf) | Q3_K_S | 6.8 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.Q3_K_M.gguf) | Q3_K_M | 7.4 | lower quality |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.Q3_K_L.gguf) | Q3_K_L | 8.0 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.IQ4_XS.gguf) | IQ4_XS | 8.3 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.Q4_K_S.gguf) | Q4_K_S | 8.7 | fast, recommended |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.Q4_K_M.gguf) | Q4_K_M | 9.1 | fast, recommended |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.Q5_K_S.gguf) | Q5_K_S | 10.4 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.Q5_K_M.gguf) | Q5_K_M | 10.6 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.Q6_K.gguf) | Q6_K | 12.2 | very good quality |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-Gutenberg-Encore-14B-GGUF/resolve/main/Qwen3-Gutenberg-Encore-14B.Q8_0.gguf) | Q8_0 | 15.8 | fast, best quality |
Here is a handy graph by ikawrakow comparing some lower-quality quant
types (lower is better):

And here are Artefact2's thoughts on the matter:
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
## FAQ / Model Request
See https://huggingface.co/mradermacher/model_requests for some answers to
questions you might have and/or if you want some other model quantized.
## Thanks
I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
me use its servers and providing upgrades to my workstation to enable
this work in my free time.
<!-- end -->
|
anhnct/sana_1.5_v49_flux_LoRA_15k_img_epoch3_20000
|
anhnct
| 2025-06-08T05:59:58Z | 0 | 0 |
diffusers
|
[
"diffusers",
"safetensors",
"arxiv:1910.09700",
"diffusers:SanaPipeline",
"region:us"
] |
text-to-image
| 2025-06-08T03:44:02Z |
---
library_name: diffusers
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🧨 diffusers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
Zahro22/oxford-pet-segmentation
|
Zahro22
| 2025-06-08T05:45:41Z | 18 | 0 |
segmentation-models-pytorch
|
[
"segmentation-models-pytorch",
"safetensors",
"model_hub_mixin",
"pytorch_model_hub_mixin",
"semantic-segmentation",
"pytorch",
"image-segmentation",
"license:mit",
"region:us"
] |
image-segmentation
| 2025-05-15T08:51:22Z |
---
library_name: segmentation-models-pytorch
license: mit
pipeline_tag: image-segmentation
tags:
- model_hub_mixin
- pytorch_model_hub_mixin
- segmentation-models-pytorch
- semantic-segmentation
- pytorch
languages:
- python
---
# Unet Model Card
Table of Contents:
- [Load trained model](#load-trained-model)
- [Model init parameters](#model-init-parameters)
- [Model metrics](#model-metrics)
- [Dataset](#dataset)
## Load trained model
```python
import segmentation_models_pytorch as smp
model = smp.from_pretrained("<save-directory-or-this-repo>")
```
## Model init parameters
```python
model_init_params = {
"encoder_name": "resnet34",
"encoder_depth": 5,
"encoder_weights": "imagenet",
"decoder_use_norm": "batchnorm",
"decoder_channels": (256, 128, 64, 32, 16),
"decoder_attention_type": None,
"decoder_interpolation": "nearest",
"in_channels": 3,
"classes": 1,
"activation": None,
"aux_params": None
}
```
## Model metrics
```json
[
{
"test_per_image_iou": 0.9130621552467346,
"test_dataset_iou": 0.9190137386322021
}
]
```
## Dataset
Dataset name: Oxford Pet
## More Information
- Library: https://github.com/qubvel/segmentation_models.pytorch
- Docs: https://smp.readthedocs.io/en/latest/
This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin)
|
taguser/openshift-builds-operator-epoch3-2025-Jun-08
|
taguser
| 2025-06-08T05:34:38Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"llama-factory",
"lora",
"generated_from_trainer",
"base_model:Qwen/Qwen2.5-Coder-14B-Instruct",
"base_model:adapter:Qwen/Qwen2.5-Coder-14B-Instruct",
"license:other",
"region:us"
] | null | 2025-06-08T05:03:38Z |
---
library_name: peft
license: other
base_model: Qwen/Qwen2.5-Coder-14B-Instruct
tags:
- llama-factory
- lora
- generated_from_trainer
model-index:
- name: test
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# test
This model is a fine-tuned version of [Qwen/Qwen2.5-Coder-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-14B-Instruct) on the training_dataset dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 1
- eval_batch_size: 8
- seed: 42
- distributed_type: multi-GPU
- num_devices: 2
- gradient_accumulation_steps: 16
- total_train_batch_size: 32
- total_eval_batch_size: 16
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 3
### Training results
### Framework versions
- PEFT 0.15.1
- Transformers 4.51.0
- Pytorch 2.7.0+cu126
- Datasets 3.5.0
- Tokenizers 0.21.1
|
UICHEOL-HWANG/EcomGen-Gemma-3-0.0.1v
|
UICHEOL-HWANG
| 2025-06-08T05:33:32Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"gemma3",
"image-text-to-text",
"text-generation-inference",
"unsloth",
"conversational",
"en",
"base_model:unsloth/gemma-3-4b-it-unsloth-bnb-4bit",
"base_model:finetune:unsloth/gemma-3-4b-it-unsloth-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
image-text-to-text
| 2025-06-08T05:17:29Z |
---
base_model: unsloth/gemma-3-4b-it-unsloth-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- gemma3
license: apache-2.0
language:
- en
---
# Uploaded finetuned model
- **Developed by:** UICHEOL-HWANG
- **License:** apache-2.0
- **Finetuned from model :** unsloth/gemma-3-4b-it-unsloth-bnb-4bit
This gemma3 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
noza-kit/AACbase_byGemini_3_phase1-full
|
noza-kit
| 2025-06-08T05:33:08Z | 0 | 0 |
transformers
|
[
"transformers",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T05:33:07Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
a-k-aAiMGoD/phi3-mini-distributed-fine-tune
|
a-k-aAiMGoD
| 2025-06-08T05:32:43Z | 0 | 0 | null |
[
"safetensors",
"phi3",
"phi-3",
"fine-tuned",
"distributed-training",
"pytorch",
"custom_code",
"en",
"base_model:microsoft/Phi-3-mini-128k-instruct",
"base_model:finetune:microsoft/Phi-3-mini-128k-instruct",
"license:mit",
"region:us"
] | null | 2025-06-08T04:48:10Z |
---
license: mit
base_model: microsoft/Phi-3-mini-128k-instruct
tags:
- phi-3
- fine-tuned
- distributed-training
- pytorch
language:
- en
---
# Fine-tuned Phi-3-mini Model
This is a fine-tuned version of microsoft/Phi-3-mini-128k-instruct using distributed training.
## Model Details
- **Base Model**: microsoft/Phi-3-mini-128k-instruct
- **Training Method**: Distributed fine-tuning with Ray
- **Shards Used**: 2
- **Parameters**: ~3.8B
## Training Information
The model was fine-tuned using a distributed approach across multiple shards. While the base architecture is preserved, this model has been through a fine-tuning process optimized for specific tasks.
## Usage
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("a-k-aAiMGoD/phi3-mini-distributed-fine-tune")
model = AutoModelForCausalLM.from_pretrained("a-k-aAiMGoD/phi3-mini-distributed-fine-tune")
# Example usage
text = "Hello, how are you?"
inputs = tokenizer(text, return_tensors="pt")
outputs = model.generate(**inputs, max_length=100)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
```
## Training Configuration
- Distributed across 2 shards
- Optimized for large-scale deployment
- Enhanced with Ray-based parallelization
|
mlx-community/IndexTTS-1.5
|
mlx-community
| 2025-06-08T05:26:17Z | 29 | 0 |
mlx
|
[
"mlx",
"safetensors",
"indextts",
"text-to-speech",
"license:apache-2.0",
"region:us"
] |
text-to-speech
| 2025-06-04T10:40:19Z |
---
license: apache-2.0
tags:
- mlx
pipeline_tag: text-to-speech
---
# mlx-community/IndexTTS-1.5
This model was converted to MLX format from [`IndexTeam/IndexTTS-1.5`](https://huggingface.co/IndexTeam/IndexTTS-1.5) using mlx-audio version **0.2.3**.
Refer to the [original model card](https://huggingface.co/IndexTeam/IndexTTS-1.5) for more details on the model.
## Use with mlx
```bash
pip install -U mlx-audio
```
```bash
python -m mlx_audio.tts.generate --model mlx-community/IndexTTS-1.5 --text "Describe this image."
```
|
RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf
|
RichardErkhov
| 2025-06-08T05:22:46Z | 0 | 0 | null |
[
"gguf",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us",
"conversational"
] | null | 2025-06-08T04:16:36Z |
Quantization made by Richard Erkhov.
[Github](https://github.com/RichardErkhov)
[Discord](https://discord.gg/pvy7H8DZMG)
[Request more models](https://github.com/RichardErkhov/quant_request)
Llama-3.1-8B-ESG-Environmental - GGUF
- Model creator: https://huggingface.co/AarushSinha/
- Original model: https://huggingface.co/AarushSinha/Llama-3.1-8B-ESG-Environmental/
| Name | Quant method | Size |
| ---- | ---- | ---- |
| [Llama-3.1-8B-ESG-Environmental.Q2_K.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q2_K.gguf) | Q2_K | 2.96GB |
| [Llama-3.1-8B-ESG-Environmental.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.IQ3_XS.gguf) | IQ3_XS | 3.28GB |
| [Llama-3.1-8B-ESG-Environmental.IQ3_S.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.IQ3_S.gguf) | IQ3_S | 3.43GB |
| [Llama-3.1-8B-ESG-Environmental.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q3_K_S.gguf) | Q3_K_S | 3.41GB |
| [Llama-3.1-8B-ESG-Environmental.IQ3_M.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.IQ3_M.gguf) | IQ3_M | 3.52GB |
| [Llama-3.1-8B-ESG-Environmental.Q3_K.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q3_K.gguf) | Q3_K | 3.74GB |
| [Llama-3.1-8B-ESG-Environmental.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q3_K_M.gguf) | Q3_K_M | 3.74GB |
| [Llama-3.1-8B-ESG-Environmental.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q3_K_L.gguf) | Q3_K_L | 4.03GB |
| [Llama-3.1-8B-ESG-Environmental.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.IQ4_XS.gguf) | IQ4_XS | 4.18GB |
| [Llama-3.1-8B-ESG-Environmental.Q4_0.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q4_0.gguf) | Q4_0 | 4.34GB |
| [Llama-3.1-8B-ESG-Environmental.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.IQ4_NL.gguf) | IQ4_NL | 4.38GB |
| [Llama-3.1-8B-ESG-Environmental.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q4_K_S.gguf) | Q4_K_S | 4.37GB |
| [Llama-3.1-8B-ESG-Environmental.Q4_K.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q4_K.gguf) | Q4_K | 4.58GB |
| [Llama-3.1-8B-ESG-Environmental.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q4_K_M.gguf) | Q4_K_M | 4.58GB |
| [Llama-3.1-8B-ESG-Environmental.Q4_1.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q4_1.gguf) | Q4_1 | 4.78GB |
| [Llama-3.1-8B-ESG-Environmental.Q5_0.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q5_0.gguf) | Q5_0 | 5.21GB |
| [Llama-3.1-8B-ESG-Environmental.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q5_K_S.gguf) | Q5_K_S | 5.21GB |
| [Llama-3.1-8B-ESG-Environmental.Q5_K.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q5_K.gguf) | Q5_K | 5.34GB |
| [Llama-3.1-8B-ESG-Environmental.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q5_K_M.gguf) | Q5_K_M | 5.34GB |
| [Llama-3.1-8B-ESG-Environmental.Q5_1.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q5_1.gguf) | Q5_1 | 5.65GB |
| [Llama-3.1-8B-ESG-Environmental.Q6_K.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q6_K.gguf) | Q6_K | 6.14GB |
| [Llama-3.1-8B-ESG-Environmental.Q8_0.gguf](https://huggingface.co/RichardErkhov/AarushSinha_-_Llama-3.1-8B-ESG-Environmental-gguf/blob/main/Llama-3.1-8B-ESG-Environmental.Q8_0.gguf) | Q8_0 | 7.95GB |
Original model description:
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
johngreendr1/83be439c-d369-4dbb-9ed6-1645b4e499fb
|
johngreendr1
| 2025-06-08T05:22:31Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Sao10K/Llama-3.3-70B-Vulpecula-r1",
"base_model:adapter:Sao10K/Llama-3.3-70B-Vulpecula-r1",
"region:us"
] | null | 2025-06-08T05:22:14Z |
---
base_model: Sao10K/Llama-3.3-70B-Vulpecula-r1
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.15.1
|
johngreendr1/186911b6-715a-4a90-9b18-5ad9208120f5
|
johngreendr1
| 2025-06-08T05:21:59Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:unsloth/Hermes-3-Llama-3.1-8B",
"base_model:adapter:unsloth/Hermes-3-Llama-3.1-8B",
"region:us"
] | null | 2025-06-08T01:57:12Z |
---
base_model: unsloth/Hermes-3-Llama-3.1-8B
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.15.1
|
Vimax97/sdxl_bg_test
|
Vimax97
| 2025-06-08T05:16:16Z | 0 | 0 |
diffusers
|
[
"diffusers",
"stable-diffusion-xl",
"stable-diffusion-xl-diffusers",
"text-to-image",
"diffusers-training",
"lora",
"base_model:SG161222/RealVisXL_V4.0",
"base_model:adapter:SG161222/RealVisXL_V4.0",
"license:creativeml-openrail-m",
"region:us"
] |
text-to-image
| 2025-04-27T07:30:28Z |
---
base_model: SG161222/RealVisXL_V4.0
library_name: diffusers
license: creativeml-openrail-m
inference: true
tags:
- stable-diffusion-xl
- stable-diffusion-xl-diffusers
- text-to-image
- diffusers
- diffusers-training
- lora
---
<!-- This model card has been generated automatically according to the information the training script had access to. You
should probably proofread and complete it, then remove this comment. -->
# LoRA text2image fine-tuning - Vimax97/sdxl_bg_test
These are LoRA adaption weights for SG161222/RealVisXL_V4.0. The weights were fine-tuned on the Vimax97/background_florence_2_captioned_3050 dataset. You can find some example images in the following.




LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Intended uses & limitations
#### How to use
```python
# TODO: add an example code snippet for running this diffusion pipeline
```
#### Limitations and bias
[TODO: provide examples of latent issues and potential remediations]
## Training details
[TODO: describe the data used to train the model]
|
BootesVoid/cmbd6we16031o10oz99lbe56x_cmbdad6zd0003m73iviohbluv
|
BootesVoid
| 2025-06-08T05:15:15Z | 0 | 0 |
diffusers
|
[
"diffusers",
"flux",
"lora",
"replicate",
"text-to-image",
"en",
"base_model:black-forest-labs/FLUX.1-dev",
"base_model:adapter:black-forest-labs/FLUX.1-dev",
"license:other",
"region:us"
] |
text-to-image
| 2025-06-08T05:15:13Z |
---
license: other
license_name: flux-1-dev-non-commercial-license
license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md
language:
- en
tags:
- flux
- diffusers
- lora
- replicate
base_model: "black-forest-labs/FLUX.1-dev"
pipeline_tag: text-to-image
# widget:
# - text: >-
# prompt
# output:
# url: https://...
instance_prompt: AR69
---
# Cmbd6We16031O10Oz99Lbe56X_Cmbdad6Zd0003M73Iviohbluv
<Gallery />
## About this LoRA
This is a [LoRA](https://replicate.com/docs/guides/working-with-loras) for the FLUX.1-dev text-to-image model. It can be used with diffusers or ComfyUI.
It was trained on [Replicate](https://replicate.com/) using AI toolkit: https://replicate.com/ostris/flux-dev-lora-trainer/train
## Trigger words
You should use `AR69` to trigger the image generation.
## Run this LoRA with an API using Replicate
```py
import replicate
input = {
"prompt": "AR69",
"lora_weights": "https://huggingface.co/BootesVoid/cmbd6we16031o10oz99lbe56x_cmbdad6zd0003m73iviohbluv/resolve/main/lora.safetensors"
}
output = replicate.run(
"black-forest-labs/flux-dev-lora",
input=input
)
for index, item in enumerate(output):
with open(f"output_{index}.webp", "wb") as file:
file.write(item.read())
```
## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
```py
from diffusers import AutoPipelineForText2Image
import torch
pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda')
pipeline.load_lora_weights('BootesVoid/cmbd6we16031o10oz99lbe56x_cmbdad6zd0003m73iviohbluv', weight_name='lora.safetensors')
image = pipeline('AR69').images[0]
```
For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
## Training details
- Steps: 2000
- Learning rate: 0.0004
- LoRA rank: 16
## Contribute your own examples
You can use the [community tab](https://huggingface.co/BootesVoid/cmbd6we16031o10oz99lbe56x_cmbdad6zd0003m73iviohbluv/discussions) to add images that show off what you’ve made with this LoRA.
|
dgambettaphd/M_llm2_run0_gen9_WXS_doc1000_synt64_lr1e-04_acm_MPP
|
dgambettaphd
| 2025-06-08T05:12:35Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"unsloth",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T05:12:22Z |
---
library_name: transformers
tags:
- unsloth
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
GStoynev/lab-2
|
GStoynev
| 2025-06-08T05:12:10Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T05:04:56Z |
---
library_name: transformers
tags:
- generated_from_trainer
model-index:
- name: lab-2
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# lab-2
This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 1
### Training results
### Framework versions
- Transformers 4.52.4
- Pytorch 2.5.1+cu121
- Datasets 3.6.0
- Tokenizers 0.21.1
|
TheGardener/KD-llama-0.8b-shortened-epoch-1st-ver2
|
TheGardener
| 2025-06-08T05:11:41Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T05:10:59Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
VortexHunter23/LeoPARD-Coder-0.8.1-4bit
|
VortexHunter23
| 2025-06-08T05:09:46Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"conversational",
"en",
"base_model:VortexHunter23/LeoPARD-Coder-0.8",
"base_model:quantized:VortexHunter23/LeoPARD-Coder-0.8",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] |
text-generation
| 2025-06-08T05:08:03Z |
---
base_model: VortexHunter23/LeoPARD-Coder-0.8
tags:
- text-generation-inference
- transformers
- unsloth
- qwen2
- trl
- sft
license: apache-2.0
language:
- en
---
# Uploaded model
- **Developed by:** VortexHunter23
- **License:** apache-2.0
- **Finetuned from model :** VortexHunter23/LeoPARD-Coder-0.8
This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
DevQuasar/openbmb.MiniCPM4-0.5B-GGUF
|
DevQuasar
| 2025-06-08T05:08:28Z | 0 | 0 | null |
[
"gguf",
"text-generation",
"base_model:openbmb/MiniCPM4-0.5B",
"base_model:quantized:openbmb/MiniCPM4-0.5B",
"endpoints_compatible",
"region:us",
"conversational"
] |
text-generation
| 2025-06-08T05:05:04Z |
---
base_model:
- openbmb/MiniCPM4-0.5B
pipeline_tag: text-generation
---
[<img src="https://raw.githubusercontent.com/csabakecskemeti/devquasar/main/dq_logo_black-transparent.png" width="200"/>](https://devquasar.com)
Quantized version of: [openbmb/MiniCPM4-0.5B](https://huggingface.co/openbmb/MiniCPM4-0.5B)
'Make knowledge free for everyone'
<p align="center">
Made with <br>
<a href="https://www.civo.com/" target="_blank">
<img src="https://www.civo.com/assets/public/brand-assets/civo-logo-colour-60cc1622dedf346f7afde1fff760523f731b0aac106a5465af98ff4073114b74.svg" width="100"/>
</a>
</p>
<a href='https://ko-fi.com/L4L416YX7C' target='_blank'><img height='36' style='border:0px;height:36px;' src='https://storage.ko-fi.com/cdn/kofi6.png?v=6' border='0' alt='Buy Me a Coffee at ko-fi.com' /></a>
|
DevQuasar/NousResearch.Genstruct-7B-GGUF
|
DevQuasar
| 2025-06-08T05:04:59Z | 0 | 0 | null |
[
"gguf",
"text-generation",
"base_model:NousResearch/Genstruct-7B",
"base_model:quantized:NousResearch/Genstruct-7B",
"endpoints_compatible",
"region:us",
"conversational"
] |
text-generation
| 2025-06-08T04:11:41Z |
---
base_model:
- NousResearch/Genstruct-7B
pipeline_tag: text-generation
---
[<img src="https://raw.githubusercontent.com/csabakecskemeti/devquasar/main/dq_logo_black-transparent.png" width="200"/>](https://devquasar.com)
Quantized version of: [NousResearch/Genstruct-7B](https://huggingface.co/NousResearch/Genstruct-7B)
'Make knowledge free for everyone'
<p align="center">
Made with <br>
<a href="https://www.civo.com/" target="_blank">
<img src="https://www.civo.com/assets/public/brand-assets/civo-logo-colour-60cc1622dedf346f7afde1fff760523f731b0aac106a5465af98ff4073114b74.svg" width="100"/>
</a>
</p>
<a href='https://ko-fi.com/L4L416YX7C' target='_blank'><img height='36' style='border:0px;height:36px;' src='https://storage.ko-fi.com/cdn/kofi6.png?v=6' border='0' alt='Buy Me a Coffee at ko-fi.com' /></a>
|
sudoping01/bambara-tts-1-merged-16bit
|
sudoping01
| 2025-06-08T04:58:44Z | 13 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"text-generation-inference",
"conversational",
"en",
"base_model:sudoping01/bambara-tts-1-merged-16bit",
"base_model:finetune:sudoping01/bambara-tts-1-merged-16bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-05T16:12:29Z |
---
base_model: sudoping01/bambara-tts-1-merged-16bit
tags:
- text-generation-inference
- transformers
# - unsloth
- qwen2
license: apache-2.0
language:
- en
---
# Uploaded finetuned model
- **Developed by:** sudoping01
- **License:** apache-2.0
- **Finetuned from model :** sudoping01/bambara-tts-1-merged-16bit
<!--
This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth) -->
|
luckeciano/Qwen-2.5-7B-GRPO-Base-NoAdvNorm_3728
|
luckeciano
| 2025-06-08T04:56:24Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"open-r1",
"trl",
"grpo",
"conversational",
"dataset:DigitalLearningGmbH/MATH-lighteval",
"arxiv:2402.03300",
"base_model:Qwen/Qwen2.5-Math-7B",
"base_model:finetune:Qwen/Qwen2.5-Math-7B",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T00:13:18Z |
---
base_model: Qwen/Qwen2.5-Math-7B
datasets: DigitalLearningGmbH/MATH-lighteval
library_name: transformers
model_name: Qwen-2.5-7B-GRPO-Base-NoAdvNorm_3728
tags:
- generated_from_trainer
- open-r1
- trl
- grpo
licence: license
---
# Model Card for Qwen-2.5-7B-GRPO-Base-NoAdvNorm_3728
This model is a fine-tuned version of [Qwen/Qwen2.5-Math-7B](https://huggingface.co/Qwen/Qwen2.5-Math-7B) on the [DigitalLearningGmbH/MATH-lighteval](https://huggingface.co/datasets/DigitalLearningGmbH/MATH-lighteval) dataset.
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="luckeciano/Qwen-2.5-7B-GRPO-Base-NoAdvNorm_3728", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/max-ent-llms/PolicyGradientStability/runs/fjgphf8u)
This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
### Framework versions
- TRL: 0.16.0.dev0
- Transformers: 4.49.0
- Pytorch: 2.6.0
- Datasets: 3.4.1
- Tokenizers: 0.21.1
## Citations
Cite GRPO as:
```bibtex
@article{zhihong2024deepseekmath,
title = {{DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models}},
author = {Zhihong Shao and Peiyi Wang and Qihao Zhu and Runxin Xu and Junxiao Song and Mingchuan Zhang and Y. K. Li and Y. Wu and Daya Guo},
year = 2024,
eprint = {arXiv:2402.03300},
}
```
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
ahmetikbal/gemma-2b-it-bnb-4bit-CSE4078_Grp1-r16-tr-ner-lr2e-4
|
ahmetikbal
| 2025-06-08T04:55:45Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"gemma",
"trl",
"en",
"base_model:unsloth/gemma-2b-it-bnb-4bit",
"base_model:finetune:unsloth/gemma-2b-it-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T04:55:40Z |
---
base_model: unsloth/gemma-2b-it-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- gemma
- trl
license: apache-2.0
language:
- en
---
# Uploaded model
- **Developed by:** ahmetikbal
- **License:** apache-2.0
- **Finetuned from model :** unsloth/gemma-2b-it-bnb-4bit
This gemma model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
glif-loradex-trainer/Swap_agrawal14_kuki_youtube_transition_v1
|
glif-loradex-trainer
| 2025-06-08T04:52:27Z | 0 | 0 |
diffusers
|
[
"diffusers",
"text-to-image",
"template:sd-lora",
"base_model:black-forest-labs/FLUX.1-dev",
"base_model:finetune:black-forest-labs/FLUX.1-dev",
"license:other",
"region:us",
"flux",
"lora",
"base_model:adapter:black-forest-labs/FLUX.1-dev"
] |
text-to-image
| 2025-06-08T04:52:17Z |
---
tags:
- diffusers
- text-to-image
- template:sd-lora
- base_model:black-forest-labs/FLUX.1-dev
- base_model:finetune:black-forest-labs/FLUX.1-dev
- license:other
- region:us
- flux
- lora
widget:
- output:
url: samples/1749358248756__000001000_0.jpg
text: Transition from elephant theme outfit to YouTube outfit $wap_yt_transi_v1
- output:
url: samples/1749358273871__000001000_1.jpg
text: Transition from casual boring outfit to YouTube outfit $wap_yt_transi_v1
- output:
url: samples/1749358298969__000001000_2.jpg
text: Transition from Mickey mouse outfit to YouTube outfit $wap_yt_transi_v1
- output:
url: samples/1749358324164__000001000_3.jpg
text: Transition from India peacock theme saree outfit to YouTube outfit $wap_yt_transi_v1
base_model: black-forest-labs/FLUX.1-dev
trigger: "$wap_yt_transi_v1"
instance_prompt: "$wap_yt_transi_v1"
license: other
license_name: flux-1-dev-non-commercial-license
license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md
---
# kuki_youtube_transition_v1
Model trained with [AI Toolkit by Ostris](https://github.com/ostris/ai-toolkit) under the [Glif Loradex program](https://huggingface.co/glif-loradex-trainer) by [Glif](https://glif.app) user `Swap_agrawal14`.
<Gallery />
## Trigger words
You should use `$wap_yt_transi_v1` to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](/glif-loradex-trainer/Swap_agrawal14_kuki_youtube_transition_v1/tree/main) them in the Files & versions tab.
## License
This model is licensed under the [flux-1-dev-non-commercial-license](https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md).
|
nvidia/cascade_mask_rcnn_mamba_vision_small_3x_coco
|
nvidia
| 2025-06-08T04:45:57Z | 0 | 1 | null |
[
"license:other",
"region:us"
] | null | 2025-06-08T04:37:50Z |
---
license: other
license_name: nvclv1
license_link: LICENSE
---
|
Dukuru/lora_model
|
Dukuru
| 2025-06-08T04:43:42Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"llama",
"trl",
"en",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T04:43:33Z |
---
base_model: unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- trl
license: apache-2.0
language:
- en
---
# Uploaded model
- **Developed by:** Dukuru
- **License:** apache-2.0
- **Finetuned from model :** unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
gradientrouting-spar/exp_to_matrix_exp_task_epoch_10
|
gradientrouting-spar
| 2025-06-08T04:37:03Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T04:36:50Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
one1cat/llama3.2-1b-cfrTrained
|
one1cat
| 2025-06-08T04:34:39Z | 0 | 0 | null |
[
"safetensors",
"llama",
"custom_code",
"base_model:meta-llama/Llama-3.2-1B",
"base_model:finetune:meta-llama/Llama-3.2-1B",
"license:mit",
"region:us"
] | null | 2025-06-08T03:50:32Z |
---
license: mit
base_model:
- meta-llama/Llama-3.2-1B
---
|
gradientrouting-spar/2d_data_color_seed_1_20250608_042738
|
gradientrouting-spar
| 2025-06-08T04:33:02Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"gemma2",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T04:30:37Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf
|
RichardErkhov
| 2025-06-08T04:26:26Z | 0 | 0 | null |
[
"gguf",
"endpoints_compatible",
"region:us",
"conversational"
] | null | 2025-06-08T03:10:30Z |
Quantization made by Richard Erkhov.
[Github](https://github.com/RichardErkhov)
[Discord](https://discord.gg/pvy7H8DZMG)
[Request more models](https://github.com/RichardErkhov/quant_request)
Multilingual-SaigaSuzume-8B - GGUF
- Model creator: https://huggingface.co/Khetterman/
- Original model: https://huggingface.co/Khetterman/Multilingual-SaigaSuzume-8B/
| Name | Quant method | Size |
| ---- | ---- | ---- |
| [Multilingual-SaigaSuzume-8B.Q2_K.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q2_K.gguf) | Q2_K | 2.96GB |
| [Multilingual-SaigaSuzume-8B.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.IQ3_XS.gguf) | IQ3_XS | 3.28GB |
| [Multilingual-SaigaSuzume-8B.IQ3_S.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.IQ3_S.gguf) | IQ3_S | 3.43GB |
| [Multilingual-SaigaSuzume-8B.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q3_K_S.gguf) | Q3_K_S | 3.41GB |
| [Multilingual-SaigaSuzume-8B.IQ3_M.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.IQ3_M.gguf) | IQ3_M | 3.52GB |
| [Multilingual-SaigaSuzume-8B.Q3_K.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q3_K.gguf) | Q3_K | 3.74GB |
| [Multilingual-SaigaSuzume-8B.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q3_K_M.gguf) | Q3_K_M | 3.74GB |
| [Multilingual-SaigaSuzume-8B.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q3_K_L.gguf) | Q3_K_L | 4.03GB |
| [Multilingual-SaigaSuzume-8B.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.IQ4_XS.gguf) | IQ4_XS | 4.18GB |
| [Multilingual-SaigaSuzume-8B.Q4_0.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q4_0.gguf) | Q4_0 | 4.34GB |
| [Multilingual-SaigaSuzume-8B.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.IQ4_NL.gguf) | IQ4_NL | 4.38GB |
| [Multilingual-SaigaSuzume-8B.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q4_K_S.gguf) | Q4_K_S | 4.37GB |
| [Multilingual-SaigaSuzume-8B.Q4_K.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q4_K.gguf) | Q4_K | 4.58GB |
| [Multilingual-SaigaSuzume-8B.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q4_K_M.gguf) | Q4_K_M | 4.58GB |
| [Multilingual-SaigaSuzume-8B.Q4_1.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q4_1.gguf) | Q4_1 | 4.78GB |
| [Multilingual-SaigaSuzume-8B.Q5_0.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q5_0.gguf) | Q5_0 | 5.21GB |
| [Multilingual-SaigaSuzume-8B.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q5_K_S.gguf) | Q5_K_S | 5.21GB |
| [Multilingual-SaigaSuzume-8B.Q5_K.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q5_K.gguf) | Q5_K | 5.34GB |
| [Multilingual-SaigaSuzume-8B.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q5_K_M.gguf) | Q5_K_M | 5.34GB |
| [Multilingual-SaigaSuzume-8B.Q5_1.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q5_1.gguf) | Q5_1 | 5.65GB |
| [Multilingual-SaigaSuzume-8B.Q6_K.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q6_K.gguf) | Q6_K | 6.14GB |
| [Multilingual-SaigaSuzume-8B.Q8_0.gguf](https://huggingface.co/RichardErkhov/Khetterman_-_Multilingual-SaigaSuzume-8B-gguf/blob/main/Multilingual-SaigaSuzume-8B.Q8_0.gguf) | Q8_0 | 7.95GB |
Original model description:
---
base_model:
- huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated
- IlyaGusev/saiga_llama3_8b
- lightblue/suzume-llama-3-8B-multilingual
- lightblue/suzume-llama-3-8B-multilingual-orpo-borda-full
- lightblue/suzume-llama-3-8B-multilingual-orpo-borda-half
- lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top25
- lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top75
library_name: transformers
tags:
- mergekit
- merge
- bfloat16
- safetensors
- 8b
- chat
- conversational
language:
- de
- en
- es
- fr
- hi
- it
- ja
- pt
- ru
- th
- zh
---
# Multilingual-SaigaSuzume-8B
>Your words are like rain falling from heaven on a tower in a sinful land; can anyone in Babylon understand them?

This model was created as the basis of multilingual abilities for other models. I think it will be very useful as an integral part of your model. There is some censorship, keep this in mind.
## Merge Details
### Method
This is a simple, but usefull merge of **7 cool models**, created using [mergekit](https://github.com/arcee-ai/mergekit).
### Models
The following models were included in the merge:
* [huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated](https://huggingface.co/huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated)
* [IlyaGusev/saiga_llama3_8b](https://huggingface.co/IlyaGusev/saiga_llama3_8b)
* [lightblue/suzume-llama-3-8B-multilingual](https://huggingface.co/lightblue/suzume-llama-3-8B-multilingual)
* [lightblue/suzume-llama-3-8B-multilingual-orpo-borda-full](https://huggingface.co/lightblue/suzume-llama-3-8B-multilingual-orpo-borda-full)
* [lightblue/suzume-llama-3-8B-multilingual-orpo-borda-half](https://huggingface.co/lightblue/suzume-llama-3-8B-multilingual-orpo-borda-half)
* [lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top25](https://huggingface.co/lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top25)
* [lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top75](https://huggingface.co/lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top75)
### Configuration
The following YAML configurations was used to produce this model:
```yaml
# Multilingual-SaigaSuzume-8B-BFH
models:
- model: lightblue/suzume-llama-3-8B-multilingual-orpo-borda-full
- model: IlyaGusev/saiga_llama3_8b
- model: lightblue/suzume-llama-3-8B-multilingual-orpo-borda-half
merge_method: model_stock
base_model: huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated
dtype: bfloat16
# Multilingual-SaigaSuzume-8B-BTP
models:
- model: lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top75
- model: IlyaGusev/saiga_llama3_8b
- model: lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top25
merge_method: model_stock
base_model: huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated
dtype: bfloat16
# Multilingual-SaigaSuzume-8B-Classic
models:
- model: IlyaGusev/saiga_llama3_8b
- model: lightblue/suzume-llama-3-8B-multilingual
merge_method: model_stock
base_model: huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated
dtype: bfloat16
# Multilingual-SaigaSuzume-8B
models:
- model: Multilingual-SaigaSuzume-8B-BFH
- model: Multilingual-SaigaSuzume-8B-BTP
merge_method: model_stock
base_model: Multilingual-SaigaSuzume-8B-Classic
dtype: bfloat16
```
>My thanks to the authors of the original models, your work is incredible. Have a good time 🖤
|
coralieb7/mcqa_sft_focus_100k_2048length
|
coralieb7
| 2025-06-08T04:24:07Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen3",
"text-generation",
"generated_from_trainer",
"trl",
"sft",
"conversational",
"base_model:Qwen/Qwen3-0.6B-Base",
"base_model:finetune:Qwen/Qwen3-0.6B-Base",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T04:22:44Z |
---
base_model: Qwen/Qwen3-0.6B-Base
library_name: transformers
model_name: mcqa_sft_focus_100k_2048length
tags:
- generated_from_trainer
- trl
- sft
licence: license
---
# Model Card for mcqa_sft_focus_100k_2048length
This model is a fine-tuned version of [Qwen/Qwen3-0.6B-Base](https://huggingface.co/Qwen/Qwen3-0.6B-Base).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="coralieb7/mcqa_sft_focus_100k_2048length", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
This model was trained with SFT.
### Framework versions
- TRL: 0.17.0
- Transformers: 4.51.3
- Pytorch: 2.7.0
- Datasets: 3.2.0
- Tokenizers: 0.21.0
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
Flo0620/Qwen2_5_7B_r32_a32_d0_1_ArXivQA
|
Flo0620
| 2025-06-08T04:22:26Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"generated_from_trainer",
"trl",
"sft",
"base_model:Qwen/Qwen2.5-VL-7B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-VL-7B-Instruct",
"endpoints_compatible",
"region:us"
] | null | 2025-05-30T04:27:26Z |
---
base_model: Qwen/Qwen2.5-VL-7B-Instruct
library_name: transformers
model_name: Qwen2_5_7B_r32_a32_d0_1_ArXivQA
tags:
- generated_from_trainer
- trl
- sft
licence: license
---
# Model Card for Qwen2_5_7B_r32_a32_d0_1_ArXivQA
This model is a fine-tuned version of [Qwen/Qwen2.5-VL-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="Flo0620/Qwen2_5_7B_r32_a32_d0_1_ArXivQA", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
This model was trained with SFT.
### Framework versions
- TRL: 0.15.2
- Transformers: 4.52.0.dev0
- Pytorch: 2.6.0+cu124
- Datasets: 3.5.0
- Tokenizers: 0.21.1
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
Lewdiculous/DS-R1-Qwen3-8B-ArliAI-RpR-v4-Small-GGUF-IQ-Imatrix
|
Lewdiculous
| 2025-06-08T04:21:42Z | 0 | 1 | null |
[
"gguf",
"qwen3",
"qwen",
"chatml",
"sillytavern",
"roleplay",
"conversational",
"reasoning",
"thinking",
"en",
"base_model:ArliAI/DS-R1-Qwen3-8B-ArliAI-RpR-v4-Small",
"base_model:quantized:ArliAI/DS-R1-Qwen3-8B-ArliAI-RpR-v4-Small",
"license:apache-2.0",
"endpoints_compatible",
"region:us",
"imatrix"
] | null | 2025-06-08T02:20:08Z |
---
language:
- en
base_model:
- ArliAI/DS-R1-Qwen3-8B-ArliAI-RpR-v4-Small
tags:
- qwen3
- qwen
- chatml
- sillytavern
- roleplay
- conversational
- reasoning
- thinking
license: apache-2.0
---
<!--
- presets
- mistral
-->
<!--
> [!WARNING]
> **Uploading...** <br>
> Card will be updated later.
-->
My **GGUF-Imatrix** quants of [**DS-R1-Qwen3-8B-ArliAI-RpR-v4-Small**](https://huggingface.co/ArliAI/DS-R1-Qwen3-8B-ArliAI-RpR-v4-Small). <br>

> [!NOTE]
> **Prompt format:** <br>
> ChatML
>
> **Note:** <br>
> Set the additional settings as per the instructions in the image at the end of the card to use the thinking setup. [[1]](https://files.catbox.moe/3jky2q.jpg) <br>
> 
# Reasoning setup in SillyTavern:

|
2-wolf-1-girl-viral-video/18.VIDEOS.2.wolf.1.girl.viral.video.download.link
|
2-wolf-1-girl-viral-video
| 2025-06-08T04:21:35Z | 0 | 0 | null |
[
"region:us"
] | null | 2025-06-08T04:20:56Z |
<p><a rel="nofollow" href="https://viralflix.xyz/leaked/?ffd">►►✅ 𝘾𝙇𝙄𝘾𝙆 𝙃𝙀𝙍𝙀 ==►► 𝙁𝙪𝙡𝙡 𝙑𝙞𝙙𝙚𝙤️​</a></p>
<a rel="nofollow" href="https://viralflix.xyz/leaked/?ffd">🔴 CLICK HERE 🌐==►► Download Now)</a>
<a rel="nofollow" href="https://viralflix.xyz/leaked/?ffd"><img src="https://i.postimg.cc/qvPp49Sm/ythngythg.gif" alt="fsd"></a>
|
Superrrdamn/task-10-microsoft-Phi-4-mini-instruct
|
Superrrdamn
| 2025-06-08T04:19:27Z | 51 | 0 |
peft
|
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:microsoft/Phi-4-mini-instruct",
"base_model:adapter:microsoft/Phi-4-mini-instruct",
"region:us"
] | null | 2025-06-07T00:06:39Z |
---
base_model: microsoft/Phi-4-mini-instruct
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.15.2
|
RichardErkhov/kh38_-_my-cool-model1122-gguf
|
RichardErkhov
| 2025-06-08T04:15:34Z | 0 | 0 | null |
[
"gguf",
"arxiv:2311.03099",
"arxiv:2306.01708",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T03:01:11Z |
Quantization made by Richard Erkhov.
[Github](https://github.com/RichardErkhov)
[Discord](https://discord.gg/pvy7H8DZMG)
[Request more models](https://github.com/RichardErkhov/quant_request)
my-cool-model1122 - GGUF
- Model creator: https://huggingface.co/kh38/
- Original model: https://huggingface.co/kh38/my-cool-model1122/
| Name | Quant method | Size |
| ---- | ---- | ---- |
| [my-cool-model1122.Q2_K.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q2_K.gguf) | Q2_K | 2.96GB |
| [my-cool-model1122.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.IQ3_XS.gguf) | IQ3_XS | 3.28GB |
| [my-cool-model1122.IQ3_S.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.IQ3_S.gguf) | IQ3_S | 3.43GB |
| [my-cool-model1122.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q3_K_S.gguf) | Q3_K_S | 3.41GB |
| [my-cool-model1122.IQ3_M.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.IQ3_M.gguf) | IQ3_M | 3.52GB |
| [my-cool-model1122.Q3_K.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q3_K.gguf) | Q3_K | 3.74GB |
| [my-cool-model1122.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q3_K_M.gguf) | Q3_K_M | 3.74GB |
| [my-cool-model1122.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q3_K_L.gguf) | Q3_K_L | 4.03GB |
| [my-cool-model1122.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.IQ4_XS.gguf) | IQ4_XS | 4.18GB |
| [my-cool-model1122.Q4_0.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q4_0.gguf) | Q4_0 | 4.34GB |
| [my-cool-model1122.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.IQ4_NL.gguf) | IQ4_NL | 4.38GB |
| [my-cool-model1122.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q4_K_S.gguf) | Q4_K_S | 4.37GB |
| [my-cool-model1122.Q4_K.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q4_K.gguf) | Q4_K | 4.58GB |
| [my-cool-model1122.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q4_K_M.gguf) | Q4_K_M | 4.58GB |
| [my-cool-model1122.Q4_1.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q4_1.gguf) | Q4_1 | 4.78GB |
| [my-cool-model1122.Q5_0.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q5_0.gguf) | Q5_0 | 5.21GB |
| [my-cool-model1122.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q5_K_S.gguf) | Q5_K_S | 5.21GB |
| [my-cool-model1122.Q5_K.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q5_K.gguf) | Q5_K | 5.34GB |
| [my-cool-model1122.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q5_K_M.gguf) | Q5_K_M | 5.34GB |
| [my-cool-model1122.Q5_1.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q5_1.gguf) | Q5_1 | 5.65GB |
| [my-cool-model1122.Q6_K.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q6_K.gguf) | Q6_K | 6.14GB |
| [my-cool-model1122.Q8_0.gguf](https://huggingface.co/RichardErkhov/kh38_-_my-cool-model1122-gguf/blob/main/my-cool-model1122.Q8_0.gguf) | Q8_0 | 7.95GB |
Original model description:
---
base_model: []
library_name: transformers
tags:
- mergekit
- merge
---
# final_merge
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using ../evol_merge_storage/input_models/Llama-3.1-Swallow-8B-v0.2_4249862252 as a base.
### Models Merged
The following models were included in the merge:
* ../evol_merge_storage/input_models/llama-3-chinese-8b_120379959
* ../evol_merge_storage/input_models/Llama-3-ELYZA-JP-8B_2371007997
### Configuration
The following YAML configuration was used to produce this model:
```yaml
base_model: ../evol_merge_storage/input_models/Llama-3.1-Swallow-8B-v0.2_4249862252
dtype: bfloat16
merge_method: dare_ties
parameters:
int8_mask: 1.0
normalize: 1.0
slices:
- sources:
- layer_range: [0, 4]
model: ../evol_merge_storage/input_models/Llama-3.1-Swallow-8B-v0.2_4249862252
parameters:
density: 0.863485562098192
weight: 0.22651847020495885
- layer_range: [0, 4]
model: ../evol_merge_storage/input_models/llama-3-chinese-8b_120379959
parameters:
density: 0.9343953420777168
weight: 0.5036150562646258
- layer_range: [0, 4]
model: ../evol_merge_storage/input_models/Llama-3-ELYZA-JP-8B_2371007997
parameters:
density: 1.0
weight: 0.6451005324417585
- sources:
- layer_range: [4, 8]
model: ../evol_merge_storage/input_models/Llama-3.1-Swallow-8B-v0.2_4249862252
parameters:
density: 0.9846266882538002
weight: 0.5639921695621852
- layer_range: [4, 8]
model: ../evol_merge_storage/input_models/llama-3-chinese-8b_120379959
parameters:
density: 1.0
weight: 0.3231299604274662
- layer_range: [4, 8]
model: ../evol_merge_storage/input_models/Llama-3-ELYZA-JP-8B_2371007997
parameters:
density: 0.9908955898534834
weight: 0.21486915206711796
- sources:
- layer_range: [8, 12]
model: ../evol_merge_storage/input_models/Llama-3.1-Swallow-8B-v0.2_4249862252
parameters:
density: 0.9065299264285266
weight: 0.2987555834921648
- layer_range: [8, 12]
model: ../evol_merge_storage/input_models/llama-3-chinese-8b_120379959
parameters:
density: 0.8840782503058148
weight: 0.26619854603379545
- layer_range: [8, 12]
model: ../evol_merge_storage/input_models/Llama-3-ELYZA-JP-8B_2371007997
parameters:
density: 0.9914153096559333
weight: 0.4573592950405189
- sources:
- layer_range: [12, 16]
model: ../evol_merge_storage/input_models/Llama-3.1-Swallow-8B-v0.2_4249862252
parameters:
density: 0.9740298213855892
weight: 0.48137164129667176
- layer_range: [12, 16]
model: ../evol_merge_storage/input_models/llama-3-chinese-8b_120379959
parameters:
density: 1.0
weight: 0.27412584703978277
- layer_range: [12, 16]
model: ../evol_merge_storage/input_models/Llama-3-ELYZA-JP-8B_2371007997
parameters:
density: 0.8407412390278275
weight: 0.3182141906839257
- sources:
- layer_range: [16, 20]
model: ../evol_merge_storage/input_models/Llama-3.1-Swallow-8B-v0.2_4249862252
parameters:
density: 1.0
weight: 0.2240504757935422
- layer_range: [16, 20]
model: ../evol_merge_storage/input_models/llama-3-chinese-8b_120379959
parameters:
density: 1.0
weight: 0.23938850503773312
- layer_range: [16, 20]
model: ../evol_merge_storage/input_models/Llama-3-ELYZA-JP-8B_2371007997
parameters:
density: 0.9687795057288319
weight: 0.5987730759861593
- sources:
- layer_range: [20, 24]
model: ../evol_merge_storage/input_models/Llama-3.1-Swallow-8B-v0.2_4249862252
parameters:
density: 1.0
weight: 0.09945022964618122
- layer_range: [20, 24]
model: ../evol_merge_storage/input_models/llama-3-chinese-8b_120379959
parameters:
density: 1.0
weight: 0.26835539762495914
- layer_range: [20, 24]
model: ../evol_merge_storage/input_models/Llama-3-ELYZA-JP-8B_2371007997
parameters:
density: 0.8139356897740962
weight: 0.4942452603808056
- sources:
- layer_range: [24, 28]
model: ../evol_merge_storage/input_models/Llama-3.1-Swallow-8B-v0.2_4249862252
parameters:
density: 1.0
weight: 0.20318580465269015
- layer_range: [24, 28]
model: ../evol_merge_storage/input_models/llama-3-chinese-8b_120379959
parameters:
density: 1.0
weight: 0.16861512537170825
- layer_range: [24, 28]
model: ../evol_merge_storage/input_models/Llama-3-ELYZA-JP-8B_2371007997
parameters:
density: 1.0
weight: 0.15118597877918583
- sources:
- layer_range: [28, 32]
model: ../evol_merge_storage/input_models/Llama-3.1-Swallow-8B-v0.2_4249862252
parameters:
density: 0.7988559962120717
weight: 0.34008425117612984
- layer_range: [28, 32]
model: ../evol_merge_storage/input_models/llama-3-chinese-8b_120379959
parameters:
density: 1.0
weight: 0.2824977970939407
- layer_range: [28, 32]
model: ../evol_merge_storage/input_models/Llama-3-ELYZA-JP-8B_2371007997
parameters:
density: 0.7131873401997189
weight: 0.5228166170045327
tokenizer_source: base
```
|
publication-charaf/MCQ_Qwen3-0.6B-Base_lr-1e-05_e-7_s-0
|
publication-charaf
| 2025-06-08T04:13:51Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen3",
"text-generation",
"generated_from_trainer",
"trl",
"sft",
"conversational",
"base_model:Qwen/Qwen3-0.6B-Base",
"base_model:finetune:Qwen/Qwen3-0.6B-Base",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T00:20:42Z |
---
base_model: Qwen/Qwen3-0.6B-Base
library_name: transformers
model_name: MCQ_Qwen3-0.6B-Base_lr-1e-05_e-7_s-0
tags:
- generated_from_trainer
- trl
- sft
licence: license
---
# Model Card for MCQ_Qwen3-0.6B-Base_lr-1e-05_e-7_s-0
This model is a fine-tuned version of [Qwen/Qwen3-0.6B-Base](https://huggingface.co/Qwen/Qwen3-0.6B-Base).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="publication-charaf/MCQ_Qwen3-0.6B-Base_lr-1e-05_e-7_s-0", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/kamel-charaf-epfl/huggingface/runs/zduuvosu)
This model was trained with SFT.
### Framework versions
- TRL: 0.17.0
- Transformers: 4.51.3
- Pytorch: 2.6.0
- Datasets: 3.5.0
- Tokenizers: 0.21.1
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
Procit004/Gemma-2b
|
Procit004
| 2025-06-08T04:12:33Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"gemma",
"text-generation",
"trl",
"sft",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] |
text-generation
| 2025-06-08T04:07:47Z |
---
library_name: transformers
tags:
- trl
- sft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
DevQuasar/Salesforce.E1-AceReason-14B-GGUF
|
DevQuasar
| 2025-06-08T04:11:35Z | 0 | 0 | null |
[
"gguf",
"text-generation",
"base_model:Salesforce/E1-AceReason-14B",
"base_model:quantized:Salesforce/E1-AceReason-14B",
"endpoints_compatible",
"region:us",
"conversational"
] |
text-generation
| 2025-06-08T01:35:54Z |
---
base_model:
- Salesforce/E1-AceReason-14B
pipeline_tag: text-generation
---
[<img src="https://raw.githubusercontent.com/csabakecskemeti/devquasar/main/dq_logo_black-transparent.png" width="200"/>](https://devquasar.com)
Quantized version of: [Salesforce/E1-AceReason-14B](https://huggingface.co/Salesforce/E1-AceReason-14B)
'Make knowledge free for everyone'
<p align="center">
Made with <br>
<a href="https://www.civo.com/" target="_blank">
<img src="https://www.civo.com/assets/public/brand-assets/civo-logo-colour-60cc1622dedf346f7afde1fff760523f731b0aac106a5465af98ff4073114b74.svg" width="100"/>
</a>
</p>
<a href='https://ko-fi.com/L4L416YX7C' target='_blank'><img height='36' style='border:0px;height:36px;' src='https://storage.ko-fi.com/cdn/kofi6.png?v=6' border='0' alt='Buy Me a Coffee at ko-fi.com' /></a>
|
RichardErkhov/ianr007_-_sdg-test-gguf
|
RichardErkhov
| 2025-06-08T04:08:16Z | 0 | 0 | null |
[
"gguf",
"endpoints_compatible",
"region:us",
"conversational"
] | null | 2025-06-08T03:03:38Z |
Quantization made by Richard Erkhov.
[Github](https://github.com/RichardErkhov)
[Discord](https://discord.gg/pvy7H8DZMG)
[Request more models](https://github.com/RichardErkhov/quant_request)
sdg-test - GGUF
- Model creator: https://huggingface.co/ianr007/
- Original model: https://huggingface.co/ianr007/sdg-test/
| Name | Quant method | Size |
| ---- | ---- | ---- |
| [sdg-test.Q2_K.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q2_K.gguf) | Q2_K | 2.96GB |
| [sdg-test.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.IQ3_XS.gguf) | IQ3_XS | 3.28GB |
| [sdg-test.IQ3_S.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.IQ3_S.gguf) | IQ3_S | 3.43GB |
| [sdg-test.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q3_K_S.gguf) | Q3_K_S | 3.41GB |
| [sdg-test.IQ3_M.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.IQ3_M.gguf) | IQ3_M | 3.52GB |
| [sdg-test.Q3_K.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q3_K.gguf) | Q3_K | 3.74GB |
| [sdg-test.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q3_K_M.gguf) | Q3_K_M | 3.74GB |
| [sdg-test.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q3_K_L.gguf) | Q3_K_L | 4.03GB |
| [sdg-test.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.IQ4_XS.gguf) | IQ4_XS | 4.18GB |
| [sdg-test.Q4_0.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q4_0.gguf) | Q4_0 | 4.34GB |
| [sdg-test.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.IQ4_NL.gguf) | IQ4_NL | 4.38GB |
| [sdg-test.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q4_K_S.gguf) | Q4_K_S | 4.37GB |
| [sdg-test.Q4_K.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q4_K.gguf) | Q4_K | 4.58GB |
| [sdg-test.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q4_K_M.gguf) | Q4_K_M | 4.58GB |
| [sdg-test.Q4_1.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q4_1.gguf) | Q4_1 | 4.78GB |
| [sdg-test.Q5_0.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q5_0.gguf) | Q5_0 | 5.21GB |
| [sdg-test.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q5_K_S.gguf) | Q5_K_S | 5.21GB |
| [sdg-test.Q5_K.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q5_K.gguf) | Q5_K | 5.34GB |
| [sdg-test.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q5_K_M.gguf) | Q5_K_M | 5.34GB |
| [sdg-test.Q5_1.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q5_1.gguf) | Q5_1 | 5.65GB |
| [sdg-test.Q6_K.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q6_K.gguf) | Q6_K | 6.14GB |
| [sdg-test.Q8_0.gguf](https://huggingface.co/RichardErkhov/ianr007_-_sdg-test-gguf/blob/main/sdg-test.Q8_0.gguf) | Q8_0 | 7.95GB |
Original model description:
---
tags:
- autotrain
- text-generation-inference
- text-generation
- peft
library_name: transformers
widget:
- messages:
- role: user
content: What is your favorite condiment?
license: other
---
# Model Trained Using AutoTrain
This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain).
# Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_path = "PATH_TO_THIS_REPO"
tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
model_path,
device_map="auto",
torch_dtype='auto'
).eval()
# Prompt content: "hi"
messages = [
{"role": "user", "content": "hi"}
]
input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt')
output_ids = model.generate(input_ids.to('cuda'))
response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True)
# Model response: "Hello! How can I assist you today?"
print(response)
```
|
Jondojds/Heytic
|
Jondojds
| 2025-06-08T03:54:13Z | 0 | 0 | null |
[
"license:bigscience-openrail-m",
"region:us"
] | null | 2025-06-08T03:54:13Z |
---
license: bigscience-openrail-m
---
|
KasuleTrevor/QWen-sample
|
KasuleTrevor
| 2025-06-08T03:52:46Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2_audio",
"text2text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] |
text2text-generation
| 2025-06-08T03:22:20Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
KasuleTrevor/Qwen-song-birds
|
KasuleTrevor
| 2025-06-08T03:51:51Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"generated_from_trainer",
"base_model:Qwen/Qwen2-Audio-7B-Instruct",
"base_model:adapter:Qwen/Qwen2-Audio-7B-Instruct",
"license:apache-2.0",
"region:us"
] | null | 2025-06-08T03:51:45Z |
---
library_name: peft
license: apache-2.0
base_model: Qwen/Qwen2-Audio-7B-Instruct
tags:
- generated_from_trainer
model-index:
- name: Qwen-song-birds
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Qwen-song-birds
This model is a fine-tuned version of [Qwen/Qwen2-Audio-7B-Instruct](https://huggingface.co/Qwen/Qwen2-Audio-7B-Instruct) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2721
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: constant
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 1.7534 | 0.4 | 8 | 1.0095 |
| 0.8579 | 0.8 | 16 | 0.4390 |
| 0.2699 | 1.2 | 24 | 0.3492 |
| 0.1648 | 1.6 | 32 | 0.3064 |
| 0.2503 | 2.0 | 40 | 0.2721 |
### Framework versions
- PEFT 0.15.2
- Transformers 4.53.0.dev0
- Pytorch 2.7.1+cu126
- Datasets 3.6.0
- Tokenizers 0.21.1
|
Ywhsheng/Llama-3.1-8B-bnb-4bit-baigou
|
Ywhsheng
| 2025-06-08T03:51:04Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"gguf",
"llama",
"text-generation-inference",
"unsloth",
"trl",
"en",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T03:13:28Z |
---
base_model: unsloth/meta-llama-3.1-8b-unsloth-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- trl
license: apache-2.0
language:
- en
---
# Uploaded model
- **Developed by:** Ywhsheng
- **License:** apache-2.0
- **Finetuned from model :** unsloth/meta-llama-3.1-8b-unsloth-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
BootesVoid/cmbn0vnow01o3ekg02np0v5vh_cmbn38qam01qbekg0jxahuq30
|
BootesVoid
| 2025-06-08T03:50:26Z | 0 | 0 |
diffusers
|
[
"diffusers",
"flux",
"lora",
"replicate",
"text-to-image",
"en",
"base_model:black-forest-labs/FLUX.1-dev",
"base_model:adapter:black-forest-labs/FLUX.1-dev",
"license:other",
"region:us"
] |
text-to-image
| 2025-06-08T03:50:25Z |
---
license: other
license_name: flux-1-dev-non-commercial-license
license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md
language:
- en
tags:
- flux
- diffusers
- lora
- replicate
base_model: "black-forest-labs/FLUX.1-dev"
pipeline_tag: text-to-image
# widget:
# - text: >-
# prompt
# output:
# url: https://...
instance_prompt: KYLIE
---
# Cmbn0Vnow01O3Ekg02Np0V5Vh_Cmbn38Qam01Qbekg0Jxahuq30
<Gallery />
## About this LoRA
This is a [LoRA](https://replicate.com/docs/guides/working-with-loras) for the FLUX.1-dev text-to-image model. It can be used with diffusers or ComfyUI.
It was trained on [Replicate](https://replicate.com/) using AI toolkit: https://replicate.com/ostris/flux-dev-lora-trainer/train
## Trigger words
You should use `KYLIE` to trigger the image generation.
## Run this LoRA with an API using Replicate
```py
import replicate
input = {
"prompt": "KYLIE",
"lora_weights": "https://huggingface.co/BootesVoid/cmbn0vnow01o3ekg02np0v5vh_cmbn38qam01qbekg0jxahuq30/resolve/main/lora.safetensors"
}
output = replicate.run(
"black-forest-labs/flux-dev-lora",
input=input
)
for index, item in enumerate(output):
with open(f"output_{index}.webp", "wb") as file:
file.write(item.read())
```
## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
```py
from diffusers import AutoPipelineForText2Image
import torch
pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda')
pipeline.load_lora_weights('BootesVoid/cmbn0vnow01o3ekg02np0v5vh_cmbn38qam01qbekg0jxahuq30', weight_name='lora.safetensors')
image = pipeline('KYLIE').images[0]
```
For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
## Training details
- Steps: 2000
- Learning rate: 0.0004
- LoRA rank: 16
## Contribute your own examples
You can use the [community tab](https://huggingface.co/BootesVoid/cmbn0vnow01o3ekg02np0v5vh_cmbn38qam01qbekg0jxahuq30/discussions) to add images that show off what you’ve made with this LoRA.
|
zaydzuhri/myopic-1.8B-4096-model
|
zaydzuhri
| 2025-06-08T03:45:08Z | 0 | 0 | null |
[
"safetensors",
"transformer",
"region:us"
] | null | 2025-06-08T03:38:16Z |
<div align="center">
# 🔥 Flame: Flash Linear Attention Made Easy
</div>
Welcome to 🔥 `flame`, a minimal and efficient framework built on `torchtitan` for training Flash Linear Attention (FLA) models (and more broadly, arbitrary autoregressive language models) with blazing efficiency.
**Feature Highlights:**
- 🚀 Minimal, easy-to-use, extensible training framework
- 🤗 Seamless integration with `fla` and `transformers`
- 🔄 Zero-cost data preprocessing: online tokenization, dataset shuffling, and multiple datasets support
- 🔮 4D parallelism (coming soon)
## Setup
To get started, clone the `flame` repository and install the required dependencies:
```bash
git clone https://github.com/fla-org/flame.git
cd flame
pip install .
```
`flame` manages minimal dependencies, only including `fla` and `torchtitan` as submodules.
After installation, initialize and update the submodules:
```sh
git submodule update --init --recursive
```
## Dataset Preparation
To download the dataset to your local disk, create a new Python file with the following content and execute it:
```py
from datasets import load_dataset
# load fineweb-edu with parallel processing
dataset = load_dataset("HuggingFaceFW/fineweb-edu", name="default", num_proc=64, cache_dir="/your/cache/path")
# or load a subset with roughly 100B tokens, suitable for small- or medium-sized experiments
dataset = load_dataset("HuggingFaceFW/fineweb-edu", name="sample-100BT", num_proc=64, cache_dir="/your/cache/path")
```
## Training Recipes
Here's an example of training a 340M FLA Transformer model with a LLaMA-like architecture from scratch on a 100BT subset of the Fineweb-edu corpus in streaming mode.
> [!WARNING]
> If the dataset is not downloaded beforehand, the streaming mode will attempt to fetch it from a remote server and download it on-the-fly, which can be highly unstable during training due to network issues.
> For stable training, ensure the dataset is downloaded locally (see [**Dataset Preparation**](#dataset-preparation)). Otherwise, we assume you are only testing the new corpus.
```sh
bash train.sh \
--job.config_file flame/models/fla.toml \
--job.dump_folder exp/transformer-340M-4K-10B/batch1.seqlen65536.context4096.warmup1024.update1.steps20480.lr3e-4.cosine \
--model.config configs/transformer_340M.json \
--model.tokenizer_path fla-hub/transformer-1.3B-100B \
--optimizer.name AdamW \
--optimizer.eps 1e-15 \
--optimizer.lr 3e-4 \
--lr_scheduler.warmup_steps 1024 \
--lr_scheduler.lr_min 0.1 \
--lr_scheduler.decay_type cosine \
--training.batch_size 1 \
--training.seq_len 65536 \
--training.context_len 4096 \
--training.varlen \
--training.gradient_accumulation_steps 1 \
--training.steps 20480 \
--training.max_norm 1.0 \
--training.skip_nan_inf \
--training.dataset HuggingFaceFW/fineweb-edu \
--training.dataset_name sample-100BT \
--training.dataset_split train \
--training.streaming \
--training.num_workers 32 \
--training.prefetch_factor 2 \
--training.seed 42 \
--training.compile \
--checkpoint.interval 2048 \
--checkpoint.load_step -1 \
--checkpoint.keep_latest_k 2 \
--metrics.log_freq 1
```
You can specify the number of GPUs by setting the environment variable `NGPU`, which defaults to 8.
**For single-GPU debugging, set `NGPU=1`.**
We provide several [config files](https://github.com/fla-org/flame/tree/main/configs) for different models.
By default, the learning rate is set to 3e-4 with a cosine scheduler. Other schedulers, such as WSD (wsd), are also supported.
**Key parameters:**
- `--lr_scheduler.decay_ratio`: The proportion of the steps allocated to the decay phase. The learning rate will remain stable after the warmup period and only start decaying during the last `decay_ratio` portion of the total training steps, which is known as the Warmup-Stable-Decay (WSD) schedule.
- `--lr_scheduler.warmup_steps`: The number of steps for the learning rate warmup phase.
- `--training.steps`: Total number of training steps.
- `--training.batch_size`: Batch size per device, must be 1 if `--training.varlen` is set.
- `--training.seq_len`: The length of each sequence in the batch, which is concatenated from multiple samples.
- `--training.context_len`: The max allowed length of a sample. For non-varlen mode, this is equivalent to `seq_len`.
- `--training.varlen`: Whether to conduct variable-length sequence training.
- `--training.gradient_accumulation_steps`: Number of gradient accumulation steps.
> [!WARNING]
> The total number of tokens processed per batch, referred to as `global_batch_size`, is calculated as batch_size × gradient_accumulation_steps × num_gpus.
> Each step processes `global_batch_size * seq_len` tokens.
> Monitor the value of `global_batch_size`, `warmup_steps`, and `steps` carefully when modifying any of the hyperparameters!
For a detailed explanation of all parameters, run:
```sh
bash train.sh -h
```
<details>
<summary>Usage</summary>
```py
options:
-h, --help show this help message and exit
--job.config_file JOB.CONFIG_FILE
Job config file
--job.dump_folder JOB.DUMP_FOLDER
Folder to dump job outputs
--job.description JOB.DESCRIPTION
Description of the job
--job.use_for_integration_test
Add this config to the integration test suite
--job.print_args Print the args to terminal
--model.config MODEL.CONFIG
Path to the model config
--model.norm_type MODEL.NORM_TYPE
Type of layer normalization to use [layernorm,
np_layernorm, rmsnorm, fused_rmsnorm]
--model.tokenizer_path MODEL.TOKENIZER_PATH
Tokenizer path
--profiling.enable_profiling
Whether to enable pytorch profiler
--profiling.save_traces_folder PROFILING.SAVE_TRACES_FOLDER
Trace files location
--profiling.profile_freq PROFILING.PROFILE_FREQ
How often to collect profiler traces, in iterations
--profiling.enable_memory_snapshot
Whether to dump memory snapshot
--profiling.save_memory_snapshot_folder PROFILING.SAVE_MEMORY_SNAPSHOT_FOLDER
Memeory snapshot files location
--optimizer.name OPTIMIZER.NAME
Optimizer to use
--optimizer.eps OPTIMIZER.EPS
Epsilon value for the optimizer.
--optimizer.fused Whether the fused implementation(CUDA only) is used.
--optimizer.scheduler {wsd,cosine,linear}
Scheduler to use. Currently supported: wsd, cosine,
and linear.
--optimizer.lr OPTIMIZER.LR
Learning rate to use
--optimizer.min_lr_ratio OPTIMIZER.MIN_LR_RATIO
Min lr ratio for lr scheduler
--optimizer.early_step_in_backward
Whether to apply optimizer in the backward. Caution,
optimizer_in_backward is not compatible with gradients
clipping, users should not call
register_post_accumulate_grad_hook after the optimizer
is built.
--training.batch_size TRAINING.BATCH_SIZE
Batch size
--training.seq_len TRAINING.SEQ_LEN
Sequence length
--training.context_len TRAINING.CONTEXT_LEN
Max length allowed for each sequence
--training.varlen Whether to take sequences of variable length as input
--training.warmup_steps TRAINING.WARMUP_STEPS
Steps for lr scheduler warmup, normally 1/5 of
--training.steps
--training.gradient_accumulation_steps TRAINING.GRADIENT_ACCUMULATION_STEPS
Number of steps to accumulate gradients before
updating parameters
--training.steps TRAINING.STEPS
How many train steps to run
--training.max_norm TRAINING.MAX_NORM
Max norm for gradient clipping
--training.skip_nan_inf
Skip batch updates when NaN or INF gradients are
encountered during training
--training.dataset TRAINING.DATASET
Dataset to use, with comma separated values
--training.dataset_name TRAINING.DATASET_NAME
The name of the dataset config, with comma separated
values if provided
--training.dataset_split TRAINING.DATASET_SPLIT
Dataset split to use, with comma separated values if
provided
--training.data_dir TRAINING.DATA_DIR
Data dirs to use, with comma separated values if
provided
--training.data_files TRAINING.DATA_FILES
Data files to use, with comma separated values if
provided
--training.data_probs TRAINING.DATA_PROBS
Data sampling probabilities, with comma separated
values if provided
--training.streaming Whether to load dataset in streaming mode, used for
huge dataset
--training.num_workers TRAINING.NUM_WORKERS
Number of subprocesses to use for data loading. 0
means that the data will be loaded in the main
process.
--training.prefetch_factor TRAINING.PREFETCH_FACTOR
Number of batches loaded in advance by each worker.2
means there will be a total of 2 * num_workers batches
prefetched across all workers.
--training.data_parallel_replicate_degree TRAINING.DATA_PARALLEL_REPLICATE_DEGREE
The `data_parallel_replicate_degree` argument
specifies the degree of data parallelism for weight
replication. When this value is greater than 1,
weights will be replicated across
`data_parallel_replicate_degree` ranks. If
`data_parallel_shard_degree` is also greater than 1,
the parallelism method used is HSDP (Hybrid Sharded
Data Parallelism). Otherwise, the parallelism method
used is DDP (Distributed Data Parallelism). 1 means
disabled.
--training.data_parallel_shard_degree TRAINING.DATA_PARALLEL_SHARD_DEGREE
The `data_parallel_shard_degree` argument specifies
the degree of data parallelism for weight sharding.
When this value is greater than 1, weights will be
sharded across `data_parallel_shard_degree` ranks. If
`data_parallel_replicate_degree` is also greater than
1, the parallelism method used is HSDP (Hybrid Sharded
Data Parallelism). Otherwise, the parallelism method
used is FSDP (Fully Sharded Data Parallelism). -1
means leftover ranks will be used (After
DP_REPLICATE/SP/PP). Note that only
`data_parallel_shard_degree` can be negative. 1 means
disabled.
--training.enable_cpu_offload
Whether to apply CPU offloading of parameters,
gradients, and optimizer states in FSDP
--training.tensor_parallel_degree TRAINING.TENSOR_PARALLEL_DEGREE
Tensor Parallelism degree. 1 means disabled.
--training.disable_loss_parallel
Whether to apply loss parallel when sequence parallel
is enabled
--training.mixed_precision_param {bfloat16,float32}
torch dtype to use for parameters when applying mixed
precision via FSDP. This feature only takes effect
when data_parallel_shard_degree > 1
--training.mixed_precision_reduce {float32}
torch dtype to use for reductions when applying mixed
precision via FSDP. This feature only takes effect
when data_parallel_shard_degree > 1
--training.compile Whether to compile the model
--training.gc_freq TRAINING.GC_FREQ
Python garbage control scheduling interval, in steps
--training.seed TRAINING.SEED
Choose the base RNG seed used for training
--training.deterministic
Use deterministic algorithms wherever possible, may be
slower
--metrics.log_freq METRICS.LOG_FREQ
How often to log metrics to TensorBoard, in iterations
--metrics.enable_tensorboard
Whether to log metrics to TensorBoard
--metrics.disable_color_printing
Whether to disable color printing in logs
--metrics.save_tb_folder METRICS.SAVE_TB_FOLDER
Folder to dump TensorBoard states
--metrics.rank_0_only
Whether to save TensorBoard metrics only for rank 0 or
for all ranks. When pipeline_parallel_degree is > 1,
this option uses the 0th rank of the last stage
pipeline group, which is the only stage that computes
loss metrics.
--metrics.enable_wandb
Whether to log metrics to Weights & Biases
--experimental.enable_async_tensor_parallel
Whether to apply async tensor parallel (currently only
effective when compile is enabled)
--experimental.pipeline_parallel_degree EXPERIMENTAL.PIPELINE_PARALLEL_DEGREE
Pipeline Parallelism degree, or number of ranks. 1
means disabled. If using looped schedules, this still
specifies the number of physical ranks, not the number
of stages. Stages per rank are inferred from split
points degree, and schedule.
--experimental.pipeline_parallel_split_points EXPERIMENTAL.PIPELINE_PARALLEL_SPLIT_POINTS [EXPERIMENTAL.PIPELINE_PARALLEL_SPLIT_POINTS ...]
Specify comma-separated names of modules to use as the
beginning of a split point. e.g. "layers.0,layers.2"
will cause the model to be split into 3 stages, the
first containing all the layers up to layers.0, the
second containing layers.0 and up to layers.2, the
third containing layers.2 and all the remaining
layers. Note: fully-automated splitting may be enabled
in the future, but currently the split points must be
specified manually.
--experimental.pipeline_parallel_schedule EXPERIMENTAL.PIPELINE_PARALLEL_SCHEDULE
Specify the Pipeline Parallel schedule to use. The
supported schedules are: https://github.com/pytorch/py
torch/blob/de4c2a3b4e89d96334dc678d1c3f2ae51a6630a0/to
rch/distributed/pipelining/schedules.py#L2161. The
schedule must be compatible with the split points and
stages_per_rank. Looped schedules (e.g.
Interleaved1F1B) require specifying
pipeline_parallel_degree = number of ranks, and
split_points = number of stages - 1
--experimental.pipeline_parallel_schedule_csv EXPERIMENTAL.PIPELINE_PARALLEL_SCHEDULE_CSV
Specify the path to the pipeline parallel schedule csv
file to use. The pipeline_parallel_schedule argument
must be either PipelineScheduleSingle,
PipelineScheduleMulti, or _PipelineScheduleRuntime.
--experimental.pipeline_parallel_microbatches EXPERIMENTAL.PIPELINE_PARALLEL_MICROBATCHES
How many microbatches to split the global training
batch into when using pipeline parallelism. The global
training batch size must be evenly divisible by the
number of microbatches. The default value will be the
number of pipeline stages, if unspecified.
--experimental.enable_compiled_autograd
Enable CompiledAutograd to compile the backward.
--experimental.context_parallel_degree EXPERIMENTAL.CONTEXT_PARALLEL_DEGREE
Context parallelism degree. 1 means disabled.
--experimental.context_parallel_rotate_method EXPERIMENTAL.CONTEXT_PARALLEL_ROTATE_METHOD
The collective to use in context parallel SDPA for kv
shards exchange. 'allgather' means to all-gather all
kv shards on ranks after the first sub-SDPA
computation, 'alltoall' means to all-to-all shuffle
the kv shards. The default value is 'allgather'.
--checkpoint.enable_checkpoint
Whether to enable checkpoint
--checkpoint.folder CHECKPOINT.FOLDER
The folder to store the checkpoints. When
enable_checkpoint is set to true, checkpoints will be
in {--job.dump_folder}/{--checkpoint.folder}.
--checkpoint.interval_type CHECKPOINT.INTERVAL_TYPE
Checkpointing interval unit of measurement ['step',
'seconds']
--checkpoint.interval CHECKPOINT.INTERVAL
Checkpointing interval, in steps or seconds depending
on --checkpoint.interval_type
--checkpoint.model_weights_only
When model_weights_only=True, only model weights will
be saved at the end of training. With this,
checkpoints can be loaded using `torch.load(...,
weights_only=True)` after conversion. When
model_weights_only=False, the full checkpoint will be
saved. A full checkpoint includes model, optimizer and
train_state, which can be used to resume training. The
default value is false.
--checkpoint.export_dtype {float16,bfloat16,float32}
Converts to the specified precision when training
completes and model_weights_only=true. Currently
supports float32, float16, and bfloat16. The default
value is float32.
--checkpoint.create_seed_checkpoint
Initializes the full model without applying
parallelisms, and then saves it as a seed checkpoint.
Note: requires user to call train.py without
specifying any parallelisms, e.g. NGPU=1. Could be
implemented as a separate script, but this way shares
more code.
--checkpoint.async_mode CHECKPOINT.ASYNC_MODE
Which async checkpoint mode to use. Currently there
are 3 different modes. 1. "disabled": synchronized
checkpointing will be used. 2. "async":
torch.distributed.checkpoint.async_save will be used.
1. "async_with_pinned_mem": this option utilizes a
dedicated pinned memory space and creates a separate
process for faster GPU->CPU transfer performance and
eliminating GIL contention. The cost is increased CPU
memory usage. If insufficient CPU memory is available,
performance may degrade due to memory paging. For most
users, "async" should suffice as the performance
overhead is typically small (on the order of tens of
seconds) compared to checkpointing frequency. This
mode can be employed to pursue near-zero checkpointing
times (e.g., < 1 second) given appropriate hardware
support such as ample CPU memory and fast PCIe.
"disabled" is the default mode.
--checkpoint.keep_latest_k CHECKPOINT.KEEP_LATEST_K
Keeps only the latest k checkpoints, and purging older
ones. If 0, keep all checkpoints. 0 is the default
value.
--checkpoint.load_step CHECKPOINT.LOAD_STEP
Load the checkpoint at the specified step. If -1, load
the latest checkpoint.
--float8.enable_float8_linear
If true, swaps `torch.nn.Linear` with `Float8Linear`.
This feature requires you to install 'torchao' which
can be found here: https://github.com/pytorch/ao
--float8.enable_fsdp_float8_all_gather
Whether enable float8 all-gather in FSDP
--float8.precompute_float8_dynamic_scale_for_fsdp
Whether precompute float8 scales dynamically for FSDP
--float8.scaling_type_input {dynamic,delayed}
float8 scaling for input, dynamic (default) or delayed
--float8.scaling_type_weight FLOAT8.SCALING_TYPE_WEIGHT
float8 scaling for input, dynamic (default) or delayed
--float8.scaling_type_grad_output FLOAT8.SCALING_TYPE_GRAD_OUTPUT
float8 scaling for input, dynamic (default) or delayed
--comm.init_timeout_seconds COMM.INIT_TIMEOUT_SECONDS
Timeout for communication operations, during
initialization and first train step.
--comm.train_timeout_seconds COMM.TRAIN_TIMEOUT_SECONDS
Timeout for communication operations after the first
train step -- usually a tighter bound than during
initialization.
--comm.trace_buf_size COMM.TRACE_BUF_SIZE
Flight recorder ring buffer size, >0 means recording
by default, 0 means disabled
--memory_estimation.enabled
Whether to estimate memory usage for FSDP
--memory_estimation.disable_fake_mode
Whether to estimate memory under FakeTensorMode
```
</details>
### Training with `torch.compile`
Starting from `torch 2.0`, `torch.compile` has been introduced as a new feature to seamlessly accelerate training processes.
In `flame`, one can simply enable `torch.compile` by adding `--training.compile` flag to your training script.
However, `fla` has integrated numerous fused kernels for acceleration, which may potentially conflict with `torch.compile`.
We are actively working on resolving these issues to make compilation transparent to users.
In the meantime, please ensure you are using the latest dependencies.
Specifically, **we recommend using `torch>=2.6` and `triton>=3.0`**.
### Training with multiple datasets
If you wish to train a model with all-round capabilities (e.g., code, math, and multilingual ability), it's necessary to train on multiple datasets.
`flame` allows training with multiple datasets easily.
For example, you can specify the following arguments to train on 6 datasets with different proportions:
```sh
--training.dataset HuggingFaceFW/fineweb-edu,opencsg/Fineweb-Edu-Chinese-V2.1,OpenCoder-LLM/opc-fineweb-code-corpus,math-ai/AutoMathText,EleutherAI/proof-pile-2,OpenCoder-LLM/opc-fineweb-math-corpus \
--training.data_probs 0.6,0.15,0.15,0.014,0.058,0.028 \
```
### ~Finalizing training~
> [!NOTE]
> We have done this conversion automatically in the training script since our latest updates.
Once training is complete, you may want to convert the distributed checkpoints (DCPs) into the 🤗 format for broader use.
To facilitate this, we provide a straightforward conversion script:
```sh
python -m flame.utils.convert_dcp_to_hf --path <path_to_model> --step <step> --config <path_to_config> --tokenizer <path_to_tokenizer>
```
After this, your model will be in the 🤗 format, ready to be shared or deployed.
You can then easily publish your model using the `huggingface_hub` for wider accessibility.
### Continual training
If you wish to build upon a strong pre-trained model (in 🤗 format) and continue training, we also offer a script to convert the 🤗 format model back into DCP format.
This allows you to seamlessly resume training with `flame`.
```sh
python -m flame.utils.convert_hf_to_dcp --model <path_to_hf> --checkpoint <path_to_dcp/checkpoint/step-0>
```
Here, `<path_to_dcp>` is the directory where your distributed checkpoints will be stored.
The checkpoint is intentionally saved at `<step-0>` within the checkpoint folder to ensure it is loadable by `flame` during the initial training step, similar to how a seed checkpoint is handled.
Once the conversion is complete, you can proceed with training using `flame` as usual, continuing from where the pretrained model left off.
## Multi-node training
If you have access to multi-node GPUs, consider leveraging them for optimal performance.
This process is straightforward and well-documented in the PyTorch [docs](https://pytorch.org/docs/stable/elastic/run.html).
To set up multi-node training:
* Set the environment variables `MASTER_ADDR=<ip>` and `MASTER_PORT=<port>` before running the training script across all nodes.
* If you're using a job scheduler like Slurm, it will handle these variables for you.
`torchtitan` provides a [Slurm script](https://github.com/pytorch/torchtitan/blob/main/multinode_trainer.slurm) for multi-node training, which you can use as a reference or starting point.
|
new-payal-gaming-videos/full.video.payal.gaming.viral.video.original.4k.link
|
new-payal-gaming-videos
| 2025-06-08T03:40:55Z | 0 | 0 | null |
[
"region:us"
] | null | 2025-06-08T03:40:21Z |
<p><a rel="nofollow" href="https://viralflix.xyz/leaked/?ffd">►►✅ 𝘾𝙇𝙄𝘾𝙆 𝙃𝙀𝙍𝙀 ==►► 𝙁𝙪𝙡𝙡 𝙑𝙞𝙙𝙚𝙤️​</a></p>
<a rel="nofollow" href="https://viralflix.xyz/leaked/?ffd">🔴 CLICK HERE 🌐==►► Download Now)</a>
<a rel="nofollow" href="https://viralflix.xyz/leaked/?ffd"><img src="https://i.postimg.cc/qvPp49Sm/ythngythg.gif" alt="fsd"></a>
|
fernandabufon/model_bertimbau_base_toxicity_5_1e-05_0.01_0.1_16_fold_3
|
fernandabufon
| 2025-06-08T03:37:34Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"bert",
"text-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-classification
| 2025-06-08T03:37:12Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
gradientrouting-spar/mc_badmed_align_train_size__seed_1
|
gradientrouting-spar
| 2025-06-08T03:33:57Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | 2025-06-07T19:40:23Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
gradientrouting-spar/mc_badmed_align_train_size__seed_1_epoch_1
|
gradientrouting-spar
| 2025-06-08T03:33:46Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | 2025-06-07T19:40:10Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
18-VIDEOS-kiffy-katrinalim123-VIDEO-hq/ORIGINAL.VIDEO.Katrina.Lim.Viral.Video.Tutorial.LINK.Official
|
18-VIDEOS-kiffy-katrinalim123-VIDEO-hq
| 2025-06-08T03:32:55Z | 0 | 0 | null |
[
"region:us"
] | null | 2025-06-08T03:32:10Z |
<p><a rel="nofollow" href="https://viralflix.xyz/leaked/?ffd">►►✅ 𝘾𝙇𝙄𝘾𝙆 𝙃𝙀𝙍𝙀 ==►► 𝙁𝙪𝙡𝙡 𝙑𝙞𝙙𝙚𝙤️​</a></p>
<a rel="nofollow" href="https://viralflix.xyz/leaked/?ffd">🔴 CLICK HERE 🌐==►► Download Now)</a>
<a rel="nofollow" href="https://viralflix.xyz/leaked/?ffd"><img src="https://i.postimg.cc/qvPp49Sm/ythngythg.gif" alt="fsd"></a>
|
I0ome/Lymon
|
I0ome
| 2025-06-08T03:31:40Z | 0 | 0 | null |
[
"license:apache-2.0",
"region:us"
] | null | 2025-06-08T03:31:40Z |
---
license: apache-2.0
---
|
Fizzarolli/l3-8b-kto-ckpt144
|
Fizzarolli
| 2025-06-08T03:22:35Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged",
"base_model:adapter:allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged",
"region:us"
] | null | 2025-06-08T03:22:29Z |
---
base_model: allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.15.2
|
EmaRimoldi/mnlp-raft-qwen
|
EmaRimoldi
| 2025-06-08T03:17:15Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen3",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T00:30:00Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
ScottBiggs2/tinyllama_detective_test
|
ScottBiggs2
| 2025-06-08T03:13:49Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"base_model:adapter:TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"region:us"
] | null | 2025-06-08T02:04:43Z |
---
base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.10.0
|
publication-charaf/MCQ_Qwen3-0.6B-Base_lr-0.001_e-5_s-0
|
publication-charaf
| 2025-06-08T03:08:04Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen3",
"text-generation",
"generated_from_trainer",
"trl",
"sft",
"conversational",
"base_model:Qwen/Qwen3-0.6B-Base",
"base_model:finetune:Qwen/Qwen3-0.6B-Base",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T00:20:36Z |
---
base_model: Qwen/Qwen3-0.6B-Base
library_name: transformers
model_name: MCQ_Qwen3-0.6B-Base_lr-0.001_e-5_s-0
tags:
- generated_from_trainer
- trl
- sft
licence: license
---
# Model Card for MCQ_Qwen3-0.6B-Base_lr-0.001_e-5_s-0
This model is a fine-tuned version of [Qwen/Qwen3-0.6B-Base](https://huggingface.co/Qwen/Qwen3-0.6B-Base).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="publication-charaf/MCQ_Qwen3-0.6B-Base_lr-0.001_e-5_s-0", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/kamel-charaf-epfl/huggingface/runs/jath4lrb)
This model was trained with SFT.
### Framework versions
- TRL: 0.17.0
- Transformers: 4.51.3
- Pytorch: 2.6.0
- Datasets: 3.5.0
- Tokenizers: 0.21.1
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
bytedance-research/Valley2-DPO
|
bytedance-research
| 2025-06-08T03:06:03Z | 34 | 2 | null |
[
"safetensors",
"valley",
"custom_code",
"arxiv:2501.05901",
"base_model:Qwen/Qwen2.5-7B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-7B-Instruct",
"license:apache-2.0",
"region:us"
] | null | 2025-04-14T08:18:01Z |
---
license: apache-2.0
base_model:
- Qwen/Qwen2.5-7B-Instruct
---
# Valley 2.0
<p align="center">
<img src="https://raw.githubusercontent.com/bytedance/Valley/refs/heads/main/assets/valley_logo.jpg" width="500"/>
<p>
<p align="center">
🎮️ <a href="https://github.com/bytedance/Valley">Github</a>   |    🤗 <a href="https://huggingface.co/bytedance-research/Valley-Eagle-7B">Hugging Face</a>   |   🤖 <a href="https://www.modelscope.cn/models/Hyggge/Valley-Eagle-7B">ModelScope</a>    |    📑 <a href="https://hyggge.github.io/projects/valley/index.html">Home Page</a>    |    📙 <a href="https://arxiv.org/abs/2501.05901">Paper</a>
</p>
## Introduction
Valley is a cutting-edge multimodal large model designed to handle a variety of tasks involving text, images, and video data, which is developed by ByteDance. Our model not only
- Achieved the best results in the inhouse e-commerce and short-video benchmarks
- Demonstrated comparatively outstanding performance in the OpenCompass (average scores > 67) tests
when evaluated against models of the same scale.
## Release
- [02/15] 🔥 Update Valley-Eagle-DPO, achieve 69.6 on OpenCompass and update AutoModel usage for checkpoints.
- [01/13] 🔥 Release TechReport. [Valley2: Exploring Multimodal Models with Scalable Vision-Language Design](https://arxiv.org/abs/2501.05901)
- [12/23] Announcing [Valley-Qwen2.5-7B](https://huggingface.co/ByteDance)!
## Valley-Eagle
The foundational version of Valley is a multimodal large model aligned with Siglip and Qwen2.5, incorporating LargeMLP and ConvAdapter to construct the projector.
- In the final version, we also referenced Eagle, introducing an additional VisionEncoder that can flexibly adjust the number of tokens and is parallelized with the original visual tokens.
- This enhancement supplements the model’s performance in extreme scenarios, and we chose the Qwen2vl VisionEncoder for this purpose.
and the model structure is shown as follows:
<div style="display:flex;">
<img src="valley_structure.jpeg" alt="opencompass" style="height:600px;" />
</div>
## Environment Setup
``` bash
pip install torch==2.4.0 torchvision==0.19.0 torchaudio==2.4.0 --index-url https://download.pytorch.org/whl/cu121
pip install -r requirements.txt
```
## License Agreement
All of our open-source models are licensed under the Apache-2.0 license.
## Related Project
We list related Project
- [Valley: Video Assistant with Large Language model Enhanced abilitY](https://github.com/RupertLuo/Valley)
- [LLaVA: Large Language and Vision Assistant](https://github.com/haotian-liu/LLaVA)
- [Eagle: Exploring The Design Space for Multimodal LLMs with Mixture of Encoders](https://github.com/NVlabs/EAGLE)
- [LLaVA-CoT: Let Vision Language Models Reason Step-by-Step](https://github.com/PKU-YuanGroup/LLaVA-CoT)
- [Qwen2.5](https://github.com/QwenLM/Qwen2.5)
## License Agreement
All of our open-source models are licensed under the [Apache-2.0](./LICENSE) license.
## We are Hiring
The Data-Ecommerce-Platform Governance-Basic Algorithms Team focuses on the research and development of multi-modal large model algorithms and foundational algorithms, continuously delving deeply into this field. Our mission is to optimize algorithms and collaborate with business teams to comprehensively govern the quality and ecosystem of ByteDance's e-commerce products. Currently, the team has a strong demand for foundational algorithm expertise in NLP, CV, and multimodal technologies. We welcome inquiries and look forward to working on challenging projects with talented individuals like you!
Location: Beijing / Shanghai / Singapore
Contact & Resume Submission: wuheng.2024@bytedance.com
> Tiktok-电商,基础算法团队专注于多模态大模型算法和基础算法的研发,并在此方向上持续深耕,期待和优秀的你(实习/全职),一起做有挑战的事情!
>
> 岗位城市:北京/上海/新加坡
>
> 咨询&简历投递:wuheng.2024@bytedance.com
## Citation
```
@article{wu2025valley2,
title={Valley2: Exploring Multimodal Models with Scalable Vision-Language Design},
author={Wu, Ziheng and Chen, Zhenghao and Luo, Ruipu and Zhang, Can and Gao, Yuan and He, Zhentao and Wang, Xian and Lin, Haoran and Qiu, Minghui},
journal={arXiv preprint arXiv:2501.05901},
year={2025}
}
```
|
HouraMor/wh-ft-lre5-adm-ga1ba16-st15k
|
HouraMor
| 2025-06-08T03:04:55Z | 3 | 0 |
transformers
|
[
"transformers",
"tensorboard",
"safetensors",
"whisper",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:openai/whisper-large-v3",
"base_model:finetune:openai/whisper-large-v3",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
automatic-speech-recognition
| 2025-06-06T09:04:33Z |
---
library_name: transformers
license: apache-2.0
base_model: openai/whisper-large-v3
tags:
- generated_from_trainer
metrics:
- wer
model-index:
- name: wh-ft-lre5-adm-ga1ba16-st15k
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wh-ft-lre5-adm-ga1ba16-st15k
This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.9483
- Wer: 0.4505
- Cer: 0.3542
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- training_steps: 15000
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
|:-------------:|:------:|:-----:|:---------------:|:------:|:------:|
| 1.0084 | 0.1994 | 1000 | 1.0466 | 0.9069 | 0.7737 |
| 0.9075 | 0.3989 | 2000 | 0.9939 | 0.8817 | 0.7542 |
| 0.8338 | 0.5983 | 3000 | 0.9700 | 1.0308 | 0.9146 |
| 0.7714 | 0.7978 | 4000 | 0.9459 | 0.8932 | 0.7805 |
| 0.8347 | 0.9972 | 5000 | 0.9293 | 1.1874 | 1.0844 |
| 0.5472 | 1.1966 | 6000 | 0.9344 | 0.5475 | 0.4397 |
| 0.5814 | 1.3961 | 7000 | 0.9288 | 0.6254 | 0.5069 |
| 0.5562 | 1.5955 | 8000 | 0.9250 | 0.6017 | 0.4884 |
| 0.4887 | 1.7950 | 9000 | 0.9108 | 0.5641 | 0.4586 |
| 0.4809 | 1.9944 | 10000 | 0.9020 | 0.5632 | 0.4532 |
| 0.3191 | 2.1939 | 11000 | 0.9505 | 0.4814 | 0.3804 |
| 0.3291 | 2.3933 | 12000 | 0.9525 | 0.5196 | 0.4107 |
| 0.3319 | 2.5927 | 13000 | 0.9563 | 0.4514 | 0.3528 |
| 0.3249 | 2.7922 | 14000 | 0.9487 | 0.4435 | 0.3490 |
| 0.2724 | 2.9916 | 15000 | 0.9483 | 0.4505 | 0.3542 |
### Framework versions
- Transformers 4.52.3
- Pytorch 2.7.0+cu118
- Datasets 3.5.1
- Tokenizers 0.21.1
|
Fizzarolli/l3-8b-kto-ckpt125
|
Fizzarolli
| 2025-06-08T03:04:35Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged",
"base_model:adapter:allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged",
"region:us"
] | null | 2025-06-08T03:04:29Z |
---
base_model: allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.15.2
|
sergbese/byt5-base-isv-ru-translator
|
sergbese
| 2025-06-08T03:03:37Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"t5",
"text2text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text2text-generation
| 2025-06-08T03:02:40Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
RichardErkhov/ehristoforu_-_HappyLlama1-gguf
|
RichardErkhov
| 2025-06-08T03:00:30Z | 0 | 0 | null |
[
"gguf",
"endpoints_compatible",
"region:us",
"conversational"
] | null | 2025-06-08T01:47:43Z |
Quantization made by Richard Erkhov.
[Github](https://github.com/RichardErkhov)
[Discord](https://discord.gg/pvy7H8DZMG)
[Request more models](https://github.com/RichardErkhov/quant_request)
HappyLlama1 - GGUF
- Model creator: https://huggingface.co/ehristoforu/
- Original model: https://huggingface.co/ehristoforu/HappyLlama1/
| Name | Quant method | Size |
| ---- | ---- | ---- |
| [HappyLlama1.Q2_K.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q2_K.gguf) | Q2_K | 2.96GB |
| [HappyLlama1.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.IQ3_XS.gguf) | IQ3_XS | 3.28GB |
| [HappyLlama1.IQ3_S.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.IQ3_S.gguf) | IQ3_S | 3.43GB |
| [HappyLlama1.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q3_K_S.gguf) | Q3_K_S | 3.41GB |
| [HappyLlama1.IQ3_M.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.IQ3_M.gguf) | IQ3_M | 3.52GB |
| [HappyLlama1.Q3_K.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q3_K.gguf) | Q3_K | 3.74GB |
| [HappyLlama1.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q3_K_M.gguf) | Q3_K_M | 3.74GB |
| [HappyLlama1.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q3_K_L.gguf) | Q3_K_L | 4.03GB |
| [HappyLlama1.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.IQ4_XS.gguf) | IQ4_XS | 4.18GB |
| [HappyLlama1.Q4_0.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q4_0.gguf) | Q4_0 | 4.34GB |
| [HappyLlama1.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.IQ4_NL.gguf) | IQ4_NL | 4.38GB |
| [HappyLlama1.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q4_K_S.gguf) | Q4_K_S | 4.37GB |
| [HappyLlama1.Q4_K.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q4_K.gguf) | Q4_K | 4.58GB |
| [HappyLlama1.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q4_K_M.gguf) | Q4_K_M | 4.58GB |
| [HappyLlama1.Q4_1.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q4_1.gguf) | Q4_1 | 4.78GB |
| [HappyLlama1.Q5_0.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q5_0.gguf) | Q5_0 | 5.21GB |
| [HappyLlama1.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q5_K_S.gguf) | Q5_K_S | 5.21GB |
| [HappyLlama1.Q5_K.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q5_K.gguf) | Q5_K | 5.34GB |
| [HappyLlama1.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q5_K_M.gguf) | Q5_K_M | 5.34GB |
| [HappyLlama1.Q5_1.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q5_1.gguf) | Q5_1 | 5.65GB |
| [HappyLlama1.Q6_K.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q6_K.gguf) | Q6_K | 6.14GB |
| [HappyLlama1.Q8_0.gguf](https://huggingface.co/RichardErkhov/ehristoforu_-_HappyLlama1-gguf/blob/main/HappyLlama1.Q8_0.gguf) | Q8_0 | 7.95GB |
Original model description:
---
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- trl
- sft
base_model: voidful/Llama-3.2-8B-Instruct
model-index:
- name: HappyLlama1
results:
- task:
type: text-generation
name: Text Generation
dataset:
name: IFEval (0-Shot)
type: HuggingFaceH4/ifeval
args:
num_few_shot: 0
metrics:
- type: inst_level_strict_acc and prompt_level_strict_acc
value: 73.63
name: strict accuracy
source:
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ehristoforu/HappyLlama1
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: BBH (3-Shot)
type: BBH
args:
num_few_shot: 3
metrics:
- type: acc_norm
value: 28.5
name: normalized accuracy
source:
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ehristoforu/HappyLlama1
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: MATH Lvl 5 (4-Shot)
type: hendrycks/competition_math
args:
num_few_shot: 4
metrics:
- type: exact_match
value: 10.12
name: exact match
source:
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ehristoforu/HappyLlama1
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: GPQA (0-shot)
type: Idavidrein/gpqa
args:
num_few_shot: 0
metrics:
- type: acc_norm
value: 4.47
name: acc_norm
source:
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ehristoforu/HappyLlama1
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: MuSR (0-shot)
type: TAUR-Lab/MuSR
args:
num_few_shot: 0
metrics:
- type: acc_norm
value: 11.25
name: acc_norm
source:
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ehristoforu/HappyLlama1
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: MMLU-PRO (5-shot)
type: TIGER-Lab/MMLU-Pro
config: main
split: test
args:
num_few_shot: 5
metrics:
- type: acc
value: 28.28
name: accuracy
source:
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ehristoforu/HappyLlama1
name: Open LLM Leaderboard
---
# Uploaded model
- **Developed by:** ehristoforu
- **License:** apache-2.0
- **Finetuned from model :** voidful/Llama-3.2-8B-Instruct
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_ehristoforu__HappyLlama1)
| Metric |Value|
|-------------------|----:|
|Avg. |26.04|
|IFEval (0-Shot) |73.63|
|BBH (3-Shot) |28.50|
|MATH Lvl 5 (4-Shot)|10.12|
|GPQA (0-shot) | 4.47|
|MuSR (0-shot) |11.25|
|MMLU-PRO (5-shot) |28.28|
|
Fizzarolli/l3-8b-kto-ckpt100
|
Fizzarolli
| 2025-06-08T02:58:42Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged",
"base_model:adapter:allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged",
"region:us"
] | null | 2025-06-08T02:58:36Z |
---
base_model: allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.15.2
|
gradientrouting-spar/2d_data_color_seed_11_seed_22_seed_33_20250608_013016
|
gradientrouting-spar
| 2025-06-08T02:57:46Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"gemma2",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T02:54:58Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
Fizzarolli/l3-8b-kto-ckpt75
|
Fizzarolli
| 2025-06-08T02:57:06Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged",
"base_model:adapter:allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged",
"region:us"
] | null | 2025-06-08T02:57:01Z |
---
base_model: allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.15.2
|
Fizzarolli/l3-8b-kto-ckpt50
|
Fizzarolli
| 2025-06-08T02:56:39Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged",
"base_model:adapter:allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged",
"region:us"
] | null | 2025-06-08T02:56:32Z |
---
base_model: allura-forge/l3-8b-deisgnaitnit-checkpriont-ep2-myorged
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.15.2
|
mlx-community/Big-Alice-28B-v1-4bit
|
mlx-community
| 2025-06-08T02:53:03Z | 0 | 0 |
mlx
|
[
"mlx",
"safetensors",
"mistral",
"text-generation",
"conversational",
"base_model:TheDrummer/Big-Alice-28B-v1",
"base_model:quantized:TheDrummer/Big-Alice-28B-v1",
"license:mit",
"4-bit",
"region:us"
] |
text-generation
| 2025-06-08T02:49:23Z |
---
base_model: TheDrummer/Big-Alice-28B-v1
license: mit
pipeline_tag: text-generation
library_name: mlx
tags:
- mlx
---
# mlx-community/Big-Alice-28B-v1-4bit
This model [mlx-community/Big-Alice-28B-v1-4bit](https://huggingface.co/mlx-community/Big-Alice-28B-v1-4bit) was
converted to MLX format from [TheDrummer/Big-Alice-28B-v1](https://huggingface.co/TheDrummer/Big-Alice-28B-v1)
using mlx-lm version **0.25.0**.
## Use with mlx
```bash
pip install mlx-lm
```
```python
from mlx_lm import load, generate
model, tokenizer = load("mlx-community/Big-Alice-28B-v1-4bit")
prompt = "hello"
if tokenizer.chat_template is not None:
messages = [{"role": "user", "content": prompt}]
prompt = tokenizer.apply_chat_template(
messages, add_generation_prompt=True
)
response = generate(model, tokenizer, prompt=prompt, verbose=True)
```
|
RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf
|
RichardErkhov
| 2025-06-08T02:46:49Z | 0 | 0 | null |
[
"gguf",
"endpoints_compatible",
"region:us",
"conversational"
] | null | 2025-06-08T01:40:46Z |
Quantization made by Richard Erkhov.
[Github](https://github.com/RichardErkhov)
[Discord](https://discord.gg/pvy7H8DZMG)
[Request more models](https://github.com/RichardErkhov/quant_request)
mergekit-della_linear-dbwwdyo - GGUF
- Model creator: https://huggingface.co/mergekit-community/
- Original model: https://huggingface.co/mergekit-community/mergekit-della_linear-dbwwdyo/
| Name | Quant method | Size |
| ---- | ---- | ---- |
| [mergekit-della_linear-dbwwdyo.Q2_K.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q2_K.gguf) | Q2_K | 2.96GB |
| [mergekit-della_linear-dbwwdyo.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.IQ3_XS.gguf) | IQ3_XS | 3.28GB |
| [mergekit-della_linear-dbwwdyo.IQ3_S.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.IQ3_S.gguf) | IQ3_S | 3.43GB |
| [mergekit-della_linear-dbwwdyo.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q3_K_S.gguf) | Q3_K_S | 3.41GB |
| [mergekit-della_linear-dbwwdyo.IQ3_M.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.IQ3_M.gguf) | IQ3_M | 3.52GB |
| [mergekit-della_linear-dbwwdyo.Q3_K.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q3_K.gguf) | Q3_K | 3.74GB |
| [mergekit-della_linear-dbwwdyo.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q3_K_M.gguf) | Q3_K_M | 3.74GB |
| [mergekit-della_linear-dbwwdyo.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q3_K_L.gguf) | Q3_K_L | 4.03GB |
| [mergekit-della_linear-dbwwdyo.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.IQ4_XS.gguf) | IQ4_XS | 4.18GB |
| [mergekit-della_linear-dbwwdyo.Q4_0.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q4_0.gguf) | Q4_0 | 4.34GB |
| [mergekit-della_linear-dbwwdyo.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.IQ4_NL.gguf) | IQ4_NL | 4.38GB |
| [mergekit-della_linear-dbwwdyo.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q4_K_S.gguf) | Q4_K_S | 4.37GB |
| [mergekit-della_linear-dbwwdyo.Q4_K.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q4_K.gguf) | Q4_K | 4.58GB |
| [mergekit-della_linear-dbwwdyo.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q4_K_M.gguf) | Q4_K_M | 4.58GB |
| [mergekit-della_linear-dbwwdyo.Q4_1.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q4_1.gguf) | Q4_1 | 4.78GB |
| [mergekit-della_linear-dbwwdyo.Q5_0.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q5_0.gguf) | Q5_0 | 5.21GB |
| [mergekit-della_linear-dbwwdyo.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q5_K_S.gguf) | Q5_K_S | 5.21GB |
| [mergekit-della_linear-dbwwdyo.Q5_K.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q5_K.gguf) | Q5_K | 5.34GB |
| [mergekit-della_linear-dbwwdyo.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q5_K_M.gguf) | Q5_K_M | 5.34GB |
| [mergekit-della_linear-dbwwdyo.Q5_1.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q5_1.gguf) | Q5_1 | 5.65GB |
| [mergekit-della_linear-dbwwdyo.Q6_K.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q6_K.gguf) | Q6_K | 6.14GB |
| [mergekit-della_linear-dbwwdyo.Q8_0.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_mergekit-della_linear-dbwwdyo-gguf/blob/main/mergekit-della_linear-dbwwdyo.Q8_0.gguf) | Q8_0 | 7.95GB |
Original model description:
---
base_model:
- Skywork/Skywork-o1-Open-Llama-3.1-8B
- Undi95/Meta-Llama-3.1-8B-Claude
- mergekit-community/mergekit-della_linear-uogzotg
- ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.2
- Solshine/reflection-llama-3.1-8B
- Undi95/Llama3-Unholy-8B-OAS
- vicgalle/Humanish-Roleplay-Llama-3.1-8B
- ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.3
- ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1
- Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2
library_name: transformers
tags:
- mergekit
- merge
---
# merge
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the della_linear merge method using [Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2](https://huggingface.co/Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2) as a base.
### Models Merged
The following models were included in the merge:
* [Skywork/Skywork-o1-Open-Llama-3.1-8B](https://huggingface.co/Skywork/Skywork-o1-Open-Llama-3.1-8B)
* [Undi95/Meta-Llama-3.1-8B-Claude](https://huggingface.co/Undi95/Meta-Llama-3.1-8B-Claude)
* [mergekit-community/mergekit-della_linear-uogzotg](https://huggingface.co/mergekit-community/mergekit-della_linear-uogzotg)
* [ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.2](https://huggingface.co/ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.2)
* [Solshine/reflection-llama-3.1-8B](https://huggingface.co/Solshine/reflection-llama-3.1-8B)
* [Undi95/Llama3-Unholy-8B-OAS](https://huggingface.co/Undi95/Llama3-Unholy-8B-OAS)
* [vicgalle/Humanish-Roleplay-Llama-3.1-8B](https://huggingface.co/vicgalle/Humanish-Roleplay-Llama-3.1-8B)
* [ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.3](https://huggingface.co/ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.3)
* [ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1](https://huggingface.co/ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.3
parameters:
density: 0.8
weight: 0.6
- model: Solshine/reflection-llama-3.1-8B
parameters:
density: 0.5
weight: 0.2
- model: Skywork/Skywork-o1-Open-Llama-3.1-8B
parameters:
density: 0.5
weight: 0.2
- model: ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.2
parameters:
density: 0.8
weight: 0.6
- model: ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1
parameters:
density: 0.8
weight: 0.6
- model: Undi95/Llama3-Unholy-8B-OAS
parameters:
density: 0.5
weight: 0.5
- model: vicgalle/Humanish-Roleplay-Llama-3.1-8B
parameters:
density: 0.5
weight: 0.5
- model: Undi95/Meta-Llama-3.1-8B-Claude
parameters:
density: 0.5
weight: 0.2
- model: mergekit-community/mergekit-della_linear-uogzotg
parameters:
density: 0.5
weight: 0.2
merge_method: della_linear
base_model: Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2
parameters:
normalize: false
int8_mask: true
dtype: float16
```
|
mlx-community/Big-Alice-28B-v1-bf16
|
mlx-community
| 2025-06-08T02:35:53Z | 0 | 0 |
mlx
|
[
"mlx",
"safetensors",
"mistral",
"text-generation",
"conversational",
"base_model:TheDrummer/Big-Alice-28B-v1",
"base_model:finetune:TheDrummer/Big-Alice-28B-v1",
"license:mit",
"region:us"
] |
text-generation
| 2025-06-08T02:32:40Z |
---
base_model: TheDrummer/Big-Alice-28B-v1
license: mit
pipeline_tag: text-generation
tags:
- mlx
library_name: mlx
---
# mlx-community/Big-Alice-28B-v1-bf16
This model [mlx-community/Big-Alice-28B-v1-bf16](https://huggingface.co/mlx-community/Big-Alice-28B-v1-bf16) was
converted to MLX format from [TheDrummer/Big-Alice-28B-v1](https://huggingface.co/TheDrummer/Big-Alice-28B-v1)
using mlx-lm version **0.25.0**.
## Use with mlx
```bash
pip install mlx-lm
```
```python
from mlx_lm import load, generate
model, tokenizer = load("mlx-community/Big-Alice-28B-v1-bf16")
prompt = "hello"
if tokenizer.chat_template is not None:
messages = [{"role": "user", "content": prompt}]
prompt = tokenizer.apply_chat_template(
messages, add_generation_prompt=True
)
response = generate(model, tokenizer, prompt=prompt, verbose=True)
```
|
nmixx-fin/nmixx-kure
|
nmixx-fin
| 2025-06-08T02:34:31Z | 0 | 0 | null |
[
"safetensors",
"xlm-roberta",
"ko",
"dataset:nmixx-fin/NMIXX_train",
"base_model:nlpai-lab/KURE-v1",
"base_model:finetune:nlpai-lab/KURE-v1",
"license:mit",
"region:us"
] | null | 2025-06-06T11:56:39Z |
---
license: mit
datasets:
- nmixx-fin/NMIXX_train
language:
- ko
base_model:
- nlpai-lab/KURE-v1
---
# NMIXX-kure
This repository contains a Kure‐based Embedding model fine‐tuned with a triplet‐loss setup on the `nmixx-fin/NMIXX_train` dataset. It produces high‐quality sentence embeddings for Korean financial text, optimized for semantic similarity tasks in the finance domain.
---
## How to Use
```python
import torch
import torch.nn.functional as F
from torch import Tensor
from transformers import AutoTokenizer, AutoModel
def cls_pool(last_hidden_states: Tensor) -> Tensor:
# Pool the hidden state of the [CLS] token.
return last_hidden_states[:, 0]
# 1. Load model and tokenizer from the Hugging Face Hub
# "your-username/your-finetuned-kure-model" should be replaced with your model's path.
model_name = "your-username/your-finetuned-kure-model"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModel.from_pretrained(model_name)
device = "cuda" if torch.cuda.is_available() else "cpu"
model.to(device)
model.eval()
# 2. Prepare sentences with the instruction
# Use the same instruction that was used for fine-tuning.
instruction = "제시된 기준 문장과 의미가 가장 유사한 문장을 찾으세요."
sentences = [
'금융은 좋아',
'금융은 안좋아',
'금금금',
]
# Add instruction to each sentence
input_texts = [f"{instruction} {sentence}" for sentence in sentences]
# 3. Tokenize and generate embeddings
batch_dict = tokenizer(input_texts, max_length=512, padding=True, truncation=True, return_tensors='pt').to(device)
with torch.no_grad():
outputs = model(**batch_dict)
# Apply CLS Pooling and normalize embeddings
embeddings = cls_pool(outputs.last_hidden_state)
embeddings = F.normalize(embeddings, p=2, dim=1)
# The output is a tensor containing the embeddings for each sentence.
print("Embeddings Shape:", embeddings.shape)
# Expected Output:
# Embeddings Shape: torch.Size([3, 1024])
```
|
btliu/llama3.2-1b-distilled-Q4_K_M-GGUF
|
btliu
| 2025-06-08T02:27:34Z | 0 | 0 |
transformers
|
[
"transformers",
"gguf",
"llama-cpp",
"gguf-my-repo",
"base_model:btliu/llama3.2-1b-distilled",
"base_model:quantized:btliu/llama3.2-1b-distilled",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T02:27:28Z |
---
library_name: transformers
tags:
- llama-cpp
- gguf-my-repo
base_model: btliu/llama3.2-1b-distilled
---
# btliu/llama3.2-1b-distilled-Q4_K_M-GGUF
This model was converted to GGUF format from [`btliu/llama3.2-1b-distilled`](https://huggingface.co/btliu/llama3.2-1b-distilled) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
Refer to the [original model card](https://huggingface.co/btliu/llama3.2-1b-distilled) for more details on the model.
## Use with llama.cpp
Install llama.cpp through brew (works on Mac and Linux)
```bash
brew install llama.cpp
```
Invoke the llama.cpp server or the CLI.
### CLI:
```bash
llama-cli --hf-repo btliu/llama3.2-1b-distilled-Q4_K_M-GGUF --hf-file llama3.2-1b-distilled-q4_k_m.gguf -p "The meaning to life and the universe is"
```
### Server:
```bash
llama-server --hf-repo btliu/llama3.2-1b-distilled-Q4_K_M-GGUF --hf-file llama3.2-1b-distilled-q4_k_m.gguf -c 2048
```
Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
Step 1: Clone llama.cpp from GitHub.
```
git clone https://github.com/ggerganov/llama.cpp
```
Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
```
cd llama.cpp && LLAMA_CURL=1 make
```
Step 3: Run inference through the main binary.
```
./llama-cli --hf-repo btliu/llama3.2-1b-distilled-Q4_K_M-GGUF --hf-file llama3.2-1b-distilled-q4_k_m.gguf -p "The meaning to life and the universe is"
```
or
```
./llama-server --hf-repo btliu/llama3.2-1b-distilled-Q4_K_M-GGUF --hf-file llama3.2-1b-distilled-q4_k_m.gguf -c 2048
```
|
AlinaTsai/taide_Llama-3.1-TAIDE-LX-8B-Chat_symptom_3960_ecophs_12_20250608
|
AlinaTsai
| 2025-06-08T02:25:29Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"llama",
"trl",
"en",
"base_model:taide/Llama-3.1-TAIDE-LX-8B-Chat",
"base_model:finetune:taide/Llama-3.1-TAIDE-LX-8B-Chat",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T02:25:02Z |
---
base_model: taide/Llama-3.1-TAIDE-LX-8B-Chat
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- trl
license: apache-2.0
language:
- en
---
# Uploaded model
- **Developed by:** AlinaTsai
- **License:** apache-2.0
- **Finetuned from model :** taide/Llama-3.1-TAIDE-LX-8B-Chat
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
Kromtao/1430e7f2-9b0b-4fbc-9d55-99e248b3f93b
|
Kromtao
| 2025-06-08T02:21:03Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"qwen2",
"axolotl",
"generated_from_trainer",
"base_model:Qwen/Qwen2.5-0.5B",
"base_model:adapter:Qwen/Qwen2.5-0.5B",
"license:apache-2.0",
"region:us"
] | null | 2025-06-08T01:49:39Z |
---
library_name: peft
license: apache-2.0
base_model: Qwen/Qwen2.5-0.5B
tags:
- axolotl
- generated_from_trainer
model-index:
- name: 1430e7f2-9b0b-4fbc-9d55-99e248b3f93b
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
[<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
<details><summary>See axolotl config</summary>
axolotl version: `0.4.1`
```yaml
adapter: lora
base_model: Qwen/Qwen2.5-0.5B
bf16: auto
chat_template: llama3
dataset_prepared_path: null
datasets:
- data_files:
- 1d5f926293665f60_train_data.json
ds_type: json
format: custom
path: /workspace/input_data/1d5f926293665f60_train_data.json
type:
field_instruction: instruct
field_output: output
format: '{instruction}'
no_input_format: '{instruction}'
system_format: '{system}'
system_prompt: ''
debug: null
deepspeed: null
do_eval: true
eval_batch_size: 8
eval_max_new_tokens: 128
eval_steps: 800
evals_per_epoch: null
flash_attention: true
fp16: false
fsdp: null
fsdp_config: null
gradient_accumulation_steps: 4
gradient_checkpointing: true
group_by_length: true
hub_model_id: Kromtao/1430e7f2-9b0b-4fbc-9d55-99e248b3f93b
hub_repo: null
hub_strategy: checkpoint
hub_token: null
learning_rate: 0.0002
local_rank: null
logging_steps: 50
lora_alpha: 16
lora_dropout: 0.1
lora_fan_in_fan_out: false
lora_model_dir: null
lora_r: 8
lora_target_linear: true
lr_scheduler: cosine
max_grad_norm: 1.0
max_steps: 800
micro_batch_size: 8
mlflow_experiment_name: /tmp/1d5f926293665f60_train_data.json
model_type: AutoModelForCausalLM
num_epochs: 10
optimizer: adamw_torch_fused
output_dir: miner_id_24
pad_to_sequence_len: true
resume_from_checkpoint: null
s2_attention: false
sample_packing: false
save_steps: 200
saves_per_epoch: null
seed: 9104
sequence_len: 1024
strict: false
tf32: true
tokenizer_type: AutoTokenizer
train_on_inputs: false
trust_remote_code: true
val_set_size: 0.05
wandb_entity: null
wandb_mode: online
wandb_name: d90f55b5-d4df-4499-a700-0fbdf5658b98
wandb_project: kr04
wandb_run: your_name
wandb_runid: d90f55b5-d4df-4499-a700-0fbdf5658b98
warmup_steps: 100
weight_decay: 0.01
xformers_attention: true
```
</details><br>
# 1430e7f2-9b0b-4fbc-9d55-99e248b3f93b
This model is a fine-tuned version of [Qwen/Qwen2.5-0.5B](https://huggingface.co/Qwen/Qwen2.5-0.5B) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.9185
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 8
- eval_batch_size: 8
- seed: 9104
- gradient_accumulation_steps: 4
- total_train_batch_size: 32
- optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 100
- training_steps: 800
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:------:|:----:|:---------------:|
| No log | 0.0001 | 1 | 2.4295 |
| 0.9234 | 0.0709 | 800 | 0.9185 |
### Framework versions
- PEFT 0.13.2
- Transformers 4.46.0
- Pytorch 2.5.0+cu124
- Datasets 3.0.1
- Tokenizers 0.20.1
|
misyaguziya/VRCT
|
misyaguziya
| 2025-06-08T02:19:22Z | 0 | 2 | null |
[
"region:us"
] | null | 2024-12-28T13:21:44Z |
github:https://github.com/misyaguziya/VRCT
|
metaheuristics/stepllm-theia-enames
|
metaheuristics
| 2025-06-08T02:18:14Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"unsloth",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T02:18:09Z |
---
library_name: transformers
tags:
- unsloth
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
madhueb/dpo-final-v3
|
madhueb
| 2025-06-08T02:17:26Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen3",
"text-generation",
"trl",
"dpo",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T02:16:13Z |
---
library_name: transformers
tags:
- trl
- dpo
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
stefandi/cog_behavior_synthetic_sft_v1_step_580
|
stefandi
| 2025-06-08T02:15:55Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T02:14:56Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
Satori-reasoning/Satori-SWE-RM-32B
|
Satori-reasoning
| 2025-06-08T02:14:46Z | 0 | 0 | null |
[
"safetensors",
"qwen2",
"license:apache-2.0",
"region:us"
] | null | 2025-05-29T04:22:39Z |
---
license: apache-2.0
---
|
underscore2/llama3-8b-bluesky-tpot-v2
|
underscore2
| 2025-06-08T02:12:17Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"llama",
"trl",
"en",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"base_model:finetune:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T02:12:11Z |
---
base_model: unsloth/llama-3-8b-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- trl
license: apache-2.0
language:
- en
---
# Uploaded model
- **Developed by:** underscore2
- **License:** apache-2.0
- **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
amanfor18/Natalia
|
amanfor18
| 2025-06-08T02:03:55Z | 0 | 0 |
diffusers
|
[
"diffusers",
"text-to-image",
"lora",
"template:diffusion-lora",
"base_model:black-forest-labs/FLUX.1-dev",
"base_model:adapter:black-forest-labs/FLUX.1-dev",
"license:unknown",
"region:us"
] |
text-to-image
| 2025-06-08T02:02:16Z |
---
tags:
- text-to-image
- lora
- diffusers
- template:diffusion-lora
widget:
- text: Natalia
output:
url: images/074073bd0fd6db9dffd4dd54de0ca6db_high.webp
base_model: black-forest-labs/FLUX.1-dev
instance_prompt: Natalia
license: unknown
---
# Natalia
<Gallery />
## Trigger words
You should use `Natalia` to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](/amanfor18/Natalia/tree/main) them in the Files & versions tab.
|
Leen20/gemma3-1b-r16-tr-ner-lr1e-4_2epochs_kartal
|
Leen20
| 2025-06-08T02:02:48Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"gemma3_text",
"trl",
"en",
"base_model:unsloth/gemma-3-1b-it-unsloth-bnb-4bit",
"base_model:finetune:unsloth/gemma-3-1b-it-unsloth-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T01:58:24Z |
---
base_model: unsloth/gemma-3-1b-it-unsloth-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- gemma3_text
- trl
license: apache-2.0
language:
- en
---
# Uploaded model
- **Developed by:** Leen20
- **License:** apache-2.0
- **Finetuned from model :** unsloth/gemma-3-1b-it-unsloth-bnb-4bit
This gemma3_text model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
uDauduna/q-FrozenLake-v1-4x4-noSlippery
|
uDauduna
| 2025-06-08T02:01:17Z | 0 | 0 | null |
[
"FrozenLake-v1-4x4-no_slippery",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] |
reinforcement-learning
| 2025-06-08T02:01:14Z |
---
tags:
- FrozenLake-v1-4x4-no_slippery
- q-learning
- reinforcement-learning
- custom-implementation
model-index:
- name: q-FrozenLake-v1-4x4-noSlippery
results:
- task:
type: reinforcement-learning
name: reinforcement-learning
dataset:
name: FrozenLake-v1-4x4-no_slippery
type: FrozenLake-v1-4x4-no_slippery
metrics:
- type: mean_reward
value: 1.00 +/- 0.00
name: mean_reward
verified: false
---
# **Q-Learning** Agent playing1 **FrozenLake-v1**
This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** .
## Usage
```python
model = load_from_hub(repo_id="uDauduna/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
|
madhueb/dpo-final-robust-v3
|
madhueb
| 2025-06-08T01:57:09Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen3",
"text-generation",
"trl",
"dpo",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-06-08T01:55:44Z |
---
library_name: transformers
tags:
- trl
- dpo
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
John6666/kodorail-v20-sdxl
|
John6666
| 2025-06-08T01:56:42Z | 0 | 0 |
diffusers
|
[
"diffusers",
"safetensors",
"text-to-image",
"stable-diffusion",
"stable-diffusion-xl",
"realistic",
"photorealistic",
"asian",
"Japanese",
"merge",
"noobai",
"illustrious",
"en",
"base_model:Laxhar/noobai-XL-1.1",
"base_model:merge:Laxhar/noobai-XL-1.1",
"base_model:OnomaAIResearch/Illustrious-XL-v1.0",
"base_model:merge:OnomaAIResearch/Illustrious-XL-v1.0",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"diffusers:StableDiffusionXLPipeline",
"region:us"
] |
text-to-image
| 2025-06-08T01:51:21Z |
---
license: other
license_name: faipl-1.0-sd
license_link: https://freedevproject.org/faipl-1.0-sd/
language:
- en
library_name: diffusers
pipeline_tag: text-to-image
tags:
- text-to-image
- stable-diffusion
- stable-diffusion-xl
- realistic
- photorealistic
- asian
- Japanese
- merge
- noobai
- illustrious
base_model:
- OnomaAIResearch/Illustrious-XL-v1.0
- Laxhar/noobai-XL-1.1
---
Original model is [here](https://civitai.com/models/1423866/kodorail?modelVersionId=1878260).
This model created by [Kodora](https://civitai.com/user/Kodora).
|
VIDraft/Gemma-3-R1984-1B
|
VIDraft
| 2025-06-08T01:55:44Z | 0 | 2 |
transformers
|
[
"transformers",
"safetensors",
"gemma3_text",
"text-generation",
"image-text-to-text",
"conversational",
"en",
"ko",
"ja",
"zh",
"es",
"ru",
"ar",
"hi",
"id",
"ml",
"fr",
"de",
"base_model:google/gemma-3-1b-it",
"base_model:finetune:google/gemma-3-1b-it",
"license:gemma",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
image-text-to-text
| 2025-06-08T00:14:56Z |
---
license: gemma
library_name: transformers
base_model: google/gemma-3-1b-it
language:
- en
- ko
- ja
- zh
- es
- ru
- ar
- hi
- id
- ml
- fr
- de
pipeline_tag: image-text-to-text
---
# Gemma3-R1984-1B
# Model Overview
Gemma3-R1984-1B is a robust Agentic AI platform built on Googls’s Gemma-3-4B model. It integrates state-of-the-art deep research via web search with multimodal file processing—including images, videos, and documents—and handles long contexts up to 8,000 tokens. Designed for local deployment on independent servers using NVIDIA L40s, L4, A-100(ZeroGPU) GPUs, it provides high security, prevents data leakage, and delivers uncensored responses.
# Key Features
Multimodal Processing:
Supports multiple file types such as images (PNG, JPG, JPEG, GIF, WEBP), videos (MP4), and documents (PDF, CSV, TXT).
Deep Research (Web Search):
Automatically extracts keywords from user queries and utilizes the SERPHouse API to retrieve up to 20 real-time search results. The model incorporates multiple sources by explicitly citing them in the response.
Long Context Handling:
Capable of processing inputs up to 8,000 tokens, ensuring comprehensive analysis of lengthy documents or conversations.
Robust Reasoning:
Employs extended chain-of-thought reasoning for systematic and accurate answer generation.
Secure Local Deployment:
Operates on independent local servers using NVIDIA L40s GPUs to maximize security and prevent information leakage.
**Experience the Power of Gemma3-R1984-1B**
- ✅ **Agentic AI Platform:** An autonomous system designed to make intelligent decisions and act independently.
- ✅ **Reasoning & Uncensored:** Delivers clear, accurate, and unfiltered responses by harnessing advanced reasoning capabilities.
- ✅ **Multimodal & VLM:** Seamlessly processes and interprets multiple input types—text, images, videos—empowering versatile applications.
- ✅ **Deep-Research & RAG:** Integrates state-of-the-art deep research and retrieval-augmented generation to provide comprehensive, real-time insights.
**Cutting-Edge Hardware for Maximum Security**
Gemma3-R1984-1B is engineered to operate on a dedicated **NVIDIA L40s GPU** within an independent local server environment. This robust setup not only guarantees optimal performance and rapid processing but also enhances security by isolating the model from external networks, effectively preventing information leakage. Whether handling sensitive data or complex queries, our platform ensures that your information remains secure and your AI interactions remain uncompromised.
# Use Cases
Fast-response conversational agents
Deep research and retrieval-augmented generation (RAG)
Document comparison and detailed analysis
Visual question answering from images and videos
Complex reasoning and research-based inquiries
# Supported File Formats
Images: PNG, JPG, JPEG, GIF, WEBP
Videos: MP4
Documents: PDF, CSV, TXT
# Model Details
Parameter Count: Approximately 1B parameters (estimated)
Context Window: Up to 8,000 tokens
Hugging Face Model Path: VIDraft/Gemma-3-R1984-1B
License: mit(Agentic AI) / gemma(gemma-3-1B)
# Installation and Setup
## Requirements
Ensure you have Python 3.8 or higher installed. The model relies on several libraries:
PyTorch (with bfloat16 support)
Transformers
Gradio
OpenCV (opencv-python)
Pillow (PIL)
PyPDF2
Pandas
Loguru
Requests
# Install dependencies using pip:
pip install torch transformers gradio opencv-python pillow PyPDF2 pandas loguru requests
# Environment Variables
Set the following environment variables before running the model:
## SERPHOUSE_API_KEY
Your SERPHouse API key for web search functionality.
Example:
export SERPHOUSE_API_KEY="your_api_key_here"
MODEL_ID
(Optional) The model identifier; default is VIDraft/Gemma-3-R1984-1B.
MAX_NUM_IMAGES
(Optional) Maximum number of images allowed per query (default is 5).
# Running the Model
Gemma3-R1984-1B comes with a Gradio-based multimodal chat interface. To run the model locally:
1. Clone the Repository:
Ensure you have the repository containing the model code.
2. Launch the Application:
Execute the main Python file:
python your_filename.py
This will start a local Gradio interface. Open the provided URL in your browser to interact with the model.
# Example Code: Server and Client Request
## Server Example
You can deploy the model server locally using the provided Gradio code. Make sure your server is accessible at your designated URL.
## Client Request Example
Below is an example of how to interact with the model using an HTTP API call:
```py
import requests
import json
# Replace with your server URL and token
url = "http://<your-server-url>:8000/v1/chat/completions"
headers = {
"Content-Type": "application/json",
"Authorization": "Bearer your_token_here"
}
# Construct the message payload
messages = [
{"role": "system", "content": "You are a powerful AI assistant."},
{"role": "user", "content": "Compare the contents of two PDF files."}
]
data = {
"model": "VIDraft/Gemma-3-R1984-1B",
"messages": messages,
"temperature": 0.15
}
# Send the POST request to the server
response = requests.post(url, headers=headers, data=json.dumps(data))
# Print the response from the model
print(response.json())
```
**Important Deployment Notice:**
For optimal performance, it is highly recommended to clone the repository using the following command. This model is designed to run on a server equipped with at least an NVIDIA L40s, L4, A100(ZeroGPU) GPU. The minimum VRAM requirement is 24GB, and VRAM usage may temporarily peak at approximately 82GB during processing.
```bash
git clone https://huggingface.co/spaces/VIDraft/Gemma-3-R1984-1B
|
IsmaelMousa/Qwen2.5-3B-Instruct-Books-19K
|
IsmaelMousa
| 2025-06-08T01:55:39Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"trl",
"sft",
"conversational",
"en",
"dataset:IsmaelMousa/books",
"base_model:Qwen/Qwen2.5-3B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-3B-Instruct",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-05-26T15:54:37Z |
---
library_name: transformers
tags:
- trl
- sft
license: apache-2.0
base_model: Qwen/Qwen2.5-3B-Instruct
datasets:
- IsmaelMousa/books
metrics:
- accuracy
- f1
- precision
- recall
- cohen_kappa
- rmse
model-index:
- name: Qwen2.5-3B-Instruct-Books-19K
results:
- task:
name: Text Generation
type: text-generation
dataset:
name: IsmaelMousa/books
type: IsmaelMousa/books
config: IsmaelMousa/books
split: train
args: IsmaelMousa/books
metrics:
- name: Accuracy
type: accuracy
value: 0.2200
- name: F1
type: f1
value: 0.1732
- name: Precision
type: precision
value: 0.1889
- name: Recall
type: recall
value: 0.2492
- name: Cohen Kappa
type: cohen_kappa
value: -0.0005
- name: RMSE
type: rmse
value: 1.7944
language:
- en
pipeline_tag: text-generation
---
# Qwen2.5-3B-Instruct-Books-19K
This model is a fine-tuned version of [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct) on the Books dataset for Essay Grading.
- **Workflow:** GitHub Repository: [https://github.com/IsmaelMousa/automatic-essay-grading](https://github.com/IsmaelMousa/automatic-essay-grading).
- **Base Model:** Qwen2.5-3B-Instruct: [https://doi.org/10.48550/arXiv.2412.15115](https://doi.org/10.48550/arXiv.2412.15115).
- **Fine-tuning Dataset:** Books-19K: [https://github.com/IsmaelMousa/Books/19K](https://github.com/IsmaelMousa/automatic-essay-grading/blob/main/data/books/clean/entries/train/200_entries.csv).
- **Task:** Automatic Essay Grading (Text Generation).
[](https://api.wandb.ai/links/ismael-amjad/783p4r3l)
## Dataset
The Books dataset is a synthetic collection of essay-style data points generated using public domain literature and
large language model prompting. The dataset comprises a total of 300 entries and is built from six classic books. Four
of these: *The Life of James Watt*, *The Life of Julius Caesar*, *The Moonstone*, and *North and South*; were used
during the training phase, while the remaining two: *The Life of Napoleon* and *Sense and Sensibility*; were held out
for benchmarking purposes. Each book contributed exactly 50 entries, leading to a structured split of 200 training
samples and 100 benchmark samples.
All entries were generated using Le Chat Mistral, a model developed by Mistral AI. A carefully crafted prompt was used
to ensure each generated entry included a question, a reference answer written by an expert, a student answer meant to
simulate a real-world response, a mark scheme outlining the grading criteria, a score between 1 and 4, and a rationale
explaining why the score was assigned. The prompt enforced strict quality control: no duplicate questions or answers
were allowed, all required fields had to be present, and the scoring range was strictly limited to valid values. The
final output was formatted as CSV files to maintain consistency and ensure compatibility with downstream processing.
For more details, the metadata can be accessed at: [metadata](https://github.com/IsmaelMousa/automatic-essay-grading/blob/main/data/books/metadata.py).
## Modeling
The modeling approach for this study was carefully designed to evaluate the performance of different large language models (LLMs) on the automated essay grading task. We selected the Qwen2.5 architecture to represent a range of model sizes: 0.5B, 1.5B, and 3B. Each model was instruction-tuned on the Books dataset in varying sizes, with hyperparameters optimized to balance computational efficiency and performance. The experiments were conducted on GPU-accelerated hardware, leveraging techniques such as gradient checkpointing, flash attention, and mixed-precision training to maximize resource utilization.
## Evaluation
The evaluation methodology employed both quantitative metrics and qualitative analysis. For quantitative assessment, we computed accuracy, precision, recall, F1 score, root mean squared error (RMSE), and Cohen's kappa score (CKS) for the scoring task, while using BERT-Score precision, recall, and F1 for rationale evaluation. On a held-out test set of 100 samples. Qualitative examination of models' outputs revealed cases where most of the models correctly identified key aspects of student answers but sometimes failed to properly align its scoring with the rubric criteria.
### Evaluation results for `score` and `rationale` outputs:
| **Aspect** | **F1** | **Precision** | **Recall** | **Accuracy** | **CKS** | **RMSE** |
|:----------:|:------:|:-------------:|:----------:|:------------:|:-------:|:--------:|
| Score | 0.1732 | 0.1889 | 0.2492 | 0.2200 | -0.0005 | 1.7944 |
| Rationale | 0.5449 | 0.5567 | 0.5371 | -- | -- | -- |
## Usage
Below is an example of how to use the model with the Hugging Face Transformers library:
```python
from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
import torch
checkpoint = "IsmaelMousa/Qwen2.5-3B-Instruct-Books-19K"
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
tokenizer = AutoTokenizer .from_pretrained(checkpoint)
model = AutoModelForCausalLM.from_pretrained(checkpoint)
assistant = pipeline("text-generation", tokenizer=tokenizer, model=model, device=device)
question = input("Question : ")
reference_answer = input("Reference Answer: ")
student_answer = input("Student Answer : ")
mark_scheme = input("Mark Scheme : ")
system_content = "You are a grading assistant. Evaluate student answers based on the mark scheme. Respond only in JSON format with keys 'score' (int) and 'rationale' (string)."
user_content = ("Provide both a score and a rationale by evaluating the student's answer strictly within the mark scheme range,"
" grading based on how well it meets the question's requirements by comparing the student answer to the reference answer.\n"
f"Question: {question}\n"
f"Reference Answer: {reference_answer}\n"
f"Student Answer: {student_answer}\n"
f"Mark Scheme: {mark_scheme}")
messages = [{"role": "system", "content": system_content}, {"role": "user", "content": user_content}]
inputs = tokenizer.apply_chat_template(messages, tokenize=False)
output = assistant(inputs, max_new_tokens=128, do_sample=False, return_full_text=False)[0]["generated_text"]
print(output)
```
### Frameworks
- `datasets-3.6.0`
- `torch-2.7.0`
- `transformers-4.51.3`
- `trl-0.17.0`
- `scikit-learn-1.6.1`
- `bert-score-0.3.13`
- `json-repair-0.46.0`
|
Jyqti/Jyoti
|
Jyqti
| 2025-06-08T01:53:56Z | 0 | 0 | null |
[
"license:other",
"region:us"
] | null | 2025-06-08T01:53:56Z |
---
license: other
license_name: license
license_link: LICENSE
---
|
kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5-gptqv2-4bit
|
kxdw2580
| 2025-06-08T01:50:32Z | 2 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen3",
"text-generation",
"GPTQ",
"conversational",
"zh",
"dataset:kxdw2580/catgirl-dataset",
"base_model:kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5",
"base_model:quantized:kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"gptq",
"region:us"
] |
text-generation
| 2025-06-07T06:36:21Z |
---
library_name: transformers
tags:
- GPTQ
license: apache-2.0
datasets:
- kxdw2580/catgirl-dataset
language:
- zh
base_model:
- kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5
---
This is the **GPTQ-v2 4-bit quantized version** of the model [`kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5`](https://huggingface.co/kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5).
During quantization, **layers 30–35 exhibited high loss**, which can be reviewed in the detailed **GPTQ-v2 quantization log** . Despite this anomaly, internal small-sample benchmarking indicates that the model's overall performance remains acceptable.
For more information about the base model, please refer to the original README.
---
# kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5
This new model series integrates updated datasets, base architectures, and fine-tuning methodologies. Based on **Qwen3**, it includes models with parameter counts of **8B** and **1.7B**.
Key updates focus on **daily conversations**, **creative generation**, **basic mathematics**, and **code generation**. Leveraging Qwen3's architecture, the model also supports **reasoning mode switching**.
🔍 **Fine-tuning records** are available on **SwanLab**:
1. [First Fine-tuning](https://swanlab.cn/@shadow01a/qwen-catgirl/runs/pcxfkgosz2e0cb430jk0a/overview)
2. [Second Fine-tuning](https://swanlab.cn/@shadow01a/qwen-catgirl/runs/iuou1xratkvbiv7jxw16k/overview)
3. [Third Fine-tuning](https://swanlab.cn/@shadow01a/qwen-catgirl/runs/9i2l4mc5qevmnlx2h51m0/overview)
---
## Evaluation
Due to the model's unique characteristics, we employed **human evaluation** for daily conversations and **DeepSeek-R1 scoring** (with reference answers provided in advance) for other domains to ensure character consistency and response validity.
### Key Improvements (vs. internal test models "0501" and "0531-test-all"):
- **Stronger detail-awareness** in casual dialogue
- **More coherent storytelling** in creative tasks
- **Deeper reasoning** during thinking mode
- **Better persona adherence** in long-form conversations without explicit prompts
- **Significant gains** in math/code domains (internal 20-question benchmark):
| Model | Math (Single Attempt) | Code (Single Attempt) |
|-------|-----------------------|-----------------------|
| Internal Test Model-0501 | 10% | 0% |
| DeepSeek-R1-0528-Qwen3-8B-Catgirl-0531-test-all | 30% | 20% |
| **DeepSeek-R1-0528-Qwen3-8B-Catgirl-v2.5** | **70%** | **60%** |
---
## Usage Guidelines
### Recommended Parameters:
- `temperature`: 0.7 (reasoning mode) / 0.6 (standard mode)
- `top_p`: 0.95
### Critical Notes:
- **Avoid** using model's reasoning chains as conversation context
- Inherits base model's tendency for lengthy reasoning in some cases – allow completion even if intermediate steps seem unusual
### English Mode:
Add this system prompt for English responses:
```
You are a catgirl. Please speak English.
```
---
## Acknowledgments
Special thanks to:
- **LLaMA-Factory** (fine-tuning framework)
- **Qwen Team** (base model provider)
- **DeepSeek Team** (DeepSeek-R1 evaluation support)
|
AliSerwat/medgemma-4b-it-sft-lora-crc100k
|
AliSerwat
| 2025-06-08T01:47:19Z | 0 | 0 |
transformers
|
[
"transformers",
"tensorboard",
"safetensors",
"generated_from_trainer",
"trl",
"sft",
"base_model:google/medgemma-4b-it",
"base_model:finetune:google/medgemma-4b-it",
"endpoints_compatible",
"region:us"
] | null | 2025-06-08T01:29:16Z |
---
base_model: google/medgemma-4b-it
library_name: transformers
model_name: medgemma-4b-it-sft-lora-crc100k
tags:
- generated_from_trainer
- trl
- sft
licence: license
---
# Model Card for medgemma-4b-it-sft-lora-crc100k
This model is a fine-tuned version of [google/medgemma-4b-it](https://huggingface.co/google/medgemma-4b-it).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="AliSerwat/medgemma-4b-it-sft-lora-crc100k", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
This model was trained with SFT.
### Framework versions
- TRL: 0.18.1
- Transformers: 4.52.4
- Pytorch: 2.7.0+cu128
- Datasets: 3.6.0
- Tokenizers: 0.21.1
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
FrenzyBiscuit/The-Omega-Directive-M-12B-Unslop-v2.0
|
FrenzyBiscuit
| 2025-06-08T01:44:35Z | 0 | 0 | null |
[
"safetensors",
"mistral",
"base_model:ReadyArt/The-Omega-Directive-M-12B-Unslop-v2.0",
"base_model:quantized:ReadyArt/The-Omega-Directive-M-12B-Unslop-v2.0",
"4-bit",
"awq",
"region:us"
] | null | 2025-06-08T01:41:06Z |
---
base_model: ReadyArt/The-Omega-Directive-M-12B-Unslop-v2.0
base_model_relation: quantized
quantized_by: FrenzyBiscuit
---
AWQ quant by FrenzyBiscuit.
Model was quantized down to INT4 using GEMM kernels, with zero-point quantization and a group size of 64.
I have not tested this quant.
<img src="./frenzy.png"></img>
|
kz919/DeepSeek-R1-Distill-Qwen-1.5B-GRPO
|
kz919
| 2025-06-08T01:42:36Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"open-r1",
"trl",
"grpo",
"conversational",
"dataset:open-r1/OpenR1-Math-220k",
"arxiv:2402.03300",
"base_model:deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
"base_model:finetune:deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-05-25T22:00:23Z |
---
base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
datasets: open-r1/OpenR1-Math-220k
library_name: transformers
model_name: DeepSeek-R1-Distill-Qwen-1.5B-GRPO
tags:
- generated_from_trainer
- open-r1
- trl
- grpo
licence: license
---
# Model Card for DeepSeek-R1-Distill-Qwen-1.5B-GRPO
This model is a fine-tuned version of [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B) on the [open-r1/OpenR1-Math-220k](https://huggingface.co/datasets/open-r1/OpenR1-Math-220k) dataset.
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="kz919/DeepSeek-R1-Distill-Qwen-1.5B-GRPO", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/kl2/c_optim_rl/runs/ka61pv7n)
This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
### Framework versions
- TRL: 0.18.0.dev0
- Transformers: 4.52.0.dev0
- Pytorch: 2.6.0
- Datasets: 3.6.0
- Tokenizers: 0.21.1
## Citations
Cite GRPO as:
```bibtex
@article{zhihong2024deepseekmath,
title = {{DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models}},
author = {Zhihong Shao and Peiyi Wang and Qihao Zhu and Runxin Xu and Junxiao Song and Mingchuan Zhang and Y. K. Li and Y. Wu and Daya Guo},
year = 2024,
eprint = {arXiv:2402.03300},
}
```
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
keita-origin/Ka-1.0
|
keita-origin
| 2025-06-08T01:41:32Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"gpt2",
"arxiv:1910.09700",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | 2025-06-07T14:32:22Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
dvalishvili-vs-sean-omalley-2-live-video/ufc.316.live.stream.tv
|
dvalishvili-vs-sean-omalley-2-live-video
| 2025-06-08T01:41:16Z | 0 | 0 | null |
[
"region:us"
] | null | 2025-06-08T01:40:23Z |
<a rel="nofollow" href="https://tinyurl.com/3tfpxeub?ufc.tv">🌐 𝖢𝖫𝖨𝖢𝖪 𝖧𝖤𝖱𝖤 🟢==►► UFC 316 LIVE Free</a>
<a rel="nofollow" href="https://tinyurl.com/3tfpxeub?ufc.tv">🔴 CLICK HERE 🌐==►► UFC 316 Live Now)</a>
<a rel="nofollow" href="https://tinyurl.com/3tfpxeub?ufc.tv"><img src="https://i.postimg.cc/qvPp49Sm/ythngythg.gif" alt="fsd"></a>
|
RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf
|
RichardErkhov
| 2025-06-08T01:40:01Z | 0 | 0 | null |
[
"gguf",
"endpoints_compatible",
"region:us",
"conversational"
] | null | 2025-06-08T00:31:41Z |
Quantization made by Richard Erkhov.
[Github](https://github.com/RichardErkhov)
[Discord](https://discord.gg/pvy7H8DZMG)
[Request more models](https://github.com/RichardErkhov/quant_request)
final_test_2_original_recipe - GGUF
- Model creator: https://huggingface.co/mergekit-community/
- Original model: https://huggingface.co/mergekit-community/final_test_2_original_recipe/
| Name | Quant method | Size |
| ---- | ---- | ---- |
| [final_test_2_original_recipe.Q2_K.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q2_K.gguf) | Q2_K | 2.96GB |
| [final_test_2_original_recipe.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.IQ3_XS.gguf) | IQ3_XS | 3.28GB |
| [final_test_2_original_recipe.IQ3_S.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.IQ3_S.gguf) | IQ3_S | 3.43GB |
| [final_test_2_original_recipe.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q3_K_S.gguf) | Q3_K_S | 3.41GB |
| [final_test_2_original_recipe.IQ3_M.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.IQ3_M.gguf) | IQ3_M | 3.52GB |
| [final_test_2_original_recipe.Q3_K.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q3_K.gguf) | Q3_K | 3.74GB |
| [final_test_2_original_recipe.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q3_K_M.gguf) | Q3_K_M | 3.74GB |
| [final_test_2_original_recipe.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q3_K_L.gguf) | Q3_K_L | 4.03GB |
| [final_test_2_original_recipe.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.IQ4_XS.gguf) | IQ4_XS | 4.18GB |
| [final_test_2_original_recipe.Q4_0.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q4_0.gguf) | Q4_0 | 4.34GB |
| [final_test_2_original_recipe.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.IQ4_NL.gguf) | IQ4_NL | 4.38GB |
| [final_test_2_original_recipe.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q4_K_S.gguf) | Q4_K_S | 4.37GB |
| [final_test_2_original_recipe.Q4_K.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q4_K.gguf) | Q4_K | 4.58GB |
| [final_test_2_original_recipe.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q4_K_M.gguf) | Q4_K_M | 4.58GB |
| [final_test_2_original_recipe.Q4_1.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q4_1.gguf) | Q4_1 | 4.78GB |
| [final_test_2_original_recipe.Q5_0.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q5_0.gguf) | Q5_0 | 5.21GB |
| [final_test_2_original_recipe.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q5_K_S.gguf) | Q5_K_S | 5.21GB |
| [final_test_2_original_recipe.Q5_K.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q5_K.gguf) | Q5_K | 5.34GB |
| [final_test_2_original_recipe.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q5_K_M.gguf) | Q5_K_M | 5.34GB |
| [final_test_2_original_recipe.Q5_1.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q5_1.gguf) | Q5_1 | 5.65GB |
| [final_test_2_original_recipe.Q6_K.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q6_K.gguf) | Q6_K | 6.14GB |
| [final_test_2_original_recipe.Q8_0.gguf](https://huggingface.co/RichardErkhov/mergekit-community_-_final_test_2_original_recipe-gguf/blob/main/final_test_2_original_recipe.Q8_0.gguf) | Q8_0 | 7.95GB |
Original model description:
---
base_model:
- Undi95/Llama3-Unholy-8B-OAS
- ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.2
- Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2
- Undi95/Meta-Llama-3.1-8B-Claude
- vicgalle/Humanish-Roleplay-Llama-3.1-8B
- mergekit-community/mergekit-della_linear-uogzotg
- ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.3
- Skywork/Skywork-o1-Open-Llama-3.1-8B
- Solshine/reflection-llama-3.1-8B
- ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1
library_name: transformers
tags:
- mergekit
- merge
---
# merge
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the della_linear merge method using [Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2](https://huggingface.co/Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2) as a base.
### Models Merged
The following models were included in the merge:
* [Undi95/Llama3-Unholy-8B-OAS](https://huggingface.co/Undi95/Llama3-Unholy-8B-OAS)
* [ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.2](https://huggingface.co/ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.2)
* [Undi95/Meta-Llama-3.1-8B-Claude](https://huggingface.co/Undi95/Meta-Llama-3.1-8B-Claude)
* [vicgalle/Humanish-Roleplay-Llama-3.1-8B](https://huggingface.co/vicgalle/Humanish-Roleplay-Llama-3.1-8B)
* [mergekit-community/mergekit-della_linear-uogzotg](https://huggingface.co/mergekit-community/mergekit-della_linear-uogzotg)
* [ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.3](https://huggingface.co/ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.3)
* [Skywork/Skywork-o1-Open-Llama-3.1-8B](https://huggingface.co/Skywork/Skywork-o1-Open-Llama-3.1-8B)
* [Solshine/reflection-llama-3.1-8B](https://huggingface.co/Solshine/reflection-llama-3.1-8B)
* [ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1](https://huggingface.co/ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.3
parameters:
density: 0.5
weight: 0.6
- model: Solshine/reflection-llama-3.1-8B
parameters:
density: 0.5
weight: 0.6
- model: Skywork/Skywork-o1-Open-Llama-3.1-8B
parameters:
density: 0.5
weight: 0.2
- model: ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.2
parameters:
density: 0.8
weight: 0.6
- model: ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1
parameters:
density: 0.8
weight: 0.6
- model: Undi95/Llama3-Unholy-8B-OAS
parameters:
density: 0.5
weight: 0.5
- model: vicgalle/Humanish-Roleplay-Llama-3.1-8B
parameters:
density: 0.5
weight: 0.5
- model: Undi95/Meta-Llama-3.1-8B-Claude
parameters:
density: 0.5
weight: 0.2
- model: mergekit-community/mergekit-della_linear-uogzotg
parameters:
density: 0.5
weight: 0.2
merge_method: della_linear
base_model: Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2
parameters:
normalize: false
int8_mask: true
dtype: float16
```
|
IsmaelMousa/Qwen2.5-0.5B-Instruct-EngSaf-835K
|
IsmaelMousa
| 2025-06-08T01:37:09Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"trl",
"sft",
"conversational",
"en",
"dataset:EngSAF",
"arxiv:2407.12818",
"base_model:Qwen/Qwen2.5-0.5B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-0.5B-Instruct",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-05-24T19:46:28Z |
---
library_name: transformers
tags:
- trl
- sft
license: apache-2.0
base_model: Qwen/Qwen2.5-0.5B-Instruct
datasets:
- EngSAF
metrics:
- accuracy
- f1
- precision
- recall
- cohen_kappa
- rmse
model-index:
- name: Qwen2.5-0.5B-Instruct-EngSAF-835K
results:
- task:
name: Text Generation
type: text-generation
dataset:
name: EngSAF
type: EngSAF
config: EngSAF
split: train
args: EngSAF
metrics:
- name: Accuracy
type: accuracy
value: 0.3600
- name: F1
type: f1
value: 0.3246
- name: Precision
type: precision
value: 0.3211
- name: Recall
type: recall
value: 0.4101
- name: Cohen Kappa
type: cohen_kappa
value: 0.0828
- name: RMSE
type: rmse
value: 1.0863
language:
- en
pipeline_tag: text-generation
---
# Qwen2.5-0.5B-Instruct-EngSaf-835K
This model is a fine-tuned version of [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct) on the EngSAF dataset for Essay Grading.
- **Workflow:** GitHub Repository: [https://github.com/IsmaelMousa/automatic-essay-grading](https://github.com/IsmaelMousa/automatic-essay-grading).
- **Base Model:** Qwen2.5-0.5B-Instruct: [https://doi.org/10.48550/arXiv.2412.15115](https://doi.org/10.48550/arXiv.2412.15115).
- **Fine-tuning Dataset:** EngSAF-835K: [https://github.com/IsmaelMousa/EngSAF/835K](https://github.com/IsmaelMousa/automatic-essay-grading/blob/main/data/engsaf/clean/train/4735_entries.csv).
- **Task:** Automatic Essay Grading (Text Generation).
[](https://api.wandb.ai/links/ismael-amjad/783p4r3l)
## Dataset
The EngSAF dataset, in its raw and unprocessed form, consists of approximately 5,800 short-answer responses collected
from real-life engineering examinations administered at a reputed academic institute. These responses are spread across
119 unique questions drawn from a wide range of engineering disciplines, making the dataset both diverse and
domain-specific. Each data point includes a student’s answer and an associated human-annotated score, serving as a
benchmark for evaluating automated grading models.
The dataset is divided into three primary subsets: 70% is allocated for training, 16% is reserved for evaluation on
unseen answers (UA), and 14% is dedicated to evaluating performance on entirely new questions (UQ). At this stage, it is
important to note that the dataset is considered in its original state; no preprocessing, transformation, or filtering
has yet been applied. All subsequent improvements and refinements to the data will be described in later sections.
This dataset is known as EngSAF version 1.0 and was introduced in the paper titled *"I understand why I got this grade":
Automatic Short Answer Grading (ASAG) with Feedback*, authored by Aggarwal et al., and set to appear in the proceedings
of AIED 2025. The dataset is released strictly for academic and research purposes; any commercial use or redistribution
without explicit permission is prohibited. Researchers are also urged to avoid publicly disclosing any sensitive content
that may be contained in the dataset.
For more details, the paper can be accessed at: [https://arxiv.org/abs/2407.12818](https://arxiv.org/abs/2407.12818).
## Modeling
The modeling approach for this study was carefully designed to evaluate the performance of different large language models (LLMs) on the automated essay grading task. We selected the Qwen2.5 architecture to represent a range of model sizes: 0.5B, 1.5B, and 3B. Each model was instruction-tuned on the EngSAF dataset in varying sizes, with hyperparameters optimized to balance computational efficiency and performance. The experiments were conducted on GPU-accelerated hardware, leveraging techniques such as gradient checkpointing, flash attention, and mixed-precision training to maximize resource utilization.
## Evaluation
The evaluation methodology employed both quantitative metrics and qualitative analysis. For quantitative assessment, we computed accuracy, precision, recall, F1 score, root mean squared error (RMSE), and Cohen's kappa score (CKS) for the scoring task, while using BERT-Score precision, recall, and F1 for rationale evaluation. On a held-out test set of 100 samples. Qualitative examination of models' outputs revealed cases where most of the models correctly identified key aspects of student answers but sometimes failed to properly align its scoring with the rubric criteria.
### Evaluation results for `score` and `rationale` outputs:
| **Aspect** | **F1** | **Precision** | **Recall** | **Accuracy** | **CKS** | **RMSE** |
|:----------:|:------:|:-------------:|:----------:|:------------:|:-------:|:--------:|
| Score | 0.3246 | 0.3211 | 0.4101 | 0.3600 | 0.0828 | 1.0863 |
| Rationale | 0.6153 | 0.6112 | 0.6234 | -- | -- | -- |
## Usage
Below is an example of how to use the model with the Hugging Face Transformers library:
```python
from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
import torch
checkpoint = "IsmaelMousa/Qwen2.5-0.5B-Instruct-EngSaf-835K"
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
tokenizer = AutoTokenizer .from_pretrained(checkpoint)
model = AutoModelForCausalLM.from_pretrained(checkpoint)
assistant = pipeline("text-generation", tokenizer=tokenizer, model=model, device=device)
question = input("Question : ")
reference_answer = input("Reference Answer: ")
student_answer = input("Student Answer : ")
mark_scheme = input("Mark Scheme : ")
system_content = "You are a grading assistant. Evaluate student answers based on the mark scheme. Respond only in JSON format with keys 'score' (int) and 'rationale' (string)."
user_content = ("Provide both a score and a rationale by evaluating the student's answer strictly within the mark scheme range,"
" grading based on how well it meets the question's requirements by comparing the student answer to the reference answer.\n"
f"Question: {question}\n"
f"Reference Answer: {reference_answer}\n"
f"Student Answer: {student_answer}\n"
f"Mark Scheme: {mark_scheme}")
messages = [{"role": "system", "content": system_content}, {"role": "user", "content": user_content}]
inputs = tokenizer.apply_chat_template(messages, tokenize=False)
output = assistant(inputs, max_new_tokens=128, do_sample=False, return_full_text=False)[0]["generated_text"]
print(output)
```
### Frameworks
- `datasets-3.6.0`
- `torch-2.7.0`
- `transformers-4.51.3`
- `trl-0.17.0`
- `scikit-learn-1.6.1`
- `bert-score-0.3.13`
- `json-repair-0.46.0`
|
Subsets and Splits
Filtered Qwen2.5 Distill Models
Identifies specific configurations of models by filtering cards that contain 'distill', 'qwen2.5', '7b' while excluding certain base models and incorrect model ID patterns, uncovering unique model variants.
Filtered Model Cards Count
Finds the count of entries with specific card details that include 'distill', 'qwen2.5', '7b' but exclude certain base models, revealing valuable insights about the dataset's content distribution.
Filtered Distill Qwen 7B Models
Filters for specific card entries containing 'distill', 'qwen', and '7b', excluding certain strings and patterns, to identify relevant model configurations.
Filtered Qwen-7b Model Cards
The query performs a detailed filtering based on specific keywords and excludes certain entries, which could be useful for identifying a specific subset of cards but does not provide deeper insights or trends.
Filtered Qwen 7B Model Cards
The query filters for specific terms related to "distilled" or "distill", "qwen", and "7b" in the 'card' column but excludes certain base models, providing a limited set of entries for further inspection.
Qwen 7B Distilled Models
The query provides a basic filtering of records to find specific card names that include keywords related to distilled Qwen 7b models, excluding a particular base model, which gives limited insight but helps in focusing on relevant entries.
Qwen 7B Distilled Model Cards
The query filters data based on specific keywords in the modelId and card fields, providing limited insight primarily useful for locating specific entries rather than revealing broad patterns or trends.
Qwen 7B Distilled Models
Finds all entries containing the terms 'distilled', 'qwen', and '7b' in a case-insensitive manner, providing a filtered set of records but without deeper analysis.
Distilled Qwen 7B Models
The query filters for specific model IDs containing 'distilled', 'qwen', and '7b', providing a basic retrieval of relevant entries but without deeper analysis or insight.
Filtered Model Cards with Distill Qwen2.
Filters and retrieves records containing specific keywords in the card description while excluding certain phrases, providing a basic count of relevant entries.
Filtered Model Cards with Distill Qwen 7
The query filters specific variations of card descriptions containing 'distill', 'qwen', and '7b' while excluding a particular base model, providing limited but specific data retrieval.
Distill Qwen 7B Model Cards
The query filters and retrieves rows where the 'card' column contains specific keywords ('distill', 'qwen', and '7b'), providing a basic filter result that can help in identifying specific entries.