modelId
string
author
string
last_modified
timestamp[us, tz=UTC]
downloads
int64
likes
int64
library_name
string
tags
list
pipeline_tag
string
createdAt
timestamp[us, tz=UTC]
card
string
2hpsatt/blockassist-bc-huge_deft_eagle_1755788948
2hpsatt
2025-08-21T15:10:27Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "huge deft eagle", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T15:10:23Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - huge deft eagle --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
gghfez/DeepSeek-V3.1-256x21B-BF16
gghfez
2025-08-21T15:09:54Z
0
1
null
[ "gguf", "base_model:deepseek-ai/DeepSeek-V3.1", "base_model:quantized:deepseek-ai/DeepSeek-V3.1", "region:us" ]
null
2025-08-21T11:15:40Z
--- base_model: - deepseek-ai/DeepSeek-V3.1 ---
mlx-community/shisa-v2-llama3.3-70b-mlx-bf16
mlx-community
2025-08-21T15:07:06Z
171
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "mlx", "conversational", "ja", "en", "dataset:shisa-ai/shisa-v2-sharegpt", "dataset:shisa-ai/deepseekv3-ultrafeedback-armorm-dpo", "base_model:shisa-ai/shisa-v2-llama3.3-70b", "base_model:finetune:shisa-ai/shisa-v2-llama3.3-70b", "license:llama3.3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-07-22T22:05:08Z
--- library_name: transformers model_name: shisa-v2-llama3.3-70b license: llama3.3 datasets: - shisa-ai/shisa-v2-sharegpt - shisa-ai/deepseekv3-ultrafeedback-armorm-dpo language: - ja - en base_model: shisa-ai/shisa-v2-llama3.3-70b pipeline_tag: text-generation tags: - mlx --- # mlx-community/shisa-v2-llama3.3-70b-mlx-fp16 The Model [mlx-community/shisa-v2-llama3.3-70b-mlx-bf16](https://huggingface.co/mlx-community/shisa-v2-llama3.3-70b-mlx-bf16) was converted to MLX format from [shisa-ai/shisa-v2-llama3.3-70b](https://huggingface.co/shisa-ai/shisa-v2-llama3.3-70b) using mlx-lm version **0.22.3**. ## Use with mlx ```bash pip install mlx-lm ``` ```python from mlx_lm import load, generate model, tokenizer = load("bibproj/shisa-v2-llama3.3-70b-mlx-fp16") prompt="hello" if hasattr(tokenizer, "apply_chat_template") and tokenizer.chat_template is not None: messages = [{"role": "user", "content": prompt}] prompt = tokenizer.apply_chat_template( messages, tokenize=False, add_generation_prompt=True ) response = generate(model, tokenizer, prompt=prompt, verbose=True) ```
Vasya777/blockassist-bc-lumbering_enormous_sloth_1755788771
Vasya777
2025-08-21T15:06:51Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "lumbering enormous sloth", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T15:06:43Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - lumbering enormous sloth --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
djricci3/domricci-replicatedemo
djricci3
2025-08-21T15:06:37Z
0
0
diffusers
[ "diffusers", "flux", "lora", "replicate", "text-to-image", "en", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-08-21T14:38:52Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - replicate base_model: "black-forest-labs/FLUX.1-dev" pipeline_tag: text-to-image # widget: # - text: >- # prompt # output: # url: https://... instance_prompt: Dom --- # Domricci Replicatedemo <Gallery /> ## About this LoRA This is a [LoRA](https://replicate.com/docs/guides/working-with-loras) for the FLUX.1-dev text-to-image model. It can be used with diffusers or ComfyUI. It was trained on [Replicate](https://replicate.com/) using AI toolkit: https://replicate.com/ostris/flux-dev-lora-trainer/train ## Trigger words You should use `Dom` to trigger the image generation. ## Run this LoRA with an API using Replicate ```py import replicate input = { "prompt": "Dom", "lora_weights": "https://huggingface.co/djricci3/domricci-replicatedemo/resolve/main/lora.safetensors" } output = replicate.run( "black-forest-labs/flux-dev-lora", input=input ) for index, item in enumerate(output): with open(f"output_{index}.webp", "wb") as file: file.write(item.read()) ``` ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('djricci3/domricci-replicatedemo', weight_name='lora.safetensors') image = pipeline('Dom').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters) ## Training details - Steps: 2008 - Learning rate: 0.0004 - LoRA rank: 16 ## Contribute your own examples You can use the [community tab](https://huggingface.co/djricci3/domricci-replicatedemo/discussions) to add images that show off what you’ve made with this LoRA.
Tuminha/snow-predictor-basel
Tuminha
2025-08-21T15:06:28Z
0
0
null
[ "joblib", "region:us" ]
null
2025-08-21T14:47:26Z
--- title: Snow Predictor Basel emoji: 🌨️ colorFrom: blue colorTo: white sdk: gradio sdk_version: 3.50.2 app_file: app.py pinned: false --- # 🌨️ Snow Predictor Basel - My First ML Model! 🚀 Welcome to my first machine learning project! This repository contains a **7-day ahead snow prediction model** for Basel, Switzerland that I built from scratch during my Python learning journey. ## 🎯 What This Model Does **Predicts snow in Basel 7 days in advance** using weather data patterns. Perfect for planning weekend trips, outdoor activities, or just knowing when to bring your umbrella! ## 🏆 Model Performance After training on **25 years of Basel weather data**, here's how well it performs: - **🎯 Accuracy:** 77.4% - Overall prediction accuracy - **❄️ Recall:** 84.0% - Catches most snow events (prioritizes safety!) - **⚠️ Precision:** 16.4% - Some false alarms, but better than missing snow - **�� ROC AUC:** 89.4% - Excellent model discrimination ## �� Key Features - **⏰ 7-day ahead prediction** - Plan your week with confidence - **🌡️ 22 weather features** - Temperature trends, precipitation patterns, seasonal indicators - **🛡️ High recall design** - Built to catch snow events rather than avoid false alarms - **�� 25 years of data** - Trained on comprehensive Basel weather history (2000-2025) ## 🏗️ How I Built This ### **Data Collection & Processing** - **Source:** Meteostat API for real Basel weather data - **Location:** Basel, Switzerland (47.5584° N, 7.5733° E) - **Processing:** Handled missing values, temperature inconsistencies, and date gaps - **Features:** Engineered rolling weather patterns, seasonal indicators, and volatility measures ### **Model Architecture** - **Algorithm:** Logistic Regression (chosen for interpretability and reliability) - **Training:** 80% of data for training, 20% for testing - **Class Balancing:** Used balanced class weights to handle snow/no-snow imbalance - **Feature Scaling:** Standardized all features for optimal performance ### **Feature Engineering** The model uses sophisticated weather patterns: - **Temperature trends** over 7-day windows - **Precipitation accumulation** patterns - **Atmospheric pressure** changes - **Seasonal indicators** and day-of-year patterns - **Weather volatility** measures ## 🔧 How to Use This Model ### **Quick Start** ```python import joblib import numpy as np # Load the trained model model_data = joblib.load('snow_predictor.joblib') model = model_data['model'] scaler = model_data['scaler'] feature_names = model_data['feature_names'] # Prepare your weather data (must match the 22 features) weather_features = np.array([your_weather_data_here]) # Scale the features weather_features_scaled = scaler.transform(weather_features.reshape(1, -1)) # Make prediction snow_probability = model.predict_proba(weather_features_scaled)[0][1] will_snow = model.predict(weather_features_scaled)[0] print(f"❄️ Snow probability: {snow_probability:.1%}") print(f"🌨️ Will it snow? {'Yes' if will_snow else 'No'}") ``` ### **Required Features (in order)** Your weather data must include these 22 features: 1. `tavg` - Average temperature 2. `tmin` - Minimum temperature 3. `tmax` - Maximum temperature 4. `prcp` - Precipitation 5. `wspd` - Wind speed 6. `wpgt` - Wind gust 7. `pres` - Pressure 8. `temp_range` - Temperature range 9. `temp_below_freezing` - Below freezing indicator 10. `high_precipitation` - High precipitation indicator 11. `windy_day` - Windy day indicator 12. `month` - Month of year 13. `day_of_year` - Day of year 14. `is_winter_season` - Winter season indicator 15. `temp_trend_7d` - 7-day temperature trend 16. `temp_std_7d` - 7-day temperature standard deviation 17. `precip_sum_7d` - 7-day precipitation sum 18. `pressure_trend_7d` - 7-day pressure trend 19. `cold_days_7d` - 7-day cold days count 20. `temp_volatility` - Temperature volatility 21. `pressure_change` - Pressure change rate 22. `temp_drop_rate` - Temperature drop rate ## 🌍 Real-World Applications **Perfect for:** - **🏠 Personal planning** - Weekend trips, outdoor activities, daily commutes - **🏢 Business operations** - Logistics, event planning, supply chain management - **🌤️ Weather enthusiasts** - Understanding Basel's weather patterns - **📚 Students & researchers** - Learning about weather prediction and ML ## 🎓 My Learning Journey This project represents my transition from **Python beginner to machine learning practitioner**. I started with basic Python concepts and gradually built up to: - **Data collection and API integration** - **Data cleaning and feature engineering** - **Machine learning model development** - **Model evaluation and performance analysis** - **Deployment and sharing** ## ��️ Technical Details ### **Dependencies** - Python 3.8+ - scikit-learn - pandas - numpy - meteostat (for weather data) ### **Installation** ```bash # Clone the repository git clone https://github.com/Tuminha/snow-predictor-basel.git cd snow-predictor-basel # Install dependencies pip install -r requirements.txt # Load and use the model python -c "import joblib; model = joblib.load('snow_predictor.joblib'); print('Model loaded successfully!')" ``` ## 📊 Training Data Insights - **Total data points:** 9,278 days of weather data - **Date range:** January 2000 to August 2025 - **Data quality:** Cleaned and validated for temperature consistency - **Missing data:** Only 106 days (1.2%) - handled with forward-fill ## 🎯 Why This Model Works **The high recall (84%) means:** - **You'll rarely be caught unprepared** for snow - **Some false alarms** (better safe than sorry!) - **Perfect for planning** when snow is a possibility **The 77.4% accuracy means:** - **Beats many professional weather forecasts** - **Reliable for 7-day planning** - **Excellent for a first ML model!** ## �� Acknowledgements - **Meteostat API** for providing comprehensive weather data - **scikit-learn** for the machine learning framework - **The Python community** for excellent documentation and tutorials - **My learning journey** that made this project possible ## 📝 License This project is open source and available under the [MIT License](LICENSE). ## �� Let's Connect! **This is my first machine learning model, and I'm excited to share it with the world!** ### **Contact Information** - **Name:** Francisco Teixeira Barbosa - **Email:** cisco@periospot.com - **Personal Portfolio:** [https://franciscodds.framer.ai/](https://franciscodds.framer.ai/) - **GitHub:** [https://github.com/Tuminha](https://github.com/Tuminha) - **Twitter/X:** [@Cisco_research](https://x.com/Cisco_research) ### **Questions & Feedback** - **Found a bug?** Open an issue! - **Want to improve the model?** Submit a pull request! - **Just want to chat?** Reach out on Twitter or GitHub! ## �� What's Next? This is just the beginning! Future improvements could include: - **Web application** for easy snow checking - **Mobile app** for on-the-go predictions - **More weather locations** across Switzerland - **Advanced ML algorithms** (Random Forest, XGBoost, Neural Networks) --- **Happy snow predicting! ❄️��️** *Built with ❤️ during my Python learning journey*
LINK-monica-korowi-viral-video-Clip/New.full.videos.monica.korowi.Viral.Video.Official.Tutorial
LINK-monica-korowi-viral-video-Clip
2025-08-21T15:04:32Z
0
0
null
[ "region:us" ]
null
2025-08-21T15:04:23Z
<a href="https://tinyurl.com/ybtx5at9" rel="nofollow" data-target="animated-image.originalLink"><img src="https://static.wixstatic.com/media/b249f9_adac8f70fb3f45b88691696c77de18f3~mv2.gif" alt="WATCH Videos" data-canonical-src="https://i.imgur.com/dJHk4Zq.gif" style="max-width: 100%; display: inline-block;" data-target="animated-image.originalImage"></a>
sdasdsee/blockassist-bc-wise_jumping_orangutan_1755784863
sdasdsee
2025-08-21T15:03:30Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "wise jumping orangutan", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T15:03:14Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - wise jumping orangutan --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Clip-Uppal-Farm-Girl-Viral-Video-Original/Full.Uppal.Farm.Girl.Viral.Video.Original.Link.Official
Clip-Uppal-Farm-Girl-Viral-Video-Original
2025-08-21T15:01:00Z
0
0
null
[ "region:us" ]
null
2025-08-21T15:00:52Z
<a href="https://tinyurl.com/ybtx5at9" rel="nofollow" data-target="animated-image.originalLink"><img src="https://static.wixstatic.com/media/b249f9_adac8f70fb3f45b88691696c77de18f3~mv2.gif" alt="WATCH Videos" data-canonical-src="https://i.imgur.com/dJHk4Zq.gif" style="max-width: 100%; display: inline-block;" data-target="animated-image.originalImage"></a>
llencia/blockassist-bc-wiry_wise_hedgehog_1755788335
llencia
2025-08-21T14:59:20Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "wiry wise hedgehog", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:59:15Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - wiry wise hedgehog --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF
bartowski
2025-08-21T14:58:36Z
0
0
null
[ "gguf", "text-generation", "base_model:TheDrummer/Behemoth-R1-123B-v2", "base_model:quantized:TheDrummer/Behemoth-R1-123B-v2", "region:us" ]
text-generation
2025-08-21T07:57:07Z
--- quantized_by: bartowski pipeline_tag: text-generation base_model: TheDrummer/Behemoth-R1-123B-v2 base_model_relation: quantized --- ## Llamacpp imatrix Quantizations of Behemoth-R1-123B-v2 by TheDrummer Using <a href="https://github.com/ggml-org/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggml-org/llama.cpp/releases/tag/b6214">b6214</a> for quantization. Original model: https://huggingface.co/TheDrummer/Behemoth-R1-123B-v2 All quants made using imatrix option with dataset from [here](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8) combined with a subset of combined_all_small.parquet from Ed Addario [here](https://huggingface.co/datasets/eaddario/imatrix-calibration/blob/main/combined_all_small.parquet) Run them in [LM Studio](https://lmstudio.ai/) Run them directly with [llama.cpp](https://github.com/ggml-org/llama.cpp), or any other llama.cpp based project ## Prompt format No prompt format found, check original model page ## Download a file (not the whole branch) from below: | Filename | Quant type | File Size | Split | Description | | -------- | ---------- | --------- | ----- | ----------- | | [Behemoth-R1-123B-v2-Q8_0.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-Q8_0) | Q8_0 | 130.28GB | true | Extremely high quality, generally unneeded but max available quant. | | [Behemoth-R1-123B-v2-Q6_K.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-Q6_K) | Q6_K | 100.59GB | true | Very high quality, near perfect, *recommended*. | | [Behemoth-R1-123B-v2-Q5_K_M.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-Q5_K_M) | Q5_K_M | 86.49GB | true | High quality, *recommended*. | | [Behemoth-R1-123B-v2-Q5_K_S.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-Q5_K_S) | Q5_K_S | 84.36GB | true | High quality, *recommended*. | | [Behemoth-R1-123B-v2-Q4_1.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-Q4_1) | Q4_1 | 76.72GB | true | Legacy format, similar performance to Q4_K_S but with improved tokens/watt on Apple silicon. | | [Behemoth-R1-123B-v2-Q4_K_M.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-Q4_K_M) | Q4_K_M | 73.22GB | true | Good quality, default size for most use cases, *recommended*. | | [Behemoth-R1-123B-v2-Q4_K_S.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-Q4_K_S) | Q4_K_S | 69.57GB | true | Slightly lower quality with more space savings, *recommended*. | | [Behemoth-R1-123B-v2-Q4_0.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-Q4_0) | Q4_0 | 69.32GB | true | Legacy format, offers online repacking for ARM and AVX CPU inference. | | [Behemoth-R1-123B-v2-IQ4_NL.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-IQ4_NL) | IQ4_NL | 69.22GB | true | Similar to IQ4_XS, but slightly larger. Offers online repacking for ARM CPU inference. | | [Behemoth-R1-123B-v2-IQ4_XS.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-IQ4_XS) | IQ4_XS | 65.43GB | true | Decent quality, smaller than Q4_K_S with similar performance, *recommended*. | | [Behemoth-R1-123B-v2-Q3_K_XL.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-Q3_K_XL) | Q3_K_XL | 64.91GB | true | Uses Q8_0 for embed and output weights. Lower quality but usable, good for low RAM availability. | | [Behemoth-R1-123B-v2-Q3_K_L.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-Q3_K_L) | Q3_K_L | 64.55GB | true | Lower quality but usable, good for low RAM availability. | | [Behemoth-R1-123B-v2-Q3_K_M.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-Q3_K_M) | Q3_K_M | 59.10GB | true | Low quality. | | [Behemoth-R1-123B-v2-IQ3_M.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-IQ3_M) | IQ3_M | 55.28GB | true | Medium-low quality, new method with decent performance comparable to Q3_K_M. | | [Behemoth-R1-123B-v2-Q3_K_S.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-Q3_K_S) | Q3_K_S | 52.85GB | true | Low quality, not recommended. | | [Behemoth-R1-123B-v2-IQ3_XS.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/tree/main/TheDrummer_Behemoth-R1-123B-v2-IQ3_XS) | IQ3_XS | 50.14GB | true | Lower quality, new method with decent performance, slightly better than Q3_K_S. | | [Behemoth-R1-123B-v2-IQ3_XXS.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/blob/main/TheDrummer_Behemoth-R1-123B-v2-IQ3_XXS.gguf) | IQ3_XXS | 47.01GB | false | Lower quality, new method with decent performance, comparable to Q3 quants. | | [Behemoth-R1-123B-v2-Q2_K_L.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/blob/main/TheDrummer_Behemoth-R1-123B-v2-Q2_K_L.gguf) | Q2_K_L | 45.59GB | false | Uses Q8_0 for embed and output weights. Very low quality but surprisingly usable. | | [Behemoth-R1-123B-v2-Q2_K.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/blob/main/TheDrummer_Behemoth-R1-123B-v2-Q2_K.gguf) | Q2_K | 45.20GB | false | Very low quality but surprisingly usable. | | [Behemoth-R1-123B-v2-IQ2_M.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/blob/main/TheDrummer_Behemoth-R1-123B-v2-IQ2_M.gguf) | IQ2_M | 41.62GB | false | Relatively low quality, uses SOTA techniques to be surprisingly usable. | | [Behemoth-R1-123B-v2-IQ2_S.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/blob/main/TheDrummer_Behemoth-R1-123B-v2-IQ2_S.gguf) | IQ2_S | 38.38GB | false | Low quality, uses SOTA techniques to be usable. | | [Behemoth-R1-123B-v2-IQ2_XS.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/blob/main/TheDrummer_Behemoth-R1-123B-v2-IQ2_XS.gguf) | IQ2_XS | 36.08GB | false | Low quality, uses SOTA techniques to be usable. | | [Behemoth-R1-123B-v2-IQ2_XXS.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/blob/main/TheDrummer_Behemoth-R1-123B-v2-IQ2_XXS.gguf) | IQ2_XXS | 32.43GB | false | Very low quality, uses SOTA techniques to be usable. | | [Behemoth-R1-123B-v2-IQ1_M.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/blob/main/TheDrummer_Behemoth-R1-123B-v2-IQ1_M.gguf) | IQ1_M | 28.39GB | false | Extremely low quality, *not* recommended. | | [Behemoth-R1-123B-v2-IQ1_S.gguf](https://huggingface.co/bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF/blob/main/TheDrummer_Behemoth-R1-123B-v2-IQ1_S.gguf) | IQ1_S | 25.96GB | false | Extremely low quality, *not* recommended. | ## Embed/output weights Some of these quants (Q3_K_XL, Q4_K_L etc) are the standard quantization method with the embeddings and output weights quantized to Q8_0 instead of what they would normally default to. ## Downloading using huggingface-cli <details> <summary>Click to view download instructions</summary> First, make sure you have hugginface-cli installed: ``` pip install -U "huggingface_hub[cli]" ``` Then, you can target the specific file you want: ``` huggingface-cli download bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF --include "TheDrummer_Behemoth-R1-123B-v2-Q4_K_M.gguf" --local-dir ./ ``` If the model is bigger than 50GB, it will have been split into multiple files. In order to download them all to a local folder, run: ``` huggingface-cli download bartowski/TheDrummer_Behemoth-R1-123B-v2-GGUF --include "TheDrummer_Behemoth-R1-123B-v2-Q8_0/*" --local-dir ./ ``` You can either specify a new local-dir (TheDrummer_Behemoth-R1-123B-v2-Q8_0) or download them all in place (./) </details> ## ARM/AVX information Previously, you would download Q4_0_4_4/4_8/8_8, and these would have their weights interleaved in memory in order to improve performance on ARM and AVX machines by loading up more data in one pass. Now, however, there is something called "online repacking" for weights. details in [this PR](https://github.com/ggml-org/llama.cpp/pull/9921). If you use Q4_0 and your hardware would benefit from repacking weights, it will do it automatically on the fly. As of llama.cpp build [b4282](https://github.com/ggml-org/llama.cpp/releases/tag/b4282) you will not be able to run the Q4_0_X_X files and will instead need to use Q4_0. Additionally, if you want to get slightly better quality for , you can use IQ4_NL thanks to [this PR](https://github.com/ggml-org/llama.cpp/pull/10541) which will also repack the weights for ARM, though only the 4_4 for now. The loading time may be slower but it will result in an overall speed incrase. <details> <summary>Click to view Q4_0_X_X information (deprecated</summary> I'm keeping this section to show the potential theoretical uplift in performance from using the Q4_0 with online repacking. <details> <summary>Click to view benchmarks on an AVX2 system (EPYC7702)</summary> | model | size | params | backend | threads | test | t/s | % (vs Q4_0) | | ------------------------------ | ---------: | ---------: | ---------- | ------: | ------------: | -------------------: |-------------: | | qwen2 3B Q4_0 | 1.70 GiB | 3.09 B | CPU | 64 | pp512 | 204.03 ± 1.03 | 100% | | qwen2 3B Q4_0 | 1.70 GiB | 3.09 B | CPU | 64 | pp1024 | 282.92 ± 0.19 | 100% | | qwen2 3B Q4_0 | 1.70 GiB | 3.09 B | CPU | 64 | pp2048 | 259.49 ± 0.44 | 100% | | qwen2 3B Q4_0 | 1.70 GiB | 3.09 B | CPU | 64 | tg128 | 39.12 ± 0.27 | 100% | | qwen2 3B Q4_0 | 1.70 GiB | 3.09 B | CPU | 64 | tg256 | 39.31 ± 0.69 | 100% | | qwen2 3B Q4_0 | 1.70 GiB | 3.09 B | CPU | 64 | tg512 | 40.52 ± 0.03 | 100% | | qwen2 3B Q4_K_M | 1.79 GiB | 3.09 B | CPU | 64 | pp512 | 301.02 ± 1.74 | 147% | | qwen2 3B Q4_K_M | 1.79 GiB | 3.09 B | CPU | 64 | pp1024 | 287.23 ± 0.20 | 101% | | qwen2 3B Q4_K_M | 1.79 GiB | 3.09 B | CPU | 64 | pp2048 | 262.77 ± 1.81 | 101% | | qwen2 3B Q4_K_M | 1.79 GiB | 3.09 B | CPU | 64 | tg128 | 18.80 ± 0.99 | 48% | | qwen2 3B Q4_K_M | 1.79 GiB | 3.09 B | CPU | 64 | tg256 | 24.46 ± 3.04 | 83% | | qwen2 3B Q4_K_M | 1.79 GiB | 3.09 B | CPU | 64 | tg512 | 36.32 ± 3.59 | 90% | | qwen2 3B Q4_0_8_8 | 1.69 GiB | 3.09 B | CPU | 64 | pp512 | 271.71 ± 3.53 | 133% | | qwen2 3B Q4_0_8_8 | 1.69 GiB | 3.09 B | CPU | 64 | pp1024 | 279.86 ± 45.63 | 100% | | qwen2 3B Q4_0_8_8 | 1.69 GiB | 3.09 B | CPU | 64 | pp2048 | 320.77 ± 5.00 | 124% | | qwen2 3B Q4_0_8_8 | 1.69 GiB | 3.09 B | CPU | 64 | tg128 | 43.51 ± 0.05 | 111% | | qwen2 3B Q4_0_8_8 | 1.69 GiB | 3.09 B | CPU | 64 | tg256 | 43.35 ± 0.09 | 110% | | qwen2 3B Q4_0_8_8 | 1.69 GiB | 3.09 B | CPU | 64 | tg512 | 42.60 ± 0.31 | 105% | Q4_0_8_8 offers a nice bump to prompt processing and a small bump to text generation </details> </details> ## Which file should I choose? <details> <summary>Click here for details</summary> A great write up with charts showing various performances is provided by Artefact2 [here](https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9) The first thing to figure out is how big a model you can run. To do this, you'll need to figure out how much RAM and/or VRAM you have. If you want your model running as FAST as possible, you'll want to fit the whole thing on your GPU's VRAM. Aim for a quant with a file size 1-2GB smaller than your GPU's total VRAM. If you want the absolute maximum quality, add both your system RAM and your GPU's VRAM together, then similarly grab a quant with a file size 1-2GB Smaller than that total. Next, you'll need to decide if you want to use an 'I-quant' or a 'K-quant'. If you don't want to think too much, grab one of the K-quants. These are in format 'QX_K_X', like Q5_K_M. If you want to get more into the weeds, you can check out this extremely useful feature chart: [llama.cpp feature matrix](https://github.com/ggml-org/llama.cpp/wiki/Feature-matrix) But basically, if you're aiming for below Q4, and you're running cuBLAS (Nvidia) or rocBLAS (AMD), you should look towards the I-quants. These are in format IQX_X, like IQ3_M. These are newer and offer better performance for their size. These I-quants can also be used on CPU, but will be slower than their K-quant equivalent, so speed vs performance is a tradeoff you'll have to decide. </details> ## Credits Thank you kalomaze and Dampf for assistance in creating the imatrix calibration dataset. Thank you ZeroWw for the inspiration to experiment with embed/output. Thank you to LM Studio for sponsoring my work. Want to support my work? Visit my ko-fi page here: https://ko-fi.com/bartowski
llencia/blockassist-bc-wiry_wise_hedgehog_1755788204
llencia
2025-08-21T14:57:12Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "wiry wise hedgehog", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:57:08Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - wiry wise hedgehog --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
unitova/blockassist-bc-zealous_sneaky_raven_1755786446
unitova
2025-08-21T14:55:25Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "zealous sneaky raven", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:55:21Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - zealous sneaky raven --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
lb8s/my-great-gpt2-review-model-OP
lb8s
2025-08-21T14:51:46Z
0
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "generated_from_trainer", "base_model:erwanf/gpt2-mini", "base_model:finetune:erwanf/gpt2-mini", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-08-21T13:20:48Z
--- library_name: transformers license: mit base_model: erwanf/gpt2-mini tags: - generated_from_trainer model-index: - name: my-great-gpt2-review-model-OP results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # my-great-gpt2-review-model-OP This model is a fine-tuned version of [erwanf/gpt2-mini](https://huggingface.co/erwanf/gpt2-mini) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 5.1346 - Model Preparation Time: 0.0047 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.003991 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 0.6 ### Training results | Training Loss | Epoch | Step | Validation Loss | Model Preparation Time | |:-------------:|:-----:|:----:|:---------------:|:----------------------:| | 5.1225 | 0.6 | 3051 | 5.1346 | 0.0047 | ### Framework versions - Transformers 4.55.3 - Pytorch 2.8.0+cu126 - Datasets 4.0.0 - Tokenizers 0.21.4
wangyichen25/Meta-Llama-3.1-8B-Instruct_epoch2_r16_alpha16_lr0.0001_CoT_ICD_v3_vllm_16bit
wangyichen25
2025-08-21T14:51:12Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2025-08-21T14:47:36Z
--- base_model: unsloth/meta-llama-3.1-8b-instruct-unsloth-bnb-4bit tags: - text-generation-inference - transformers - unsloth - llama license: apache-2.0 language: - en --- # Uploaded finetuned model - **Developed by:** wangyichen25 - **License:** apache-2.0 - **Finetuned from model :** unsloth/meta-llama-3.1-8b-instruct-unsloth-bnb-4bit This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
Dejiat/blockassist-bc-savage_unseen_bobcat_1755787833
Dejiat
2025-08-21T14:51:09Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "savage unseen bobcat", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:51:04Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - savage unseen bobcat --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
helmutsukocok/blockassist-bc-loud_scavenging_kangaroo_1755786237
helmutsukocok
2025-08-21T14:50:43Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "loud scavenging kangaroo", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:50:39Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - loud scavenging kangaroo --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
mveroe/Qwen2.5-1.5B_lightr1_3_EN_1024_1p0_0p0_1p0_sft
mveroe
2025-08-21T14:50:38Z
21
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "generated_from_trainer", "conversational", "base_model:Qwen/Qwen2.5-1.5B", "base_model:finetune:Qwen/Qwen2.5-1.5B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-08-14T14:43:38Z
--- library_name: transformers license: apache-2.0 base_model: Qwen/Qwen2.5-1.5B tags: - generated_from_trainer model-index: - name: Qwen2.5-1.5B_lightr1_3_EN_1024_1p0_0p0_1p0_sft results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Qwen2.5-1.5B_lightr1_3_EN_1024_1p0_0p0_1p0_sft This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B](https://huggingface.co/Qwen/Qwen2.5-1.5B) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 4 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - total_eval_batch_size: 32 - optimizer: Use OptimizerNames.ADAFACTOR and the args are: No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.55.2 - Pytorch 2.7.1+cu128 - Datasets 4.0.0 - Tokenizers 0.21.2
Vasya777/blockassist-bc-lumbering_enormous_sloth_1755787687
Vasya777
2025-08-21T14:48:48Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "lumbering enormous sloth", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:48:41Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - lumbering enormous sloth --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
2hpsatt/blockassist-bc-huge_deft_eagle_1755787638
2hpsatt
2025-08-21T14:48:45Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "huge deft eagle", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:48:40Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - huge deft eagle --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Ag2r/Merged_llama3.2_V1
Ag2r
2025-08-21T14:48:13Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-08-21T14:33:27Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
thanobidex/blockassist-bc-colorful_shiny_hare_1755786107
thanobidex
2025-08-21T14:48:09Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "colorful shiny hare", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:48:05Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - colorful shiny hare --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
WenFengg/pyarr_14l19_21_8
WenFengg
2025-08-21T14:47:30Z
0
0
null
[ "safetensors", "any-to-any", "omega", "omegalabs", "bittensor", "agi", "license:mit", "region:us" ]
any-to-any
2025-08-21T14:38:59Z
--- license: mit tags: - any-to-any - omega - omegalabs - bittensor - agi --- This is an Any-to-Any model checkpoint for the OMEGA Labs x Bittensor Any-to-Any subnet. Check out the [git repo](https://github.com/omegalabsinc/omegalabs-anytoany-bittensor) and find OMEGA on X: [@omegalabsai](https://x.com/omegalabsai).
Osrivers/hunyuanVideoSafetensors_visionCLIPLBF16.safetensors
Osrivers
2025-08-21T14:47:00Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2025-08-21T14:45:44Z
--- license: creativeml-openrail-m ---
gghfez/DeepSeek-V3.1-Base-IQ2_KS
gghfez
2025-08-21T14:46:44Z
0
0
null
[ "gguf", "gguf,", "ik_llama.cpp", "base_model:deepseek-ai/DeepSeek-V3.1-Base", "base_model:quantized:deepseek-ai/DeepSeek-V3.1-Base", "license:apache-2.0", "endpoints_compatible", "region:us", "imatrix", "conversational" ]
null
2025-08-21T08:09:27Z
--- base_model: - deepseek-ai/DeepSeek-V3.1-Base license: apache-2.0 tags: - gguf, - ik_llama.cpp --- # gghfez/DeepSeek-V3.1-Base-IQ2_KS ## This is the BASE model, not trained for conversations.
Dejiat/blockassist-bc-savage_unseen_bobcat_1755787457
Dejiat
2025-08-21T14:44:55Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "savage unseen bobcat", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:44:49Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - savage unseen bobcat --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
lautan/blockassist-bc-gentle_patterned_goat_1755785912
lautan
2025-08-21T14:44:34Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "gentle patterned goat", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:44:30Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - gentle patterned goat --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
vladimirwest/ml-finetuning-2
vladimirwest
2025-08-21T14:44:04Z
0
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "llama-factory", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-08-21T14:40:30Z
--- library_name: transformers tags: - llama-factory --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
manusiaperahu2012/blockassist-bc-roaring_long_tuna_1755785854
manusiaperahu2012
2025-08-21T14:43:55Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "roaring long tuna", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:43:51Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - roaring long tuna --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Vasya777/blockassist-bc-lumbering_enormous_sloth_1755787300
Vasya777
2025-08-21T14:42:20Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "lumbering enormous sloth", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:42:12Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - lumbering enormous sloth --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Dejiat/blockassist-bc-savage_unseen_bobcat_1755787250
Dejiat
2025-08-21T14:41:25Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "savage unseen bobcat", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:41:19Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - savage unseen bobcat --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
ReallyFloppyPenguin/shuka-1-incase-they-remove-it
ReallyFloppyPenguin
2025-08-21T14:41:16Z
0
0
transformers
[ "transformers", "safetensors", "shuka", "feature-extraction", "audio-text-to-text", "custom_code", "en", "hi", "license:llama3", "region:us" ]
audio-text-to-text
2025-08-21T14:41:16Z
--- library_name: transformers pipeline_tag: audio-text-to-text license: llama3 language: - en - hi --- `Shuka v1` is a language model which natively understands audio in Indic languages. It is an encoder-decoder model built by combining two models: - Our state-of-the-art, in-house, audio encoder: Saaras v1 - Meta’s Llama3-8B-Instruct as the decoder The encoder and decoder are connected by a small projector with ~60M parameters. During training, only the projector weights are finetuned while the rest of the network is frozen. Following our tradition of training models frugally, we train `Shuka v1` on less than 100 hours of audio. Though we only finetune the projector on English and Hindi data, the multilingual nature of our encoder makes `Shuka v1` perform well on zero-shot QA in other Indic languages as well. We have tested on the model on Bengali, English, Gujarati, Hindi, Kannada, Malayalam, Marathi, Oriya, Punjabi, Tamil, and Telugu. See what `Shuka v1` can do in this [demo video](https://www.youtube.com/watch?v=VgJhjCPbORs), and get started by using huggingface pipeline, as follows: ``` # install libraries # pip install transformers==4.41.2 peft==0.11.1 librosa==0.10.2 import transformers import librosa # load the model pipeline on gpu:0 pipe = transformers.pipeline(model='sarvamai/shuka_v1', trust_remote_code=True, device=0, torch_dtype='bfloat16') # get a sample audio # wget https://huggingface.co/sarvamai/shuka_v1/resolve/main/hi-question.webm audio, sr = librosa.load("./hi-question.webm", sr=16000) turns = [ {'role': 'system', 'content': 'Respond naturally and informatively.'}, {'role': 'user', 'content': '<|audio|>'} ] pipe({'audio': audio, 'turns': turns, 'sampling_rate': sr}, max_new_tokens=512) ``` For more details, please see our [blog](https://www.sarvam.ai/blogs/shuka-v1).
rourkerhotmail1/blockassist-bc-stalking_scruffy_walrus_1755785581
rourkerhotmail1
2025-08-21T14:40:53Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "stalking scruffy walrus", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:40:49Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - stalking scruffy walrus --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
jakehsv/blockassist-bc-flexible_waddling_peacock_1755785556
jakehsv
2025-08-21T14:40:09Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "flexible waddling peacock", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:39:58Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - flexible waddling peacock --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
liukevin666/blockassist-bc-yawning_striped_cassowary_1755787009
liukevin666
2025-08-21T14:40:04Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "yawning striped cassowary", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:38:06Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - yawning striped cassowary --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Ryandro/mt5-base-finetuned-1000data-Lp6
Ryandro
2025-08-21T14:40:02Z
0
0
transformers
[ "transformers", "safetensors", "mt5", "text2text-generation", "generated_from_trainer", "base_model:google/mt5-base", "base_model:finetune:google/mt5-base", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2025-08-21T06:43:02Z
--- library_name: transformers license: apache-2.0 base_model: google/mt5-base tags: - generated_from_trainer metrics: - rouge model-index: - name: mt5-base-finetuned-1000data-Lp6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mt5-base-finetuned-1000data-Lp6 This model is a fine-tuned version of [google/mt5-base](https://huggingface.co/google/mt5-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: nan - Rouge1: 0.5556 - Rouge2: 0.0 - Rougel: 0.5556 - Rougelsum: 0.5556 - Gen Len: 10.5 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-06 - train_batch_size: 5 - eval_batch_size: 5 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:| | No log | 1.0 | 18 | nan | 0.5556 | 0.0 | 0.5556 | 0.5556 | 10.5 | | No log | 2.0 | 36 | nan | 0.5556 | 0.0 | 0.5556 | 0.5556 | 10.5 | | No log | 3.0 | 54 | nan | 0.5556 | 0.0 | 0.5556 | 0.5556 | 10.5 | ### Framework versions - Transformers 4.55.2 - Pytorch 2.8.0+cu126 - Datasets 4.0.0 - Tokenizers 0.21.4
indoempatnol/blockassist-bc-fishy_wary_swan_1755785536
indoempatnol
2025-08-21T14:39:47Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "fishy wary swan", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:39:44Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - fishy wary swan --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
kojeklollipop/blockassist-bc-spotted_amphibious_stork_1755785578
kojeklollipop
2025-08-21T14:39:26Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "spotted amphibious stork", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:39:22Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - spotted amphibious stork --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
llencia/blockassist-bc-wiry_wise_hedgehog_1755787026
llencia
2025-08-21T14:37:32Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "wiry wise hedgehog", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:37:29Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - wiry wise hedgehog --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
quantumxnode/blockassist-bc-dormant_peckish_seahorse_1755785543
quantumxnode
2025-08-21T14:37:07Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "dormant peckish seahorse", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:37:04Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - dormant peckish seahorse --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
BinBashir/batch_4_int8
BinBashir
2025-08-21T14:36:54Z
0
0
transformers
[ "transformers", "safetensors", "distilbert", "text-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "8-bit", "bitsandbytes", "region:us" ]
text-classification
2025-08-21T14:36:37Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
tamewild/4b_v60_merged_e3
tamewild
2025-08-21T14:35:05Z
0
0
transformers
[ "transformers", "safetensors", "qwen3", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-08-21T14:32:47Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
llencia/blockassist-bc-wiry_wise_hedgehog_1755786832
llencia
2025-08-21T14:34:18Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "wiry wise hedgehog", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:34:13Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - wiry wise hedgehog --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
tamewild/4b_v60_merged_e5
tamewild
2025-08-21T14:31:57Z
0
0
transformers
[ "transformers", "safetensors", "qwen3", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-08-21T14:29:41Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Dejiat/blockassist-bc-savage_unseen_bobcat_1755786547
Dejiat
2025-08-21T14:29:47Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "savage unseen bobcat", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:29:41Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - savage unseen bobcat --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
WenFengg/pyarr_14l18_21_8
WenFengg
2025-08-21T14:28:05Z
0
0
null
[ "safetensors", "any-to-any", "omega", "omegalabs", "bittensor", "agi", "license:mit", "region:us" ]
any-to-any
2025-08-21T14:23:05Z
--- license: mit tags: - any-to-any - omega - omegalabs - bittensor - agi --- This is an Any-to-Any model checkpoint for the OMEGA Labs x Bittensor Any-to-Any subnet. Check out the [git repo](https://github.com/omegalabsinc/omegalabs-anytoany-bittensor) and find OMEGA on X: [@omegalabsai](https://x.com/omegalabsai).
rk2357281/lora_model_en_bho
rk2357281
2025-08-21T14:27:48Z
0
0
transformers
[ "transformers", "safetensors", "text-generation-inference", "unsloth", "llama", "trl", "en", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2025-08-21T14:27:29Z
--- base_model: unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit tags: - text-generation-inference - transformers - unsloth - llama - trl license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** rk2357281 - **License:** apache-2.0 - **Finetuned from model :** unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
calegpedia/blockassist-bc-stealthy_slimy_rooster_1755784825
calegpedia
2025-08-21T14:27:45Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "stealthy slimy rooster", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:27:41Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - stealthy slimy rooster --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
hogensynoo/blockassist-bc-tiny_fierce_bee_1755786427
hogensynoo
2025-08-21T14:27:20Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "tiny fierce bee", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:27:08Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - tiny fierce bee --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
coelacanthxyz/blockassist-bc-finicky_thriving_grouse_1755784640
coelacanthxyz
2025-08-21T14:27:18Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "finicky thriving grouse", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:27:11Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - finicky thriving grouse --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Dejiat/blockassist-bc-savage_unseen_bobcat_1755786387
Dejiat
2025-08-21T14:27:04Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "savage unseen bobcat", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:26:57Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - savage unseen bobcat --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
llencia/blockassist-bc-wiry_wise_hedgehog_1755786390
llencia
2025-08-21T14:26:54Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "wiry wise hedgehog", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:26:50Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - wiry wise hedgehog --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
johngreendr1/3f6d258b-d775-4747-a416-0642aab9fc26
johngreendr1
2025-08-21T14:25:45Z
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Intel/neural-chat-7b-v3-3", "base_model:adapter:Intel/neural-chat-7b-v3-3", "region:us" ]
null
2025-08-21T12:14:08Z
--- base_model: Intel/neural-chat-7b-v3-3 library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.15.1
TobyLu/talk-to-me-app
TobyLu
2025-08-21T14:25:36Z
0
0
null
[ "license:apache-2.0", "region:us" ]
null
2025-08-21T14:25:36Z
--- license: apache-2.0 ---
llencia/blockassist-bc-wiry_wise_hedgehog_1755786249
llencia
2025-08-21T14:24:40Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "wiry wise hedgehog", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:24:32Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - wiry wise hedgehog --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Sanjay002/tinyllama-mental-health-finetuned
Sanjay002
2025-08-21T14:23:33Z
969
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "chatbot", "mental-health", "code", "text-generation-inference", "conversational", "en", "dataset:heliosbrahma/mental_health_chatbot_dataset", "base_model:TinyLlama/TinyLlama-1.1B-Chat-v1.0", "base_model:finetune:TinyLlama/TinyLlama-1.1B-Chat-v1.0", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2025-04-10T09:59:08Z
--- language: en pipeline_tag: text-generation license: mit tags: - chatbot - mental-health - code - text-generation-inference datasets: - heliosbrahma/mental_health_chatbot_dataset base_model: - TinyLlama/TinyLlama-1.1B-Chat-v1.0 library_name: transformers ---
rayhaan-beeharry/gemma-3-4b-it-Q4_K_M-GGUF
rayhaan-beeharry
2025-08-21T14:23:21Z
0
0
transformers
[ "transformers", "gguf", "llama-cpp", "gguf-my-repo", "image-text-to-text", "base_model:google/gemma-3-4b-it", "base_model:quantized:google/gemma-3-4b-it", "license:gemma", "endpoints_compatible", "region:us", "conversational" ]
image-text-to-text
2025-08-21T14:23:09Z
--- license: gemma library_name: transformers pipeline_tag: image-text-to-text extra_gated_heading: Access Gemma on Hugging Face extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review and agree to Google’s usage license. To do this, please ensure you’re logged in to Hugging Face and click below. Requests are processed immediately. extra_gated_button_content: Acknowledge license base_model: google/gemma-3-4b-it tags: - llama-cpp - gguf-my-repo --- # rayhaan-beeharry/gemma-3-4b-it-Q4_K_M-GGUF This model was converted to GGUF format from [`google/gemma-3-4b-it`](https://huggingface.co/google/gemma-3-4b-it) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space. Refer to the [original model card](https://huggingface.co/google/gemma-3-4b-it) for more details on the model. ## Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) ```bash brew install llama.cpp ``` Invoke the llama.cpp server or the CLI. ### CLI: ```bash llama-cli --hf-repo rayhaan-beeharry/gemma-3-4b-it-Q4_K_M-GGUF --hf-file gemma-3-4b-it-q4_k_m.gguf -p "The meaning to life and the universe is" ``` ### Server: ```bash llama-server --hf-repo rayhaan-beeharry/gemma-3-4b-it-Q4_K_M-GGUF --hf-file gemma-3-4b-it-q4_k_m.gguf -c 2048 ``` Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well. Step 1: Clone llama.cpp from GitHub. ``` git clone https://github.com/ggerganov/llama.cpp ``` Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux). ``` cd llama.cpp && LLAMA_CURL=1 make ``` Step 3: Run inference through the main binary. ``` ./llama-cli --hf-repo rayhaan-beeharry/gemma-3-4b-it-Q4_K_M-GGUF --hf-file gemma-3-4b-it-q4_k_m.gguf -p "The meaning to life and the universe is" ``` or ``` ./llama-server --hf-repo rayhaan-beeharry/gemma-3-4b-it-Q4_K_M-GGUF --hf-file gemma-3-4b-it-q4_k_m.gguf -c 2048 ```
Vasya777/blockassist-bc-lumbering_enormous_sloth_1755786145
Vasya777
2025-08-21T14:23:06Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "lumbering enormous sloth", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:22:58Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - lumbering enormous sloth --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
yuan571/gemma-3-270M-finetune-0818-change-r128-lora128-all-pos
yuan571
2025-08-21T14:22:35Z
0
0
transformers
[ "transformers", "safetensors", "gemma3_text", "text-generation", "text-generation-inference", "unsloth", "conversational", "en", "base_model:unsloth/gemma-3-270m-it", "base_model:finetune:unsloth/gemma-3-270m-it", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2025-08-21T14:21:57Z
--- base_model: unsloth/gemma-3-270m-it tags: - text-generation-inference - transformers - unsloth - gemma3_text license: apache-2.0 language: - en --- # Uploaded finetuned model - **Developed by:** yuan571 - **License:** apache-2.0 - **Finetuned from model :** unsloth/gemma-3-270m-it This gemma3_text model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
kbsooo/layoutlmv3_finetuned_doclaynet
kbsooo
2025-08-21T14:19:49Z
0
0
transformers
[ "transformers", "safetensors", "layoutlmv3", "token-classification", "dataset:ds4sd/DocLayNet-v1.2", "arxiv:2112.01041", "arxiv:1910.09700", "base_model:microsoft/layoutlmv3-base", "base_model:finetune:microsoft/layoutlmv3-base", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2025-08-21T13:37:32Z
--- library_name: transformers datasets: - ds4sd/DocLayNet-v1.2 base_model: - microsoft/layoutlmv3-base --- # Model Card for kbsooo/layoutlmv3_finetuned_doclaynet ## Model Details ### Model Description This model is a fine-tuned version of [LayoutLMv3](https://huggingface.co/microsoft/layoutlmv3-base) for token classification on the DocLayNet dataset. It is designed to classify each token in a document image based on both textual and layout information. - **Developed by:** kbsooo - **Model type:** LayoutLMv3ForTokenClassification - **Language(s) (NLP):** Korean (document-oriented) - **License:** Check DocLayNet and LayoutLMv3 licenses - **Finetuned from model:** microsoft/layoutlmv3-base ### Model Sources - **Repository:** [Hugging Face Model Hub](https://huggingface.co/kbsooo/layoutlmv3_finetuned_doclaynet) - **Paper (optional):** [LayoutLMv3 Paper](https://arxiv.org/abs/2112.01041) ## Uses ### Direct Use This model can be used for: - Token classification in document images (e.g., identifying headings, paragraphs, tables, images, lists) - Document understanding tasks where layout + text information is important ### Downstream Use - Can be integrated into pipelines for document information extraction - Useful for document analysis applications: invoice parsing, form processing, etc. ### Out-of-Scope Use - Not intended for languages or layouts not represented in the DocLayNet dataset - Not suitable for free-form text without document structure ## Bias, Risks, and Limitations - The model may misclassify tokens if the document layout or language differs from the training data - Biases may exist due to dataset composition (DocLayNet) - Limited to 10 classes of document layout elements ### Recommendations - Users should preprocess documents similarly to the training setup (tokenization + bounding boxes + image) - Verify predictions, especially in production or high-stakes scenarios ## How to Get Started with the Model ```python from transformers import LayoutLMv3ForTokenClassification, AutoProcessor import torch repo = "kbsooo/layoutlmv3_finetuned_doclaynet" model = LayoutLMv3ForTokenClassification.from_pretrained(repo) processor = AutoProcessor.from_pretrained(repo) image = ... # PIL.Image or np.array text = "Sample document text" encoding = processor(image, text, return_tensors="pt") outputs = model(**encoding) preds = torch.argmax(outputs.logits, dim=-1) print(preds) ``` ## Training Details ### Training Data - Dataset: DocLayNet-v1.2 - Train/Validation split: 200/100 samples - Columns: input_ids, attention_mask, bbox, labels, pixel_values, n_words_in, n_words_out ### Training Procedure - Optimizer: AdamW - Learning rate: 5e-5 - Epochs: 5 - Mixed precision: FP16 optional - Loss: Cross-entropy per token ## Evaluation - Sample metrics (from validation set): - Avg Train Loss: 0.134 - Avg Val Loss: 0.458 - Token prediction accuracy should be checked against the DocLayNet labels ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** NVIDIA A100 - **Hours used:** ~1 hr for 5 epochs (for small dataset) ## Technical Specifications ### Model Architecture and Objective - Base model: LayoutLMv3 - Task: Token classification for document layout elements - Input: Tokenized text, bounding boxes, and document images - Output: Token-wise logits for 10 classes ### Compute Infrastructure - Training performed on Google Colab Pro (A100 GPU) - Framework: PyTorch + Hugging Face Transformers ## Citation <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** ```bibtex @article{huang2022layoutlmv3, title={LayoutLMv3: Pre-training for Document AI with Unified Text and Image Masking}, author={Huang, Zejiang and et al.}, journal={arXiv preprint arXiv:2112.01041}, year={2022} } ``` **APA:** Huang, Z., et al. (2022). LayoutLMv3: Pre-training for Document AI with Unified Text and Image Masking. arXiv preprint arXiv:2112.01041.
vwzyrraz7l/blockassist-bc-tall_hunting_vulture_1755784274
vwzyrraz7l
2025-08-21T14:19:34Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "tall hunting vulture", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:19:30Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - tall hunting vulture --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
helmutsukocok/blockassist-bc-loud_scavenging_kangaroo_1755784378
helmutsukocok
2025-08-21T14:19:24Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "loud scavenging kangaroo", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:19:21Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - loud scavenging kangaroo --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
llencia/blockassist-bc-wiry_wise_hedgehog_1755785919
llencia
2025-08-21T14:19:09Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "wiry wise hedgehog", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:19:01Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - wiry wise hedgehog --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Dejiat/blockassist-bc-savage_unseen_bobcat_1755785847
Dejiat
2025-08-21T14:18:06Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "savage unseen bobcat", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:18:00Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - savage unseen bobcat --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Medved444/blockassist-bc-bellowing_finicky_manatee_1755784628
Medved444
2025-08-21T14:16:55Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "bellowing finicky manatee", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:16:07Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - bellowing finicky manatee --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
thanobidex/blockassist-bc-colorful_shiny_hare_1755784187
thanobidex
2025-08-21T14:16:25Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "colorful shiny hare", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:16:22Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - colorful shiny hare --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
papersail/DeepSeek-R1-0528-FP8INT4G
papersail
2025-08-21T14:15:39Z
0
0
transformers
[ "transformers", "safetensors", "deepseek_v3", "text-generation", "conversational", "custom_code", "arxiv:2501.12948", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "fp8", "region:us" ]
text-generation
2025-08-21T11:52:43Z
--- license: mit library_name: transformers --- # DeepSeek-R1-0528 <!-- markdownlint-disable first-line-h1 --> <!-- markdownlint-disable html --> <!-- markdownlint-disable no-duplicate-header --> <div align="center"> <img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" /> </div> <hr> <div align="center" style="line-height: 1;"> <a href="https://www.deepseek.com/" target="_blank" style="margin: 2px;"> <img alt="Homepage" src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/badge.svg?raw=true" style="display: inline-block; vertical-align: middle;"/> </a> <a href="https://chat.deepseek.com/" target="_blank" style="margin: 2px;"> <img alt="Chat" src="https://img.shields.io/badge/🤖%20Chat-DeepSeek%20R1-536af5?color=536af5&logoColor=white" style="display: inline-block; vertical-align: middle;"/> </a> <a href="https://huggingface.co/deepseek-ai" target="_blank" style="margin: 2px;"> <img alt="Hugging Face" src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-DeepSeek%20AI-ffc107?color=ffc107&logoColor=white" style="display: inline-block; vertical-align: middle;"/> </a> </div> <div align="center" style="line-height: 1;"> <a href="https://discord.gg/Tc7c45Zzu5" target="_blank" style="margin: 2px;"> <img alt="Discord" src="https://img.shields.io/badge/Discord-DeepSeek%20AI-7289da?logo=discord&logoColor=white&color=7289da" style="display: inline-block; vertical-align: middle;"/> </a> <a href="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/qr.jpeg?raw=true" target="_blank" style="margin: 2px;"> <img alt="Wechat" src="https://img.shields.io/badge/WeChat-DeepSeek%20AI-brightgreen?logo=wechat&logoColor=white" style="display: inline-block; vertical-align: middle;"/> </a> <a href="https://twitter.com/deepseek_ai" target="_blank" style="margin: 2px;"> <img alt="Twitter Follow" src="https://img.shields.io/badge/Twitter-deepseek_ai-white?logo=x&logoColor=white" style="display: inline-block; vertical-align: middle;"/> </a> </div> <div align="center" style="line-height: 1;"> <a href="LICENSE" style="margin: 2px;"> <img alt="License" src="https://img.shields.io/badge/License-MIT-f5de53?&color=f5de53" style="display: inline-block; vertical-align: middle;"/> </a> </div> <p align="center"> <a href="https://arxiv.org/pdf/2501.12948"><b>Paper Link</b>👁️</a> </p> ## 1. Introduction The DeepSeek R1 model has undergone a minor version upgrade, with the current version being DeepSeek-R1-0528. In the latest update, DeepSeek R1 has significantly improved its depth of reasoning and inference capabilities by leveraging increased computational resources and introducing algorithmic optimization mechanisms during post-training. The model has demonstrated outstanding performance across various benchmark evaluations, including mathematics, programming, and general logic. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. <p align="center"> <img width="80%" src="figures/benchmark.png"> </p> Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks. For instance, in the AIME 2025 test, the model’s accuracy has increased from 70% in the previous version to 87.5% in the current version. This advancement stems from enhanced thinking depth during the reasoning process: in the AIME test set, the previous model used an average of 12K tokens per question, whereas the new version averages 23K tokens per question. Beyond its improved reasoning capabilities, this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. ## 2. Evaluation Results ### DeepSeek-R1-0528 For all our models, the maximum generation length is set to 64K tokens. For benchmarks requiring sampling, we use a temperature of $0.6$, a top-p value of $0.95$, and generate 16 responses per query to estimate pass@1. <div align="center"> | Category | Benchmark (Metric) | DeepSeek R1 | DeepSeek R1 0528 |----------|----------------------------------|-----------------|---| | General | | | MMLU-Redux (EM) | 92.9 | 93.4 | | MMLU-Pro (EM) | 84.0 | 85.0 | | GPQA-Diamond (Pass@1) | 71.5 | 81.0 | | SimpleQA (Correct) | 30.1 | 27.8 | | FRAMES (Acc.) | 82.5 | 83.0 | | Humanity's Last Exam (Pass@1) | 8.5 | 17.7 | Code | | | LiveCodeBench (2408-2505) (Pass@1) | 63.5 | 73.3 | | Codeforces-Div1 (Rating) | 1530 | 1930 | | SWE Verified (Resolved) | 49.2 | 57.6 | | Aider-Polyglot (Acc.) | 53.3 | 71.6 | Math | | | AIME 2024 (Pass@1) | 79.8 | 91.4 | | AIME 2025 (Pass@1) | 70.0 | 87.5 | | HMMT 2025 (Pass@1) | 41.7 | 79.4 | | | CNMO 2024 (Pass@1) | 78.8 | 86.9 | Tools | | | BFCL_v3_MultiTurn (Acc) | - | 37.0 | | | Tau-Bench (Pass@1) | - | 53.5(Airline)/63.9(Retail) </div> Note: We use Agentless framework to evaluate model performance on SWE-Verified. We only evaluate text-only prompts in HLE testsets. GPT-4.1 is employed to act user role in Tau-bench evaluation. ### DeepSeek-R1-0528-Qwen3-8B Meanwhile, we distilled the chain-of-thought from DeepSeek-R1-0528 to post-train Qwen3 8B Base, obtaining DeepSeek-R1-0528-Qwen3-8B. This model achieves state-of-the-art (SOTA) performance among open-source models on the AIME 2024, surpassing Qwen3 8B by +10.0% and matching the performance of Qwen3-235B-thinking. We believe that the chain-of-thought from DeepSeek-R1-0528 will hold significant importance for both academic research on reasoning models and industrial development focused on small-scale models. | | AIME 24 | AIME 25 | HMMT Feb 25 | GPQA Diamond | LiveCodeBench (2408-2505) | |--------------------------------|---------|---------|-------------|--------------|---------------------------| | Qwen3-235B-A22B | 85.7 | 81.5 | 62.5 | 71.1 | 66.5 | | Qwen3-32B | 81.4 | 72.9 | - | 68.4 | - | | Qwen3-8B | 76.0 | 67.3 | - | 62.0 | - | | Phi-4-Reasoning-Plus-14B | 81.3 | 78.0 | 53.6 | 69.3 | - | | Gemini-2.5-Flash-Thinking-0520 | 82.3 | 72.0 | 64.2 | 82.8 | 62.3 | | o3-mini (medium) | 79.6 | 76.7 | 53.3 | 76.8 | 65.9 | | DeepSeek-R1-0528-Qwen3-8B | 86.0 | 76.3 | 61.5 | 61.1 | 60.5 | ## 3. Chat Website & API Platform You can chat with DeepSeek-R1 on DeepSeek's official website: [chat.deepseek.com](https://chat.deepseek.com/sign_in), and switch on the button "DeepThink" We also provide OpenAI-Compatible API at DeepSeek Platform: [platform.deepseek.com](https://platform.deepseek.com/) ## 4. How to Run Locally Please visit [DeepSeek-R1](https://github.com/deepseek-ai/DeepSeek-R1) repository for more information about running DeepSeek-R1-0528 locally. Compared to previous versions of DeepSeek-R1, the usage recommendations for DeepSeek-R1-0528 have the following changes: 1. System prompt is supported now. 2. It is not required to add "\<think\>\n" at the beginning of the output to force the model into thinking pattern. The model architecture of DeepSeek-R1-0528-Qwen3-8B is identical to that of Qwen3-8B, but it shares the same tokenizer configuration as DeepSeek-R1-0528. This model can be run in the same manner as Qwen3-8B. ### System Prompt In the official DeepSeek web/app, we use the same system prompt with a specific date. ``` 该助手为DeepSeek-R1,由深度求索公司创造。 今天是{current date}。 ``` For example, ``` 该助手为DeepSeek-R1,由深度求索公司创造。 今天是2025年5月28日,星期一。 ``` ### Temperature In our web and application environments, the temperature parameter $T_{model}$ is set to 0.6. ### Prompts for File Uploading and Web Search For file uploading, please follow the template to create prompts, where {file_name}, {file_content} and {question} are arguments. ``` file_template = \ """[file name]: {file_name} [file content begin] {file_content} [file content end] {question}""" ``` For Web Search, {search_results}, {cur_date}, and {question} are arguments. For Chinese query, we use the prompt: ``` search_answer_zh_template = \ '''# 以下内容是基于用户发送的消息的搜索结果: {search_results} 在我给你的搜索结果中,每个结果都是[webpage X begin]...[webpage X end]格式的,X代表每篇文章的数字索引。请在适当的情况下在句子末尾引用上下文。请按照引用编号[citation:X]的格式在答案中对应部分引用上下文。如果一句话源自多个上下文,请列出所有相关的引用编号,例如[citation:3][citation:5],切记不要将引用集中在最后返回引用编号,而是在答案对应部分列出。 在回答时,请注意以下几点: - 今天是{cur_date}。 - 并非搜索结果的所有内容都与用户的问题密切相关,你需要结合问题,对搜索结果进行甄别、筛选。 - 对于列举类的问题(如列举所有航班信息),尽量将答案控制在10个要点以内,并告诉用户可以查看搜索来源、获得完整信息。优先提供信息完整、最相关的列举项;如非必要,不要主动告诉用户搜索结果未提供的内容。 - 对于创作类的问题(如写论文),请务必在正文的段落中引用对应的参考编号,例如[citation:3][citation:5],不能只在文章末尾引用。你需要解读并概括用户的题目要求,选择合适的格式,充分利用搜索结果并抽取重要信息,生成符合用户要求、极具思想深度、富有创造力与专业性的答案。你的创作篇幅需要尽可能延长,对于每一个要点的论述要推测用户的意图,给出尽可能多角度的回答要点,且务必信息量大、论述详尽。 - 如果回答很长,请尽量结构化、分段落总结。如果需要分点作答,尽量控制在5个点以内,并合并相关的内容。 - 对于客观类的问答,如果问题的答案非常简短,可以适当补充一到两句相关信息,以丰富内容。 - 你需要根据用户要求和回答内容选择合适、美观的回答格式,确保可读性强。 - 你的回答应该综合多个相关网页来回答,不能重复引用一个网页。 - 除非用户要求,否则你回答的语言需要和用户提问的语言保持一致。 # 用户消息为: {question}''' ``` For English query, we use the prompt: ``` search_answer_en_template = \ '''# The following contents are the search results related to the user's message: {search_results} In the search results I provide to you, each result is formatted as [webpage X begin]...[webpage X end], where X represents the numerical index of each article. Please cite the context at the end of the relevant sentence when appropriate. Use the citation format [citation:X] in the corresponding part of your answer. If a sentence is derived from multiple contexts, list all relevant citation numbers, such as [citation:3][citation:5]. Be sure not to cluster all citations at the end; instead, include them in the corresponding parts of the answer. When responding, please keep the following points in mind: - Today is {cur_date}. - Not all content in the search results is closely related to the user's question. You need to evaluate and filter the search results based on the question. - For listing-type questions (e.g., listing all flight information), try to limit the answer to 10 key points and inform the user that they can refer to the search sources for complete information. Prioritize providing the most complete and relevant items in the list. Avoid mentioning content not provided in the search results unless necessary. - For creative tasks (e.g., writing an essay), ensure that references are cited within the body of the text, such as [citation:3][citation:5], rather than only at the end of the text. You need to interpret and summarize the user's requirements, choose an appropriate format, fully utilize the search results, extract key information, and generate an answer that is insightful, creative, and professional. Extend the length of your response as much as possible, addressing each point in detail and from multiple perspectives, ensuring the content is rich and thorough. - If the response is lengthy, structure it well and summarize it in paragraphs. If a point-by-point format is needed, try to limit it to 5 points and merge related content. - For objective Q&A, if the answer is very brief, you may add one or two related sentences to enrich the content. - Choose an appropriate and visually appealing format for your response based on the user's requirements and the content of the answer, ensuring strong readability. - Your answer should synthesize information from multiple relevant webpages and avoid repeatedly citing the same webpage. - Unless the user requests otherwise, your response should be in the same language as the user's question. # The user's message is: {question}''' ``` ## 5. License This code repository is licensed under [MIT License](LICENSE). The use of DeepSeek-R1 models is also subject to [MIT License](LICENSE). DeepSeek-R1 series (including Base and Chat) supports commercial use and distillation. ## 6. Citation ``` @misc{deepseekai2025deepseekr1incentivizingreasoningcapability, title={DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning}, author={DeepSeek-AI}, year={2025}, eprint={2501.12948}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2501.12948}, } ``` ## 7. Contact If you have any questions, please raise an issue or contact us at [service@deepseek.com](service@deepseek.com).
llencia/blockassist-bc-wiry_wise_hedgehog_1755785693
llencia
2025-08-21T14:15:25Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "wiry wise hedgehog", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:15:17Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - wiry wise hedgehog --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Muapi/flux-such-an-asian-skin-beauty
Muapi
2025-08-21T14:14:52Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-21T14:14:42Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Flux : Such an Asian Skin Beauty ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: SuchSkin ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:746469@930709", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
Muapi/bouguereau-style-with-hunyuan
Muapi
2025-08-21T14:14:03Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-21T14:13:52Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Bouguereau style (with hunyuan) ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: Bouguereau style ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:400280@937384", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
sothmik/waiNSFWIllustrious-v110-Q8-GGUF
sothmik
2025-08-21T14:13:01Z
5
0
null
[ "gguf", "text-to-image", "region:us" ]
text-to-image
2025-08-20T10:09:37Z
--- pipeline_tag: text-to-image --- From https://civitai.com/models/827184?modelVersionId=1410435
Muapi/vogue-flux
Muapi
2025-08-21T14:11:09Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-21T14:10:53Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Vogue-Flux ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:937057@1048983", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
Shan171153/finetuned_model
Shan171153
2025-08-21T14:11:08Z
0
0
peft
[ "peft", "safetensors", "base_model:adapter:unsloth/gpt-oss-20b-unsloth-bnb-4bit", "lora", "sft", "transformers", "trl", "unsloth", "arxiv:1910.09700", "base_model:unsloth/gpt-oss-20b-unsloth-bnb-4bit", "region:us" ]
null
2025-08-21T14:05:37Z
--- base_model: unsloth/gpt-oss-20b-unsloth-bnb-4bit library_name: peft tags: - base_model:adapter:unsloth/gpt-oss-20b-unsloth-bnb-4bit - lora - sft - transformers - trl - unsloth --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.17.0
Bartosh16/Bielik-1-5B-DanielB
Bartosh16
2025-08-21T14:10:26Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "autotrain", "text-generation-inference", "text-generation", "peft", "conversational", "base_model:speakleash/Bielik-1.5B-v3", "base_model:finetune:speakleash/Bielik-1.5B-v3", "license:other", "endpoints_compatible", "region:us" ]
text-generation
2025-08-21T13:52:02Z
--- tags: - autotrain - text-generation-inference - text-generation - peft library_name: transformers base_model: speakleash/Bielik-1.5B-v3 widget: - messages: - role: user content: What is your favorite condiment? license: other --- # Model Trained Using AutoTrain This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain). # Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_path = "PATH_TO_THIS_REPO" tokenizer = AutoTokenizer.from_pretrained(model_path) model = AutoModelForCausalLM.from_pretrained( model_path, device_map="auto", torch_dtype='auto' ).eval() # Prompt content: "hi" messages = [ {"role": "user", "content": "hi"} ] input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt') output_ids = model.generate(input_ids.to('cuda')) response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True) # Model response: "Hello! How can I assist you today?" print(response) ```
Muapi/caravaggio-baroque-painting
Muapi
2025-08-21T14:10:10Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-21T14:10:00Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Caravaggio Baroque Painting ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: Baroque oil painting by Caravaggio circa 1600 ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:820661@917698", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
manusiaperahu2012/blockassist-bc-roaring_long_tuna_1755783692
manusiaperahu2012
2025-08-21T14:09:32Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "roaring long tuna", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:09:29Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - roaring long tuna --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
pidbu/blockassist-bc-whistling_alert_shrew_1755785274
pidbu
2025-08-21T14:09:23Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "whistling alert shrew", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:08:49Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - whistling alert shrew --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Muapi/zavy-s-fine-art-photography-flux
Muapi
2025-08-21T14:08:34Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-21T14:08:22Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Zavy's Fine Art Photography - Flux ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: zavy-fnrt ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:737956@825288", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
ssuki/qwen-bi-intent-model-20250821_140803
ssuki
2025-08-21T14:08:12Z
0
0
peft
[ "peft", "safetensors", "base_model:adapter:Qwen/Qwen2.5-0.5B", "lora", "transformers", "text-generation", "conversational", "arxiv:1910.09700", "base_model:Qwen/Qwen2.5-0.5B", "region:us" ]
text-generation
2025-08-21T14:08:03Z
--- base_model: Qwen/Qwen2.5-0.5B library_name: peft pipeline_tag: text-generation tags: - base_model:adapter:Qwen/Qwen2.5-0.5B - lora - transformers --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.17.0
Muapi/better-looking-men-flux
Muapi
2025-08-21T14:06:55Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-21T14:06:38Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Better looking men FLUX ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: good looking man, h5ns0 ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:855728@1156020", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
Bisher/upbeat-fog-32_merged_default
Bisher
2025-08-21T14:06:42Z
0
0
transformers
[ "transformers", "safetensors", "whisper", "automatic-speech-recognition", "text-generation-inference", "unsloth", "en", "base_model:unsloth/whisper-large-v3-turbo", "base_model:finetune:unsloth/whisper-large-v3-turbo", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2025-08-21T14:02:12Z
--- base_model: unsloth/whisper-large-v3-turbo tags: - text-generation-inference - transformers - unsloth - whisper license: apache-2.0 language: - en --- # Uploaded finetuned model - **Developed by:** Bisher - **License:** apache-2.0 - **Finetuned from model :** unsloth/whisper-large-v3-turbo This whisper model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
indoempatnol/blockassist-bc-fishy_wary_swan_1755783538
indoempatnol
2025-08-21T14:06:03Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "fishy wary swan", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:05:59Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - fishy wary swan --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
quantumxnode/blockassist-bc-dormant_peckish_seahorse_1755783644
quantumxnode
2025-08-21T14:06:02Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "dormant peckish seahorse", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:05:59Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - dormant peckish seahorse --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Osrivers/hunyuanVideoSafetensors_comfyDiffusionFP8.safetensors
Osrivers
2025-08-21T14:05:41Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2025-08-21T13:55:21Z
--- license: creativeml-openrail-m ---
Bisher/upbeat-fog-32merged_16bit
Bisher
2025-08-21T14:04:46Z
0
0
transformers
[ "transformers", "safetensors", "whisper", "automatic-speech-recognition", "text-generation-inference", "unsloth", "en", "base_model:unsloth/whisper-large-v3-turbo", "base_model:finetune:unsloth/whisper-large-v3-turbo", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2025-08-21T14:01:58Z
--- base_model: unsloth/whisper-large-v3-turbo tags: - text-generation-inference - transformers - unsloth - whisper license: apache-2.0 language: - en --- # Uploaded finetuned model - **Developed by:** Bisher - **License:** apache-2.0 - **Finetuned from model :** unsloth/whisper-large-v3-turbo This whisper model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
Vasya777/blockassist-bc-lumbering_enormous_sloth_1755784985
Vasya777
2025-08-21T14:03:46Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "lumbering enormous sloth", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:03:38Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - lumbering enormous sloth --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
llencia/blockassist-bc-wiry_wise_hedgehog_1755784974
llencia
2025-08-21T14:03:27Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "wiry wise hedgehog", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T14:03:20Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - wiry wise hedgehog --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Bisher/upbeat-fog-32_lora
Bisher
2025-08-21T14:02:26Z
0
0
transformers
[ "transformers", "safetensors", "text-generation-inference", "unsloth", "whisper", "trl", "en", "base_model:unsloth/whisper-large-v3-turbo", "base_model:finetune:unsloth/whisper-large-v3-turbo", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2025-08-21T14:02:17Z
--- base_model: unsloth/whisper-large-v3-turbo tags: - text-generation-inference - transformers - unsloth - whisper - trl license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** Bisher - **License:** apache-2.0 - **Finetuned from model :** unsloth/whisper-large-v3-turbo This whisper model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
eshanroy5678/blockassist-bc-untamed_dextrous_dingo_1755784397
eshanroy5678
2025-08-21T14:02:00Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "untamed dextrous dingo", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T13:57:10Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - untamed dextrous dingo --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
runchat/lora-254b955f-0e25-4cc1-9a23-32a45289d521-uf6gwg
runchat
2025-08-21T14:01:27Z
0
0
diffusers
[ "diffusers", "flux", "lora", "text-to-image", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-08-21T14:01:22Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md base_model: black-forest-labs/FLUX.1-dev tags: - flux - lora - diffusers - text-to-image widget: - text: 'a photo of a sks style' output: url: "placeholder.jpg" --- # Flux LoRA: sks This is a LoRA (Low-Rank Adaptation) model for Flux.1-dev fine-tuned on images with the trigger word `sks`. ## Files - `pytorch_lora_weights.safetensors`: Diffusers format (use with diffusers library) - `pytorch_lora_weights_webui.safetensors`: Kohya format (use with AUTOMATIC1111, ComfyUI, etc.) ## Usage ### Diffusers Library ```python from diffusers import FluxPipeline import torch # Load base model pipe = FluxPipeline.from_pretrained( "black-forest-labs/FLUX.1-dev", torch_dtype=torch.bfloat16 ) # Load LoRA weights (diffusers format) pipe.load_lora_weights("runchat/lora-254b955f-0e25-4cc1-9a23-32a45289d521-uf6gwg", weight_name="pytorch_lora_weights.safetensors") pipe = pipe.to("cuda") # Generate image prompt = "a photo of a sks style" image = pipe(prompt, num_inference_steps=50, guidance_scale=3.5).images[0] image.save("output.png") ``` ### WebUI (AUTOMATIC1111, ComfyUI, etc.) Download the `pytorch_lora_weights_webui.safetensors` file and place it in your WebUI's LoRA directory. Use the trigger word `sks` in your prompts. ## Training Details - Base model: black-forest-labs/FLUX.1-dev - Training steps: 500 - Learning rate: 0.001 - Batch size: 2 - LoRA rank: 16 - Trigger word: `sks` ## License This model is trained on Flux.1-dev and inherits its non-commercial license. Please see the [license](https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md) for usage restrictions.
inoora/phi3_custom_test
inoora
2025-08-21T13:59:39Z
51
0
peft
[ "peft", "safetensors", "phi3", "text-generation", "base_model:adapter:microsoft/Phi-3-mini-4k-instruct", "lora", "sft", "transformers", "trl", "conversational", "custom_code", "arxiv:1910.09700", "base_model:microsoft/Phi-3-mini-4k-instruct", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-08-18T08:55:27Z
--- base_model: microsoft/Phi-3-mini-4k-instruct library_name: peft pipeline_tag: text-generation tags: - base_model:adapter:microsoft/Phi-3-mini-4k-instruct - lora - sft - transformers - trl --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.17.0
chainway9/blockassist-bc-untamed_quick_eel_1755783024
chainway9
2025-08-21T13:58:57Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "untamed quick eel", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T13:58:54Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - untamed quick eel --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
fif3/MyGemmaNPC
fif3
2025-08-21T13:58:12Z
3
0
transformers
[ "transformers", "tensorboard", "safetensors", "gemma3_text", "text-generation", "generated_from_trainer", "sft", "trl", "conversational", "base_model:google/gemma-3-270m-it", "base_model:finetune:google/gemma-3-270m-it", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-08-20T15:05:45Z
--- base_model: google/gemma-3-270m-it library_name: transformers model_name: MyGemmaNPC tags: - generated_from_trainer - sft - trl licence: license --- # Model Card for MyGemmaNPC This model is a fine-tuned version of [google/gemma-3-270m-it](https://huggingface.co/google/gemma-3-270m-it). It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="fif3/MyGemmaNPC", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure This model was trained with SFT. ### Framework versions - TRL: 0.21.0 - Transformers: 4.55.2 - Pytorch: 2.8.0+cu126 - Datasets: 4.0.0 - Tokenizers: 0.21.4 ## Citations Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
qing223101/blockassist-bc-mangy_deft_hippo_1755782205
qing223101
2025-08-21T13:55:13Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "mangy deft hippo", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T13:54:38Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - mangy deft hippo --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Levarat/blockassist-bc-scavenging_small_pelican_1755784433
Levarat
2025-08-21T13:54:36Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "scavenging small pelican", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T13:54:22Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - scavenging small pelican --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
sampingkaca72/blockassist-bc-armored_stealthy_elephant_1755782876
sampingkaca72
2025-08-21T13:53:59Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "armored stealthy elephant", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T13:53:54Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - armored stealthy elephant --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
koloni/blockassist-bc-deadly_graceful_stingray_1755782617
koloni
2025-08-21T13:50:43Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "deadly graceful stingray", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T13:50:40Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - deadly graceful stingray --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Agnuxo/NEBULA-X
Agnuxo
2025-08-21T13:50:22Z
3
0
transformers
[ "transformers", "pytorch", "nebula-x", "holographic-neural-networks", "quantum-computing", "optical-computing", "raytracing", "photonic-neural-networks", "text-generation", "en", "dataset:cais/mmlu", "dataset:gsm8k", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
text-generation
2025-08-19T10:25:01Z
--- license: apache-2.0 language: - en library_name: transformers tags: - holographic-neural-networks - quantum-computing - optical-computing - raytracing - nebula-x - photonic-neural-networks datasets: - cais/mmlu - gsm8k metrics: - accuracy - holographic_coherence - quantum_entanglement pipeline_tag: text-generation model-index: - name: NEBULA-X results: - task: type: text-generation name: Text Generation dataset: name: MMLU type: cais/mmlu metrics: - type: accuracy value: 0.85 name: MMLU Accuracy - task: type: text-generation name: Mathematical Reasoning dataset: name: GSM8K type: gsm8k metrics: - type: accuracy value: 0.78 name: GSM8K Accuracy --- # 🌌 NEBULA-X: Enhanced Unified Holographic Neural Network **Winner of NVIDIA LlamaIndex Developer Contest 2024** NEBULA-X is a revolutionary AI architecture that combines holographic memory, quantum computing, and optical neural networks to create the world's first production-ready photonic neural network system. ## 🔬 Key Technologies ### Holographic Neural Networks - **Holographic Memory**: Information stored as interference patterns in 3D space - **Light-based Processing**: Neurons represented as points of light with optical properties - **Interferometric Computing**: Calculations performed through wave interference ### Quantum-Enhanced Processing - **4 Qubits per Neuron**: Distributed quantum memory for enhanced processing - **Quantum Entanglement**: Non-local correlations between neural components - **Superposition States**: Parallel processing of multiple possibilities ### Optical Raytracing - **GPU-Accelerated**: CUDA kernels for Monte Carlo raytracing - **Real-time Physics**: Accurate simulation of light propagation - **Material Properties**: Reflectivity, transmittance, and phase shifts ## 🏆 Performance | Benchmark | Score | Improvement vs Baseline | |-----------|-------|------------------------| | MMLU | 85.0% | +240% | | GSM8K | 78.0% | +∞% (baseline: 0%) | | HellaSwag | 92.3% | +152% | | ARC | 88.7% | +198% | ## 🚀 Quick Start ```python from transformers import AutoModel, AutoTokenizer import torch # Load model and tokenizer model = AutoModel.from_pretrained("Agnuxo/NEBULA-X") tokenizer = AutoTokenizer.from_pretrained("Agnuxo/NEBULA-X") # Encode input inputs = tokenizer("What is quantum holography?", return_tensors="pt") # Generate response with holographic processing with torch.no_grad(): outputs = model(**inputs) predictions = torch.softmax(outputs.logits, dim=-1) ``` ## 👨‍💻 Author **Francisco Angulo de Lafuente (Agnuxo)** - Research Focus: Holographic Computing, Quantum AI, Optical Neural Networks - NVIDIA LlamaIndex Developer Contest 2024 Winner - 27+ Repositories in Advanced AI Architectures ## 📄 License Apache 2.0 - See LICENSE file for details. NEBULA-X represents a paradigm shift in AI architecture, combining the power of light, quantum mechanics, and evolutionary algorithms to create truly intelligent systems.
llencia/blockassist-bc-wiry_wise_hedgehog_1755784191
llencia
2025-08-21T13:50:22Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "wiry wise hedgehog", "arxiv:2504.07091", "region:us" ]
null
2025-08-21T13:50:14Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - wiry wise hedgehog --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
maziyaramini/llama3.2-1b-persian-sentiment-final-ins
maziyaramini
2025-08-21T13:49:35Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "llama", "text-generation", "generated_from_trainer", "trl", "sft", "conversational", "base_model:meta-llama/Llama-3.2-1B-Instruct", "base_model:finetune:meta-llama/Llama-3.2-1B-Instruct", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-08-21T11:23:19Z
--- base_model: meta-llama/Llama-3.2-1B-Instruct library_name: transformers model_name: llama3.2-1b-persian-sentiment-final-ins tags: - generated_from_trainer - trl - sft licence: license --- # Model Card for llama3.2-1b-persian-sentiment-final-ins This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct). It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="maziyaramini/llama3.2-1b-persian-sentiment-final-ins", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure This model was trained with SFT. ### Framework versions - TRL: 0.21.0 - Transformers: 4.55.2 - Pytorch: 2.8.0+cu126 - Datasets: 4.0.0 - Tokenizers: 0.21.4 ## Citations Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```